Request: %s
\" % self.path, \"utf-8\"))\n self.wfile.write(bytes(\"\", \"utf-8\"))\n self.wfile.write(bytes(\"This is an example web server.
\", \"utf-8\"))\n self.wfile.write(bytes(\"\", \"utf-8\"))\n\nif __name__ == \"__main__\": \n webServer = HTTPServer((host_name, port), my_server)\n print(\"Server started http://%s:%s\" % (host_name, port))\n\n try:\n webServer.serve_forever()\n except KeyboardInterrupt:\n pass\n\n webServer.server_close()\n print(\"Server stopped.\")","sub_path":"python2-web-server.py","file_name":"python2-web-server.py","file_ext":"py","file_size_in_byte":1032,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"194886568","text":"# Given a list of strings, write a function that returns the longest string chain that \n# can be buit from those strings \n\n# O(n * m^2 + nlog(n)) time, O(nm) space, n = number of string, m = length of the longest string\ndef longestStringChain(strings):\n\t# For every string, imagine the longest string chain that starts with it.\n\t# Set up every string to point to the next string in its respective longest\n\t# string chain. Also keep track of the lengths of these longest string chains.\n\tstringChains = {}\n\tfor string in strings:\n\t\tstringChains[string] = {\"nextString\": \"\", \"maxChainLength\": 1}\n\t\t\n\t\n\t# Sort the strings based on their length so that whenever we visit a \n\t# string (as we iterate through them from left to right), we can\n\t# already have computed the longest string chains of any smaller strings.\n\tsortedStrings = sorted(strings, key = len)\n\tfor string in sortedStrings:\n\t\tfindLongestStringChain(string, stringChains)\n\t\t\n\treturn buildLongestStringChain(strings, stringChains)\n\ndef findLongestStringChain(string, stringChains):\n\t# Try removing every letter of the current string to see if the\n\t# remaining strings form a string chain\n\tfor i in range(len(string)):\n\t\tsmallerString = getSmallerString(string, i)\n\t\tif smallerString not in stringChains:\n\t\t\tcontinue\n\t\ttryUpdateLongestStringChain(string, smallerString, stringChains)\n\t\t\ndef getSmallerString(string, index):\n\treturn string[0:index] + string[index + 1 : ]\n\n\ndef tryUpdateLongestStringChain(currentString, smallerString, stringChains):\n\tsmallerStringChainLength = stringChains[smallerString][\"maxChainLength\"]\n\tcurrentStringChainLength = stringChains[currentString][\"maxChainLength\"]\n\t# Update the string chain of the current string only if the smaller string leads\n\t# to a longer string chain\n\tif smallerStringChainLength + 1 > currentStringChainLength:\n\t\tstringChains[currentString][\"maxChainLength\"] = smallerStringChainLength + 1\n\t\tstringChains[currentString][\"nextString\"] = smallerString\n\t\t\ndef buildLongestStringChain(strings, stringChains):\n\t# Find the string that starts the longest string chain\n\tmaxChainLength = 0\n\tchainStartingString = \"\"\n\tfor string in strings:\n\t\tif stringChains[string][\"maxChainLength\"] > maxChainLength:\n\t\t\tmaxChainLength = stringChains[string][\"maxChainLength\"]\n\t\t\tchainStartingString = string\n\t\n\t# Starting at the string found above, build the longest string chain.\n\tourLongestStringChain = []\n\tcurrentString = chainStartingString\n\twhile currentString != \"\":\n\t\tourLongestStringChain.append(currentString)\n\t\tcurrentString = stringChains[currentString][\"nextString\"]\n\t\t\n\treturn [] if len(ourLongestStringChain) == 1 else ourLongestStringChain\n","sub_path":"Very Hard/longestStringChain.py","file_name":"longestStringChain.py","file_ext":"py","file_size_in_byte":2649,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"283010903","text":"import random\nimport copy\nimport time\nimport sys\nimport matplotlib.pyplot as plt\nimport os\nfrom multiprocessing import Pool\nimport multiprocessing as multi\nfrom ship import Ship\nfrom tqdm import tqdm\nimport slackweb\n# import own modules #\nsys.path.append('../public')\nsys.path.append('../output')\nfrom constants import *\nfrom my_modules import *\n\nclass GA_Extended:\n\n def __init__(self,oil_price_data,freight_rate_outward,freight_rate_homeward,exchange_rate,demand,supply,newbuilding,secondhand,actionlist=None,generation=None,population_size=None,crossover_rate=None,mutation_rate=None):\n self.oil_price_data = oil_price_data #oil price predicted data\n self.freight_rate_outward_data = freight_rate_outward #freight rate outward predicted data\n self.freight_rate_homeward_data = freight_rate_homeward # freight rate return predicted data\n self.exchange_rate_data = exchange_rate # exchange_rate predicted data\n self.demand_data = demand#ship demand predicted data\n self.supply_data = supply#ship supply predicted data\n self.newbuilding = newbuilding#new building ship price data\n self.secondhand = secondhand#secondhand ship price data\n self.actionlist = None if actionlist else None # decision of action parts.\n self.generation = generation if generation else DEFAULT_GENERATION # the number of generation\n self.population_size = population_size if population_size else DEFAULT_POPULATION_SIZE # the number of individual\n self.mutation_rate = mutation_rate if mutation_rate else DEFAULT_MUTATION_RATE # the rate of mutation\n self.crossover_rate = crossover_rate if crossover_rate else DEFAULT_CROSSOVER_RATE\n self.population = [] # population that has individual\n self.temp = [] #temporary group that has individuals\n self.bestpopulation = [] # group that has the best individuals in each generation\n self.averagepopulation = [] # the average value of fitness in each generation\n self.number_of_train_data = DEFAULT_PREDICT_PATTERN_NUMBER\n self.fitness_dictionary = {}\n\n def adapt_rule(self,oil_data,freight_data,own_ship,time,rule_args):\n rule_integrate = copy.deepcopy(rule_args)\n result = [[False,0],[False,0],[False,0]]\n for which_action in range(DEFAULT_NUM_OF_ACTION_INTEGRATE):\n rule = rule_integrate[which_action]\n oil_start = convert2to10_in_list(rule[2])\n oil_end = convert2to10_in_list(rule[3])\n if oil_start > oil_end:\n oil_start, oil_end = oil_end, oil_start\n freight_start = convert2to10_in_list(rule[6])\n freight_end = convert2to10_in_list(rule[7])\n if freight_start > freight_end:\n freight_start, freight_end = freight_end, freight_start\n average_oil = 0\n if oil_start == oil_end:\n average_oil = oil_data[time - oil_start]['price']\n else:\n for month_oil in range(oil_start,oil_end + 1):\n if time - month_oil < 0:\n average_oil += OIL_PREV[time - month_oil]\n else:\n average_oil += oil_data[time - month_oil]['price']\n average_oil /= (oil_end - oil_start + 1)\n average_freight = 0\n if freight_start == freight_end:\n average_freight = freight_data[time - freight_start]['price']\n else:\n for month_freight in range(freight_start,freight_end + 1):\n if time - month_freight < 0:\n average_freight += FREIGHT_PREV[time - month_freight]\n else:\n average_freight += freight_data[time - month_freight]['price']\n average_freight /= (freight_end - freight_start + 1)\n flag = True\n list_1 = [0,4,8]\n list_2 = [1,5,9]\n list_3 = [OIL_PRICE_LIST,FREIGHT_RATE_LIST,OWN_SHIP_LIST]\n list_4 = [average_oil,average_freight,own_ship]\n for cond_1, cond_2, condition_type, data_compare in zip(list_1,list_2,list_3,list_4):\n lower = condition_type[convert2to10_in_list(rule[cond_1])]\n upper = condition_type[convert2to10_in_list(rule[cond_2])]\n if (lower < data_compare or lower == DO_NOT_CARE) and (data_compare < upper or upper == DO_NOT_CARE):\n pass\n else:\n flag = False\n if flag == True:\n result[which_action][0] = True\n result[which_action][1] = + 1\n return result\n \n def generateIndividual_with_wise(self):\n population = []\n for num in range(self.population_size):\n rule_random = []\n for trade in range(DEFAULT_NUM_OF_ACTION_INTEGRATE):\n rule_random.append([])\n for condition in range(DEFAULT_NUM_OF_CONDITION*2):\n rule_random[trade].append([])\n for bit in range(DEFAULT_NUM_OF_BIT):\n rule_random[trade][condition].append(random.randint(0,1))\n rule_random.append([0,0])\n rule_random[-1][0],rule_random[-1][1] = self.fitness_function(rule_random)\n rule_string = self.return_rule_str(rule_random)\n self.fitness_dictionary[rule_string] = copy.deepcopy([rule_random[-1][0],rule_random[-1][1]])\n population.append(copy.deepcopy(rule_random))\n return population\n\n def crossover(self,a_args,b_args):\n a = copy.deepcopy(a_args)\n b = copy.deepcopy(b_args)\n temp1 = []\n temp2 = []\n which_action = random.randint(0,DEFAULT_NUM_OF_ACTION_INTEGRATE-1)\n proportion = [1,1,1,1,1]\n rand = random.randint(0,sum(proportion)-1)\n crossover_block = 0\n for tryal in range(4):\n if rand < proportion[crossover_block]:\n pass\n else:\n rand -= proportion[crossover_block]\n crossover_block += 1\n crossover_block = crossover_block*2 + random.randint(0,1)\n for index in range(DEFAULT_NUM_OF_ACTION_INTEGRATE):\n if index == which_action:\n temp1.append([])\n temp2.append([])\n for condition in range(DEFAULT_NUM_OF_CONDITION*2):\n if condition == crossover_block:\n temp1[index].append([])\n temp2[index].append([])\n length = len(a[index][condition]) - 1\n crossover_point = random.randint(1,length-1)\n for former in range(0,crossover_point):\n temp1[index][condition].append(a[index][condition][former])\n temp2[index][condition].append(b[index][condition][former])\n for latter in range(crossover_point,len(a[index][condition])):\n temp1[index][condition].append(b[index][condition][latter])\n temp2[index][condition].append(a[index][condition][latter])\n else:\n temp1[index].append(a[index][condition])\n temp2[index].append(b[index][condition])\n else:\n temp1.append(a[index])\n temp2.append(b[index])\n temp1.append([0,0])\n temp2.append([0,0])\n return [temp1,temp2]\n\n def mutation(self,individual_args):\n individual = copy.deepcopy(individual_args)\n which_action = random.randint(0,DEFAULT_NUM_OF_ACTION_INTEGRATE-1)\n proportion = [1,1,1,1,1]\n rand = random.randint(0,sum(proportion)-1)\n mutation_block = 0\n for tryal in range(4):\n if rand < proportion[mutation_block]:\n pass\n else:\n rand -= proportion[mutation_block]\n mutation_block += 1\n mutation_block = mutation_block*2 + random.randint(0,1)\n length = len(individual[which_action][mutation_block]) - 1\n point = random.randint(0,length)\n if individual[which_action][mutation_block][point] == 0:\n individual[which_action][mutation_block][point] = 1\n else:\n individual[which_action][mutation_block][point] = 0\n return individual\n\n def fitness_function(self,rule_args):\n rule = copy.deepcopy(rule_args)\n Record = []\n for pattern in range(self.number_of_train_data):\n fitness = 0\n ship = Ship(TEU_SIZE,INITIAL_SPEED,ROUTE_DISTANCE)\n for year in range(DEFAULT_PREDICT_YEARS):\n cash_flow = 0\n if year >= PAYBACK_PERIOD and ship.exist_number + ship.order_number <= 0:\n break\n for month in range(0,12,TIME_STEP):\n current_oil_price = self.oil_price_data[pattern][year*12+month]['price']\n current_freight_rate_outward = self.freight_rate_outward_data[pattern][year*12+month]['price']\n current_freight_rate_homeward = self.freight_rate_homeward_data[pattern][year*12+month]['price']\n total_freight = ( current_freight_rate_outward * LOAD_FACTOR_ASIA_TO_EUROPE + current_freight_rate_homeward * LOAD_FACTOR_EUROPE_TO_ASIA)\n current_exchange = self.exchange_rate_data[pattern][year*12+month]['price']\n current_demand = self.demand_data[pattern][year*12+month]['price']\n current_supply = self.supply_data[pattern][year*12+month]['price']\n if year < PAYBACK_PERIOD:\n current_newbuilding = self.newbuilding[pattern][year*12+month]['price']\n current_secondhand = self.secondhand[pattern][year*12+month]['price']\n result = self.adapt_rule(self.oil_price_data[pattern],self.freight_rate_outward_data[pattern],ship.total_number+ship.order_number,year*12+month,rule)\n if result[0][0]:\n cash_flow += ship.buy_new_ship(current_newbuilding,result[0][1])\n if result[1][0]:\n cash_flow += ship.buy_secondhand_ship(current_secondhand,result[1][1])\n if result[2][0]:\n cash_flow += ship.sell_ship(current_secondhand,result[2][1])\n cash_flow += ship.calculate_income_per_time_step_month(current_oil_price,total_freight,current_demand,current_supply)\n cash_flow += ship.add_age()\n DISCOUNT = (1 + DISCOUNT_RATE) ** (year + 1)\n cash_flow *= self.exchange_rate_data[pattern][year*12+11]['price']\n fitness += cash_flow / DISCOUNT\n fitness /= HUNDRED_MILLION\n fitness /= SCALING\n Record.append(fitness)\n e, sigma = calc_statistics(Record)\n return [e,sigma]\n\n def process(self,rule_args,number):\n rule = copy.deepcopy(rule_args)\n e, sigma = self.fitness_function(rule)\n return [e,sigma,self.return_rule_str(rule),number]\n\n def wrapper_process(self,args):\n return self.process(*args)\n\n def selection(self,generation):\n #store last generation's best individual unchanged\n self.population.sort(key=lambda x:x[-1][0],reverse = True)\n #roulette selection and elite storing\n #store the best 5% individual\n self.temp.sort(key=lambda x:x[-1][0],reverse = True)\n elite_number = int(self.population_size * 0.05)\n start = 1 if generation != 0 else 0\n for ith_individual in range(start,elite_number+1):\n self.population[ith_individual] = copy.deepcopy(self.temp[ith_individual])\n min_fit = self.temp[-1][-1][0]\n random.shuffle(self.temp)\n ark = 0 # the number used to roulette in crossing\n probability = 0\n for jth_individual in range(len(self.temp)):\n probability = probability + self.temp[jth_individual][-1][0] + (0.1 - min_fit)#Translation\n roulette = 0\n for kth_individual in range(elite_number+1,self.population_size):\n roulette = random.randint(0,int(probability))\n while roulette > 0:\n roulette = roulette - (self.temp[ark][-1][0] + 0.1 - min_fit)\n ark = (ark + 1) % len(self.temp)\n self.population[kth_individual] = copy.deepcopy(self.temp[ark])\n\n def exchange_rule(self):\n for individual_index in range(len(self.temp)):\n for condition_block in range(DEFAULT_NUM_OF_ACTION_INTEGRATE):\n condition = self.temp[individual_index][condition_block]\n if OIL_PRICE_LIST[convert2to10_in_list(condition[0])] > OIL_PRICE_LIST[convert2to10_in_list(condition[1])]:\n condition[0],condition[1] = copy.deepcopy(condition[1]),copy.deepcopy(condition[0])\n if FREIGHT_RATE_LIST[convert2to10_in_list(condition[2])] > FREIGHT_RATE_LIST[convert2to10_in_list(condition[3])]:\n condition[2],condition[3] = copy.deepcopy(condition[3]),copy.deepcopy(condition[2])\n if EXCHANGE_RATE_LIST[convert2to10_in_list(condition[4])] > EXCHANGE_RATE_LIST[convert2to10_in_list(condition[5])]:\n condition[4],condition[5] = copy.deepcopy(condition[5]),copy.deepcopy(condition[4])\n if OWN_SHIP_LIST[convert2to10_in_list(condition[6])] > OWN_SHIP_LIST[convert2to10_in_list(condition[7])]:\n condition[6],condition[7] = copy.deepcopy(condition[7]),copy.deepcopy(condition[6])\n if FREIGHT_RATE_LIST[convert2to10_in_list(condition[8])] > FREIGHT_RATE_LIST[convert2to10_in_list(condition[9])]:\n condition[8],condition[9] = copy.deepcopy(condition[9]),copy.deepcopy(condition[8])\n\n def store_best_and_average(self):\n self.population.sort(key=lambda x:x[-1][0],reverse = True)\n self.bestpopulation.append(self.population[0])\n random.shuffle(self.population)\n total = 0\n for e in range(self.population_size):\n total += self.population[e][-1][0]\n self.averagepopulation.append(total/self.population_size)\n\n def depict_fitness(self,gene):\n x = range(0,len(self.bestpopulation))\n y = []\n z = []\n for i in range(len(self.bestpopulation)):\n y.append(self.bestpopulation[i][-1][0])\n z.append(self.averagepopulation[i])\n plt.plot(x, y, marker='o',label='best')\n plt.plot(x, z, marker='x',label='average')\n plt.title('Transition of fitness', fontsize = 20)\n plt.xlabel('generation', fontsize = 16)\n plt.ylabel('fitness value', fontsize = 16)\n plt.tick_params(labelsize=14)\n plt.grid(True)\n plt.legend(loc = 'lower right')\n save_dir = '../output/rule-discovered'\n plt.savefig(os.path.join(save_dir, 'integrate_fitness_{}.png'.format(gene)))\n plt.close()\n\n def depict_average_variance(self,gene=None):\n x = []\n y = []\n for i in range(self.population_size): \n x.append(self.population[i][-1][0])\n y.append(self.population[i][-1][1])\n plt.scatter(x,y)\n x_min = min(x)\n x_min = x_min*0.9 if x_min>0 else x_min*1.1\n plt.xlim(0,1)\n plt.ylim(0,1)\n plt.title(\"Rule Performance\")\n plt.xlabel(\"Expectation\")\n plt.ylabel(\"Variance\")\n plt.grid(True)\n save_dir = '../output/train/image'\n if gene is not None:\n plt.savefig(os.path.join(save_dir, 'Evaluation_initial.png'))\n else:\n plt.savefig(os.path.join(save_dir, 'Evaluation.png'))\n plt.close()\n\n def check_convergence(self,target,criteria):\n flag = True\n for index in range(1,criteria+1):\n if target[-index][-1] != target[-(index+1)][-1]:\n flag = False\n break\n return flag\n\n def return_rule_str(self,lists_args):\n lists = copy.deepcopy(lists_args)\n rule_string = ''\n for rule_type in range(DEFAULT_NUM_OF_ACTION_INTEGRATE):\n for condition_block in range(DEFAULT_NUM_OF_CONDITION*2):\n block = lists[rule_type][condition_block]\n for e in block:\n rule_string += str(e)\n return rule_string\n\n def execute_GA(self):\n time_record = [0]\n first = time.time()\n #randomly generating individual group\n #for p_size in range(self.population_size):\n # self.population.append(self.generateIndividual())\n self.population = copy.deepcopy(self.generateIndividual_with_wise())\n self.depict_average_variance(0)\n\n #genetic algorithm\n for gene in tqdm(range(self.generation)):\n #crossover\n self.temp = copy.deepcopy(self.population)\n random.shuffle(self.temp)\n for selected in range(0,self.population_size,2):\n if random.random() < self.crossover_rate:\n a,b = self.crossover(self.temp[selected],self.temp[selected+1])\n else:\n a,b = self.temp[selected],self.temp[selected+1]\n self.temp.append(copy.deepcopy(a))\n self.temp.append(copy.deepcopy(b))\n \n #mutation\n for individual_index in range(self.population_size*2):\n if random.random() < self.mutation_rate:\n self.temp[individual_index] = copy.deepcopy(self.mutation(self.temp[individual_index]))\n\n #rule check\n self.exchange_rule()\n \n #fitness calculation\n num_pool = multi.cpu_count()\n num_pool = int(num_pool*0.95)\n tutumimono = []\n for individual_index in range(self.population_size*2):\n rule_string = self.return_rule_str(self.temp[individual_index])\n if rule_string in self.fitness_dictionary:\n self.temp[individual_index][-1][0] = self.fitness_dictionary[rule_string][0]\n self.temp[individual_index][-1][1] = self.fitness_dictionary[rule_string][1]\n else:\n tutumimono.append([copy.deepcopy(self.temp[individual_index]),individual_index])\n #tutumimono = [[self.temp[individual_number], individual_number] for individual_number in range(self.population_size*2)]\n '''\n with Pool(num_pool) as pool:\n p = pool.map(self.wrapper_process, tutumimono)\n #for index in range(self.population_size*2):\n for i in range(len(p)):\n index = p[i][-1]\n self.temp[index][-1][0] = p[i][0]\n self.temp[index][-1][1] = p[i][1]\n self.fitness_dictionary[p[i][2]] = copy.deepcopy([p[i][0],p[i][1]])\n '''\n for index in range(self.population_size*2):\n rule_string = self.return_rule_str(self.temp[index])\n if rule_string in self.fitness_dictionary:\n pass\n else:\n e, sigma = self.fitness_function(self.temp[index])\n self.temp[index][-1][0] = e\n self.temp[index][-1][1] = sigma\n self.fitness_dictionary[rule_string] = [e,sigma]\n #'''\n #selection\n self.selection(gene)\n\n #store best and average individual\n self.store_best_and_average()\n #if gene > 1000 and self.check_convergence(self.bestpopulation,500):\n # break\n time_record.append(time.time()-first)\n\n if gene % 500 == 0:\n export_rules_integrate_csv(self.population,gene)\n export_dictionary(self.fitness_dictionary)\n self.depict_fitness(gene)\n\n x = range(self.generation+1)\n plt.plot(x,time_record)\n save_dir = '../output/train/image'\n plt.savefig(os.path.join(save_dir, 'computationi_time.png'))\n plt.close()\n print('exploranation number ',len(self.fitness_dictionary))\n #for index in range(len(self.population)):\n # self.population[index][-1][0],self.population[index][-1][1] = self.fitness_function(self.population[index])\n self.depict_fitness(gene)\n self.depict_average_variance()\n self.population.sort(key=lambda x:x[-1][0],reverse = True)\n print(self.population[0])\n return self.population\n\n\ndef main():\n oil_data,freight_outward_data,freight_return_data,exchange_data,demand_data,supply_data,newbuilding_data,secondhand_data = load_generated_sinario()\n ga = GA_Extended(oil_data,freight_outward_data,freight_return_data,exchange_data,demand_data,supply_data,newbuilding_data,secondhand_data)\n start = time.time()\n p = ga.execute_GA()\n print(time.time()-start)\n export_rules_integrate_csv(p)\n\nif __name__ == \"__main__\":\n main()\n slack = slackweb.Slack(url=\"############\")\n slack.notify(text=\"program end!!!!!!!!!\")\n","sub_path":"models/ga_extended.py","file_name":"ga_extended.py","file_ext":"py","file_size_in_byte":21213,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"555781917","text":"\"\"\"Model definition for CNN sentiment training.\"\"\"\n\nimport os\nimport numpy as np\n\nimport tensorflow as tf\nfrom tensorflow import keras\nfrom tensorflow.keras import layers\nfrom tensorflow.keras.initializers import Constant\n\n\ndef keras_model_fn(_, config, args):\n \"\"\"Creates a CNN model for sentiment modeling.\"\"\"\n embedding_matrix = np.zeros((\n config[\"embeddings_dictionary_size\"],\n config[\"embeddings_vector_size\"]))\n\n with tf.io.gfile.GFile(config[\"embeddings_path\"], \"r\") as file:\n input_data = file.read()\n split = input_data.split(\"\\n\")\n\n for index, _ in enumerate(split):\n data = np.asarray(split[index].split()[1:], dtype='float32')\n if len(data) == config[\"embeddings_vector_size\"]:\n embedding_matrix[index + 2] = data\n else:\n padded = np.zeros((config[\"embeddings_vector_size\"]), 'float32')\n padded[:len(data)] = data\n embedding_matrix[index + 2] = padded\n\n cnn_model = keras.Sequential()\n cnn_model.add(layers.Embedding(\n input_dim=config[\"embeddings_dictionary_size\"],\n input_length=config[\"padding_size\"],\n embeddings_initializer=Constant(embedding_matrix),\n output_dim=config[\"embeddings_vector_size\"],\n trainable=True))\n cnn_filters = [\n min(1000,\n max(8, int(\n args.first_filter_size * args.cnn_layer_sizes_scale_factor**i)))\n for i in range(args.num_cnn_layers)\n ]\n for i in range(args.num_cnn_layers):\n cnn_model.add(layers.Conv1D(\n filters=cnn_filters[i],\n kernel_size=2,\n strides=1,\n padding=\"valid\",\n activation=\"relu\"))\n cnn_model.add(layers.GlobalMaxPool1D())\n dense_layers = [\n min(1024,\n max(8, int(\n args.first_layer_size * args.dense_layer_sizes_scale_factor**i)))\n for i in range(args.num_dense_layers)\n ]\n for i in range(args.num_dense_layers):\n cnn_model.add(layers.Dense(dense_layers[i], activation=\"relu\"))\n cnn_model.add(layers.Dense(1, activation=\"sigmoid\"))\n\n cnn_model.compile(\n optimizer=\"adam\",\n loss=\"binary_crossentropy\",\n metrics=[\"accuracy\"])\n return cnn_model\n\n\ndef save_model(model, output):\n \"\"\"Saves models in SaveModel format with signature to support serving.\"\"\"\n tf.saved_model.save(model, os.path.join(output, \"1\"))\n print(\"Model successfully saved at: {}\".format(output))\n","sub_path":"gcp_model_training/sentiment_model_cnn.py","file_name":"sentiment_model_cnn.py","file_ext":"py","file_size_in_byte":2469,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"51464278","text":"import argparse\nimport math\nimport os\nimport sys\nfrom linecache import getline\nfrom multiprocessing import Process\n\nimport Constants as c\n\n#In: -i in_decoded_dir -o out_dir [-t time window] [-s slide_interval] [-p num_processes]\n#Out: tab-delim txt w/ header: frame_num\\tts\\tts_delta\\tframe_len\\tip_src\\tip_dst\\thost\n\n\n#is_error is either 0 or 1\ndef print_usage(is_error): \n print(c.SLIDE_SPLIT_USAGE, file=sys.stderr) if is_error else print(c.SLIDE_SPLIT_USAGE)\n exit(is_error)\n\n\ndef get_num(str_num, description):\n bad_num = False\n try:\n num = int(str_num)\n if num < 0:\n bad_num = True\n except ValueError:\n bad_num = True\n\n if bad_num:\n print(c.NON_POS % (description, str_num), file=sys.stderr)\n num = -1\n\n return num\n\n\ndef write_file(start_idx, end_idx, in_file, out_file):\n lines = \"frame_num\\tts\\tts_delta\\tframe_len\\tip_src\\tip_dst\\thost\\n\"\n for i in range(start_idx, end_idx):\n lines += getline(in_file, i)\n with open(out_file, \"w\") as f:\n f.write(lines)\n\n\ndef run(pid, files, src, dest, slide_int, time_window):\n for fpath in files:\n with open(fpath, \"r\") as f:\n times = []\n for l in f:\n try:\n times.append(float(l.split(\"\\t\")[1]))\n except (IndexError, ValueError) as e:\n pass\n\n if len(times) == 0:\n print(c.NO_VAL_TS % fpath, file=sys.stderr)\n continue\n\n dest_file = os.path.join(dest, fpath.replace(src, \"\", 1)[:-4] + \"_part_0.txt\")\n if os.path.isfile(dest_file):\n print(\"P%s: %s exists, skipping %s\" % (pid, dest_file, fpath))\n continue\n\n print(\"P%s: IN: %s\" % (pid, fpath)) \n start_int = times[0]\n end_int = start_int + time_window\n last_poss_start = math.ceil((times[len(times) - 1] - time_window) / slide_int) * slide_int\n start_idxes = [0]\n idx = 0\n num = 0\n last_bucket = False\n for t in times:\n while t - start_int >= slide_int and not last_bucket:\n if t > last_poss_start and last_poss_start - start_int < slide_int:\n last_bucket = True\n\n start_int += slide_int\n start_idxes.append(idx)\n\n num_pop = 0\n for i in start_idxes:\n if t > end_int:\n dest_file = dest_file[:dest_file.rfind(\"_\")] + \"_%d.txt\" % num\n print(\"P%s: OUT: %s\" % (pid, dest_file))\n if not os.path.isdir(os.path.dirname(dest_file)):\n os.system(\"mkdir -pv %s\" % os.path.dirname(dest_file))\n #+2: +1 for getline is 1-based index, +1 for skipping header row\n write_file(i + 2, idx + 2, fpath, dest_file)\n num_pop += 1\n num += 1\n end_int += slide_int\n\n [ start_idxes.pop(0) for _ in range(num_pop) ]\n idx += 1\n\n while len(start_idxes) > 0:\n dest_file = dest_file[:dest_file.rfind(\"_\")] + \"_%d.txt\" % num\n print(\"P%s: OUT: %s\" % (pid, dest_file))\n if not os.path.isdir(os.path.dirname(dest_file)):\n os.system(\"mkdir -pv %s\" % os.path.dirname(dest_file))\n write_file(start_idxes[0] + 2, idx + 2, fpath, dest_file)\n start_idxes.pop(0)\n num += 1\n\n\ndef main():\n #parse arguments\n parser = argparse.ArgumentParser(usage=c.SLIDE_SPLIT_USAGE, add_help=False)\n parser.add_argument(\"-i\", dest=\"dec_dir\", default=\"\")\n parser.add_argument(\"-o\", dest=\"dest_dir\", default=\"\")\n parser.add_argument(\"-t\", dest=\"time_window\", default=\"30\")\n parser.add_argument(\"-s\", dest=\"slide_int\", default=\"5\")\n parser.add_argument(\"-p\", dest=\"num_proc\", default=\"1\")\n parser.add_argument(\"-h\", dest=\"help\", action=\"store_true\", default=False)\n args = parser.parse_args()\n\n if args.help:\n print_usage(0)\n\n print(\"Running %s...\" % c.PATH)\n\n #error checking\n errors = False\n #check -i in source\n if args.dec_dir == \"\":\n errors = True\n print(c.NO_SRC_DIR, file=sys.stderr)\n elif not os.path.isdir(args.dec_dir):\n errors = True\n print(c.INVAL % (\"Source directory\", args.dec_dir, \"directory\"), file=sys.stderr)\n else:\n if not os.access(args.dec_dir, os.R_OK):\n errors = True\n print(c.NO_PERM % (\"source directory\", args.dec_dir, \"read\"), file=sys.stderr)\n if not os.access(args.dec_dir, os.X_OK):\n errors = True\n print(c.NO_PERM % (\"source directory\", args.dec_dir, \"execute\"), file=sys.stderr)\n\n #check -o out destination\n if args.dest_dir == \"\":\n errors = True\n print(c.NO_DEST_DIR, file=sys.stderr)\n elif os.path.isdir(args.dest_dir):\n if not os.access(args.dest_dir, os.W_OK):\n errors = True\n print(c.NO_PERM % (\"destination directory\", args.dest_dir, \"write\"), file=sys.stderr)\n if not os.access(args.dest_dir, os.X_OK):\n errors = True\n print(c.NO_PERM % (\"destination directory\", args.dest_dir, \"execute\"), file=sys.stderr)\n\n #check -t time window\n time_window = get_num(args.time_window, \"time window\")\n if time_window == -1:\n errors = True\n\n #check -s slide interval\n slide_int = get_num(args.slide_int, \"slide interval\")\n if slide_int == -1:\n errors = True\n\n if slide_int > time_window:\n errors = True\n print(c.INT_GT_TIME_WIN % (slide_int, time_window), file=sys.stderr)\n\n #check -p number processes\n num_proc = get_num(args.num_proc, \"number of processes\")\n if num_proc == -1:\n errors = True\n\n if errors:\n print_usage(1)\n #end error checking\n\n if not os.path.isdir(args.dest_dir):\n os.system(\"mkdir -pv %s\" % args.dest_dir)\n\n files = [[] for _ in range(num_proc)]\n\n index = 0\n for root, dirs, fs in os.walk(args.dec_dir):\n for fname in fs:\n if fname.endswith(\".txt\"):\n files[index].append(os.path.join(root, fname))\n index += 1\n if index >= num_proc:\n index = 0\n else:\n print(c.WRONG_EXT % (\"Decoded file\", \"text (.txt)\", os.path.join(root, fname)),\n file=sys.stderr)\n\n procs = []\n for pid, files in enumerate(files):\n p = Process(target=run, args=(pid, files, args.dec_dir, args.dest_dir, slide_int, time_window))\n procs.append(p)\n p.start()\n\n\nif __name__ == \"__main__\":\n main()\n\n","sub_path":"model/src/s8_slide_split.py","file_name":"s8_slide_split.py","file_ext":"py","file_size_in_byte":6646,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"297140283","text":"# $Id: ScrollBar.py,v 1.46.2.4 2007/01/27 11:10:37 marcusva Exp $\n#\n# Copyright (c) 2004-2007, Marcus von Appen\n# All rights reserved.\n# Redistribution and use in source and binary forms, with or without\n# modification, are permitted provided that the following conditions are met:\n#\n# * Redistributions of source code must retain the above copyright notice,\n# this list of conditions and the following disclaimer.\n# * Redistributions in binary form must reproduce the above copyright notice,\n# this list of conditions and the following disclaimer in the documentation\n# and/or other materials provided with the distribution.\n#\n# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\"\n# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE\n# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE\n# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE\n# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR\n# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF\n# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS\n# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN\n# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)\n# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE\n# POSSIBILITY OF SUCH DAMAGE.\n\n\"\"\"A widget, which allows scrolling using buttons and a slider.\"\"\"\n\nfrom pygame import K_KP_PLUS, K_PLUS, K_RIGHT, K_DOWN, K_KP_MINUS, K_MINUS\nfrom pygame import K_LEFT, K_UP, K_PAGEUP, K_PAGEDOWN, K_HOME, K_END, Rect\nfrom .Range import Range\nfrom .Constants import *\nfrom .StyleInformation import StyleInformation\nfrom . import base\n\n# Timer value for the button press delay.\n_TIMER = 25\n\nclass ScrollBar (Range):\n \"\"\"ScrollBar () -> ScrollBar\n\n An abstract widget class, which is suitable for scrolling.\n\n The ScrollBar widget works much the same like a Scale widget except\n that it supports buttons for adjusting the value and that its\n minimum value always is 0. It is suitable for widgets which need\n scrolling ability and a scrolling logic.\n\n Inheriting widgets have to implement the _get_value_from_coords()\n and _get_coords_from_value() methods, which calculate the value of\n the ScrollBar using a pair of coordinates and vice versa. Example\n implementations can be found in the HScrollBar and VScrollBar widget\n classes. They also need to implement the _get_button_coords()\n method, which has to return a tuple of the both button coordinates\n [(x, y, width, height)].\n \n Default action (invoked by activate()):\n Give the ScrollBar the input focus.\n \n Mnemonic action (invoked by activate_mnemonic()):\n None\n\n Signals:\n SIG_MOUSEDOWN - Invoked, when a mouse button is pressed on the\n ScrollBar.\n SIG_MOUSEUP - Invoked, when a mouse buttor is released on the\n ScrollBar.\n SIG_MOUSEMOVE - Invoked, when the mouse moves over the ScrollBar.\n\n Attributes:\n button_dec - Indicates, if the decrease button is pressed.\n button_inc - Indicates, if the increase button is pressed.\n \"\"\"\n def __init__ (self):\n Range.__init__ (self, 0, 1, 1)\n\n # Signals.\n self._signals[SIG_MOUSEDOWN] = []\n self._signals[SIG_MOUSEMOVE] = []\n self._signals[SIG_MOUSEUP] = []\n self._signals[SIG_KEYDOWN] = None # Dummy for keyboard activation.\n self._signals[SIG_TICK] = None # Dummy for automatic scrolling.\n \n # Internal state handlers for the Events. Those need to be known by\n # the inheritors.\n self._buttondec = False\n self._buttoninc = False\n\n self._timer = _TIMER\n self._click = False\n\n def activate (self):\n \"\"\"S.activate () -> None\n\n Activates the ScrollBar default action.\n\n Activates the ScrollBar default action. This usually means giving\n the ScrollBar the input focus.\n \"\"\"\n if not self.sensitive:\n return\n self.focus = True\n \n def _get_button_coords (self, area):\n \"\"\"S._get_button_coords (...) -> tuple\n\n Gets a tuple with the coordinates of the in- and decrease buttons.\n \n This method has to be implemented by inherited widgets.\n \"\"\"\n raise NotImplementedError\n\n def _get_coords_from_value (self):\n \"\"\"S._get_coords_from_value () -> float\n\n Calculates the slider coordinates for the ScrollBar.\n \n This method has to be implemented by inherited widgets.\n \"\"\"\n raise NotImplementedError\n\n def _get_value_from_coords (self, area, coords):\n \"\"\"S._get_value_from_coords (...) -> float\n\n Calculates the slider coordinates for the ScrollBar.\n \n This method has to be implemented by inherited widgets.\n \"\"\"\n raise NotImplementedError\n\n def _get_slider_size (self):\n \"\"\"S._get_slider_size (...) -> int\n\n Calculates the size of the slider knob.\n \n This method has to be implemented by inherited widgets.\n \"\"\"\n raise NotImplementedError\n \n def _check_collision (self, pos, rect):\n \"\"\"S._check_collirion (...) -> bool\n\n Checks the collision of the given position with the passed rect.\n \"\"\"\n # Rect: (x, y, width, height), pos: (x, y).\n return (pos[0] >= rect[0]) and (pos[0] <= (rect[2] + rect[0])) and \\\n (pos[1] >= rect[1]) and (pos[1] <= (rect[3] + rect[1]))\n \n def set_minimum (self, minimum):\n \"\"\"S.set_minimum (...) -> Exception\n\n This method does not have any use.\n \"\"\"\n pass\n\n def notify (self, event):\n \"\"\"S.notify (...) -> None\n\n Notifies the ScrollBar about an event.\n \"\"\"\n if not self.sensitive:\n return\n\n if event.signal in SIGNALS_MOUSE:\n eventarea = self.rect_to_client ()\n collision = eventarea.collidepoint (event.data.pos)\n if event.signal == SIG_MOUSEDOWN and collision:\n self.focus = True\n # Act only on left clicks or scrollwheel Events.\n if event.data.button == 1:\n self.state = STATE_ACTIVE\n self.run_signal_handlers (SIG_MOUSEDOWN, event.data)\n\n if event.data.button == 1:\n buttons = self._get_button_coords (eventarea)\n if self._check_collision (event.data.pos, buttons[0]):\n self._buttondec = True\n self._buttoninc = False\n self._click = False\n self.decrease ()\n elif self._check_collision (event.data.pos, buttons[1]):\n self._buttoninc = True\n self._buttondec = False\n self._click = False\n self.increase ()\n else:\n self._click = True\n self._buttondec = False\n self._buttoninc = False\n val = self._get_value_from_coords (eventarea,\n event.data.pos)\n if val != self.value:\n self.value = val\n # Mouse wheel.\n elif event.data.button == 4:\n self.decrease ()\n elif event.data.button == 5:\n self.increase ()\n event.handled = True\n\n elif event.signal == SIG_MOUSEMOVE:\n dirty = False\n if collision:\n self.focus = True\n if self.state == STATE_NORMAL:\n self.state = STATE_ENTERED\n self.run_signal_handlers (SIG_MOUSEMOVE, event.data)\n\n buttons = self._get_button_coords (eventarea)\n if not self._check_collision (event.data.pos, buttons[0]) \\\n and self._buttondec:\n self._buttondec = False\n dirty = True\n if not self._check_collision (event.data.pos, buttons[1]) \\\n and self._buttoninc:\n self._buttoninc = False\n dirty = True\n if self._click:\n val = self._get_value_from_coords (eventarea,\n event.data.pos)\n if val != self.value:\n self.value = val\n dirty = False\n self.dirty = dirty\n event.handled = True\n\n elif self.state == STATE_ENTERED:\n self.state = STATE_NORMAL\n\n elif event.signal == SIG_MOUSEUP:\n if self._click or self._buttoninc or self._buttondec:\n self._buttondec = False\n self._buttoninc = False\n self._click = False\n\n if collision:\n if event.data.button == 1:\n if self.state == STATE_ACTIVE:\n self.state = STATE_ENTERED\n self.run_signal_handlers (SIG_MOUSEUP, event.data)\n event.handled = True\n else:\n self.state = STATE_NORMAL\n # Reset timer\n self._timer = _TIMER\n\n # The user holds the mouse clicked over one button.\n elif (self._buttondec or self._buttoninc) and \\\n (event.signal == SIG_TICK):\n # Wait half a second before starting to in/decrease.\n if self._timer > 0:\n self._timer -= 1\n else:\n if self._buttondec:\n self.decrease ()\n elif self._buttoninc:\n self.increase ()\n\n # Keyboard activation.\n elif (event.signal == SIG_KEYDOWN) and self.focus:\n if event.data.key in (K_KP_PLUS, K_PLUS, K_RIGHT, K_DOWN):\n self.increase ()\n event.handled = True\n elif event.data.key in (K_KP_MINUS, K_MINUS, K_LEFT, K_UP):\n self.decrease ()\n event.handled = True\n elif event.data.key == K_PAGEUP:\n val = self.value - 10 * self.step\n if val > self.minimum:\n self.value = val\n else:\n self.value = self.minimum\n event.handled = True\n elif event.data.key == K_PAGEDOWN:\n val = self.value + 10 * self.step\n if val < self.maximum:\n self.value = val\n else:\n self.value = self.maximum\n event.handled = True\n elif event.data.key == K_END:\n self.value = self.maximum\n event.handled = True\n elif event.data.key == K_HOME:\n self.value = self.minimum\n event.handled = True\n\n Range.notify (self, event)\n\n button_dec = property (lambda self: self._buttondec,\n doc = \"\"\"Indicates, whether the decrease\n button is pressed.\"\"\")\n button_inc = property (lambda self: self._buttoninc,\n doc = \"\"\"Indicates, whether the increase\n button is pressed.\"\"\")\n \nclass HScrollBar (ScrollBar):\n \"\"\"HScrollBar (width, scroll) -> HScrollBar\n\n A horizontal ScrollBar widget.\n\n A ScrollBar widget with a horizontal orientation. By default, its\n height is the sum of the button height (HSCROLLBAR_BUTTON_SIZE) and\n the border drawn around it (2 * SCROLLBAR_BORDER) and has the passed\n width. The scrolling area is the passed scroll value minus the width\n of the ScrollBar.\n\n Thus, if the area to scroll is 200 pixels wide and the ScrollBar is\n about 100 pixels long, the ScrollBar its value range will go from 0\n to 100 (maximum = scroll - width). If the ScrollBar is longer than\n the area to scroll (scroll < width), then the value range will be 0.\n\n Note: The minimum size of the scrollbar is at least twice its\n size[1] parameter. This means, that it can display the both\n scrolling buttons next to each other. This will override the passed\n width value in the constructor, if necessary.\n \"\"\"\n def __init__ (self, width, scroll):\n ScrollBar.__init__ (self)\n # Minimum size for the two scrolling buttons next to each other\n border = base.GlobalStyle.get_border_size \\\n (self.__class__, self.style,\n StyleInformation.get (\"SCROLLBAR_BORDER\")) * 2\n height = StyleInformation.get (\"HSCROLLBAR_BUTTON_SIZE\")[1] + border\n if width < 2 * height:\n width = 2 * height\n\n self.lock ()\n self.minsize = (width, height) # Default size.\n self.maximum = scroll\n self.unlock ()\n\n def set_maximum (self, maximum):\n \"\"\"H.set_maximum (...) -> None\n\n Sets the maximum value to scroll.\n\n The passed maximum value differs from maximum value of the\n slider. The HScrollBar also subtracts its own height from the\n scrolling maximum, so that the real maximum of its value range\n can be expressed in the formula:\n\n real_maximum = maximum - self.minsize[1]\n\n That means, that if the HScrollBar is 100 pixels high and the\n passed maximum value is 200, the scrolling range of the\n HScrollBar will go from 0 to 100 (100 + size = 200).\n\n Raises a ValueError, if the passed argument is smaller than\n the first element of the ScrollBar its size.\n \"\"\"\n if maximum < self.minsize[0]:\n raise ValueError (\"maximum must be greater than or equal to %d\"\n % self.minsize[0])\n ScrollBar.set_maximum (self, maximum - self.minsize[0])\n self.dirty = True\n\n def _get_button_coords (self, area):\n \"\"\"H._get_button_coords (...) -> tuple\n\n Gets a tuple with the coordinates of the in- and decrease buttons.\n \"\"\"\n border = base.GlobalStyle.get_border_size \\\n (self.__class__, self.style,\n StyleInformation.get (\"SCROLLBAR_BORDER\"))\n # Respect the set shadow for the ScrollBar.\n button1 = (area.left + border, area.top + border,\n area.height - 2 * border, area.height - 2 * border)\n button2 = (area.left + area.width - area.height - border,\n area.top + border, area.height - 2 * border,\n area.height - 2 * border)\n return (button1, button2)\n\n def _get_slider_size (self):\n \"\"\"H._get_slider_size () -> int\n\n Calculates the size of the slider knob.\n \"\"\"\n border = base.GlobalStyle.get_border_size \\\n (self.__class__, self.style,\n StyleInformation.get (\"SCROLLBAR_BORDER\"))\n \n # Minimum slider size, if the scrollbar is big enough.\n minsize = 10\n fullsize = self.size[0] - 2 * self.size[1]\n if fullsize == 0:\n # If only the both scrolling buttons can be displayed, we will\n # completely skip the slider.\n return 0\n\n # Full size.\n fullsize += 2 * border\n slider_width = fullsize\n if self.maximum != 0:\n slider_width = fullsize / (float (self.maximum) + fullsize) * \\\n fullsize\n if slider_width < minsize:\n slider_width = minsize\n return int (slider_width)\n \n def _get_coords_from_value (self):\n \"\"\"H._get_coords_from_value () -> int\n\n Calculates the slider coordinates for the HScrollBar.\n \"\"\"\n border = base.GlobalStyle.get_border_size \\\n (self.__class__, self.style,\n StyleInformation.get (\"SCROLLBAR_BORDER\"))\n\n val = 0\n if self.maximum > 0:\n slider = self._get_slider_size ()\n # Start offset for scrolling - this is the height\n # (button + 2 * border) - border plus the half of the\n # slider.\n sl_x = self.minsize[1] - border + float (slider) / 2\n\n # Valid sliding range.\n slide = self.minsize[0] - 2 * sl_x\n step = self.maximum / float (slide)\n val = self.value / step + sl_x\n return val\n return self.size[0] / 2\n \n def _get_value_from_coords (self, area, coords):\n \"\"\"H._get_value_from_coords (...) -> float\n\n Calculates the slider coordinates for the HScrollBar.\n \"\"\"\n border = base.GlobalStyle.get_border_size \\\n (self.__class__, self.style,\n StyleInformation.get (\"SCROLLBAR_BORDER\"))\n\n val = 0\n if self.maximum > 0:\n slider = self._get_slider_size ()\n sl_x = self.minsize[1] - border + float (slider) / 2\n slide = self.minsize[0] - 2 * sl_x\n n = coords[0] - area.left - sl_x\n step = self.maximum / float (slide)\n val = n * step\n if val > self.maximum:\n val = self.maximum\n elif val < 0:\n val = 0\n return val\n \n def draw_bg (self):\n \"\"\"H.draw_bg () -> Surface\n\n Draws the HScrollBar background surface and returns it.\n\n Creates the visible surface of the HScrollBar and returns it to\n the caller.\n \"\"\"\n return base.GlobalStyle.engine.draw_scrollbar (self,\n ORIENTATION_HORIZONTAL)\n\n def draw (self):\n \"\"\"H.draw () -> None\n\n Draws the HScrollBar surface and places its Buttons and slider on it.\n \"\"\"\n ScrollBar.draw (self)\n\n cls = self.__class__\n style = base.GlobalStyle\n st = self.style or style.get_style (cls)\n rect = self.image.get_rect ()\n draw_rect = style.engine.draw_rect\n draw_border = style.engine.draw_border\n draw_arrow = style.engine.draw_arrow\n\n # Create both buttons.\n border = style.get_border_size \\\n (cls, st, StyleInformation.get (\"SCROLLBAR_BORDER\"))\n button_type = StyleInformation.get (\"SCROLLBAR_BUTTON_BORDER\")\n\n width_button = rect.height - 2 * border\n\n # We use a temporary state here, so that just the buttons will\n # have the typical sunken effect.\n tmp_state = self.state\n if self.state == STATE_ACTIVE:\n tmp_state = STATE_NORMAL\n\n # First button.\n state_button = tmp_state\n if self.button_dec:\n state_button = STATE_ACTIVE\n button1 = draw_rect (width_button, width_button, state_button, cls, st)\n draw_border (button1, state_button, cls, st, button_type)\n rect_button1 = button1.get_rect ()\n\n # Draw the arrow.\n draw_arrow (button1, ARROW_LEFT, state_button, cls, st)\n rect_button1.x = border\n rect_button1.centery = rect.centery\n self.image.blit (button1, rect_button1)\n\n # Second button\n state_button = tmp_state\n if self.button_inc:\n state_button = STATE_ACTIVE\n \n button2 = draw_rect (width_button, width_button, state_button, cls, st)\n draw_border (button2, state_button, cls, st, button_type)\n rect_button2 = button2.get_rect ()\n\n # Draw the arrow.\n draw_arrow (button2, ARROW_RIGHT, state_button, cls, st)\n rect_button2.x = rect.width - width_button - border\n rect_button2.centery = rect.centery\n self.image.blit (button2, rect_button2)\n\n # Create the slider.\n slider_size = self._get_slider_size ()\n if slider_size > 0:\n sl = style.engine.draw_slider (slider_size, width_button,\n tmp_state, cls, st)\n r = sl.get_rect ()\n r.centerx = self._get_coords_from_value ()\n r.centery = rect.centery\n self.image.blit (sl, r)\n\nclass VScrollBar (ScrollBar):\n \"\"\"VScrollBar (height, scroll) -> VScrollBar\n\n A vertical ScrollBar widget.\n\n A ScrollBar widget with a vertical orientation. By default, its\n width is the sum of the button width (VSCROLLBAR_BUTTON_SIZE) and\n the border drawn around it (2 * SCROLLBAR_BORDER) and has the passed\n height. The scrolling area is the passed scroll value minus the\n height of the ScrollBar.\n\n Thus, if the area to scroll is 200 pixels high and the ScrollBar is\n about 100 pixels high, the ScrollBar its value range will go from 0\n to 100 (maximum = scroll - height). If the ScrollBar is longer than\n the area to scroll (scroll < height), then the value range will be 0.\n\n Note: The minimum size of the scrollbar is at least twice its\n size[0] parameter. This means, that it can display the both\n scrolling buttons next to each other. This will override the passed\n width value in the constructor, if necessary.\n \"\"\"\n def __init__ (self, height, scroll):\n ScrollBar.__init__ (self)\n # Minimum size for the two scrolling buttons next to each other.\n border = base.GlobalStyle.get_border_size \\\n (self.__class__, self.style,\n StyleInformation.get (\"SCROLLBAR_BORDER\")) * 2\n \n width = StyleInformation.get (\"VSCROLLBAR_BUTTON_SIZE\")[0] + border\n if height < 2 * width:\n height = 2 * width\n\n self.lock ()\n self.minsize = (width, height) # Default size.\n self.maximum = scroll\n self.unlock ()\n\n def set_maximum (self, maximum):\n \"\"\"V.set_maximum (...) -> None\n\n Sets the maximum value to scroll.\n\n The passed maximum value differs from maximum value of the\n slider. The VScrollBar also subtracts its own width from the\n scrolling maximum, so that the real maximum of its value range\n can be expressed in the formula:\n\n real_maximum = maximum - self.minsize[0]\n\n That means, that if the VScrollBar is 100 pixels long and the\n passed maximum value is 200, the scrolling range of the\n VScrollBar will go from 0 to 100 (100 + size = 200).\n\n Raises a ValueError, if the passed argument is smaller than\n the second element of the ScrollBar its size.\n \"\"\"\n if maximum < self.minsize[1]:\n raise ValueError (\"maximum must be greater than or equal to %d\"\n % self.minsize[1])\n ScrollBar.set_maximum (self, maximum - self.minsize[1])\n self.dirty = True\n\n def _get_button_coords (self, area):\n \"\"\"V._get_button_coords (...) -> tuple\n\n Gets a tuple with the coordinates of the in- and decrease buttons.\n \"\"\"\n border = base.GlobalStyle.get_border_size \\\n (self.__class__, self.style,\n StyleInformation.get (\"SCROLLBAR_BORDER\"))\n\n # Respect the set shadow for the ScrollBar.\n button1 = (area.left + border, area.top + border,\n area.width - 2 * border, area.width - 2 * border)\n button2 = (area.left + border,\n area.top + area.height - area.width - border,\n area.width - 2 * border, area.width - border)\n return (button1, button2)\n\n def _get_slider_size (self):\n \"\"\"V._get_slider_size () -> int\n\n Calculates the size of the slider knob.\n \"\"\"\n # Minimum slider size.\n minsize = 10\n if (self.size[1] - 2 * self.size[0]) == 0:\n # If only the both scrolling buttons can be displayed, we will\n # completely skip the slider.\n return 0\n \n # Full size.\n fullsize = self.size[1] - 2 * self.size[0]\n slider_height = fullsize\n if self.maximum != 0:\n slider_height = fullsize / (float (self.maximum) + fullsize) * \\\n fullsize\n if slider_height < minsize:\n slider_height = minsize\n return int (slider_height)\n \n def _get_coords_from_value (self):\n \"\"\"V._get_coords_from_value () -> int\n\n Calculates the slider coordinates for the VScrollBar.\n \"\"\"\n border = base.GlobalStyle.get_border_size \\\n (self.__class__, self.style,\n StyleInformation.get (\"SCROLLBAR_BORDER\"))\n \n val = 0\n if self.maximum > 0:\n slider = self._get_slider_size ()\n sl_y = self.minsize[0] - border + float (slider) / 2\n slide = self.minsize[1] - 2 * sl_y\n step = self.maximum / float (slide)\n val = self.value / step + sl_y\n return val\n return self.size[1] / 2\n \n def _get_value_from_coords (self, area, coords):\n \"\"\"V._get_value_from_coords (...) -> float\n\n Calculates the slider coordinates for the VScrollBar.\n \"\"\"\n border = base.GlobalStyle.get_border_size \\\n (self.__class__, self.style,\n StyleInformation.get (\"SCROLLBAR_BORDER\"))\n\n val = 0\n if self.maximum > 0:\n slider = self._get_slider_size ()\n\n # Start offset for scrolling - this is the width\n # (button + 2 * border) - border plus the half of the\n # slider.\n sl_y = self.minsize[0] - border + float (slider) / 2\n\n # Valid sliding range.\n slide = self.minsize[1] - 2 * sl_y\n \n n = coords[1] - area.top - sl_y\n step = self.maximum / float (slide)\n val = n * step\n if val > self.maximum:\n val = self.maximum\n elif val < 0:\n val = 0\n return val\n \n def draw_bg (self):\n \"\"\"V.draw_bg (...) -> Surface\n\n Draws the VScrollBar background surface and returns it.\n\n Creates the visible surface of the VScrollBar and returns it to\n the caller.\n \"\"\"\n return base.GlobalStyle.engine.draw_scrollbar (self,\n ORIENTATION_VERTICAL)\n\n def draw (self):\n \"\"\"V.draw () -> None\n\n Draws the VScrollBar surface and places its Buttons and slider on it.\n \"\"\"\n ScrollBar.draw (self)\n cls = self.__class__\n style = base.GlobalStyle\n st = self.style or style.get_style (cls)\n rect = self.image.get_rect ()\n draw_rect = style.engine.draw_rect\n draw_border = style.engine.draw_border\n draw_arrow = style.engine.draw_arrow\n \n # Create both buttons.\n border = style.get_border_size \\\n (cls, st, StyleInformation.get (\"SCROLLBAR_BORDER\"))\n button_type = StyleInformation.get (\"SCROLLBAR_BUTTON_BORDER\")\n \n width_button = rect.width - 2 * border\n\n # We use a temporary state here, so that just the buttons will\n # have the typical sunken effect.\n tmp_state = self.state\n if self.state == STATE_ACTIVE:\n tmp_state = STATE_NORMAL\n\n # First button.\n state_button = tmp_state\n if self.button_dec:\n state_button = STATE_ACTIVE\n button1 = draw_rect (width_button, width_button, state_button, cls, st)\n draw_border (button1, state_button, cls, st, button_type)\n rect_button1 = button1.get_rect ()\n\n # Draw the arrow.\n draw_arrow (button1, ARROW_UP, state_button, cls, st)\n rect_button1.y = border\n rect_button1.centerx = rect.centerx\n self.image.blit (button1, rect_button1)\n\n # Second button\n state_button = tmp_state\n if self.button_inc:\n state_button = STATE_ACTIVE\n \n button2 = draw_rect (width_button, width_button, state_button, cls, st)\n draw_border (button2, state_button, cls, st, button_type)\n rect_button2 = button2.get_rect ()\n\n # Draw the arrow.\n draw_arrow (button2, ARROW_DOWN, state_button, cls, st)\n rect_button2.y = rect.height - width_button - border\n rect_button2.centerx = rect.centerx\n self.image.blit (button2, rect_button2)\n\n # Create the slider.\n slider_size = self._get_slider_size ()\n if slider_size > 0:\n sl = style.engine.draw_slider (width_button, slider_size,\n tmp_state, cls, st)\n r = sl.get_rect ()\n r.centerx = rect.centerx\n r.centery = self._get_coords_from_value ()\n self.image.blit (sl, r)\n","sub_path":"ocempgui/widgets/ScrollBar.py","file_name":"ScrollBar.py","file_ext":"py","file_size_in_byte":28896,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"85194947","text":"class SinglyLinkedList:\r\n\tdef __init__(self):\r\n\t\tself.length = 0\r\n\r\n\tdef __str__(self):\r\n\t\toutput = \"[\"\r\n\t\tif(self.length == 0):\r\n\t\t\toutput += \"]\"\r\n\t\telif(self.length == 1):\r\n\t\t\toutput += \"{0}]\".format(self.head.data)\r\n\t\telse:\r\n\t\t\tnode = self.head\r\n\t\t\tfor i in range(self.length):\r\n\t\t\t\tif(i == self.length - 1):\r\n\t\t\t\t\toutput += \"{0}\".format(node.data)\r\n\t\t\t\telse:\r\n\t\t\t\t\toutput += \"{0}, \".format(node.data)\r\n\t\t\t\t\tnode = node.nextNode\r\n\t\t\toutput += \"]\"\r\n\r\n\t\treturn output\r\n\r\n\r\n\tdef access(self, index):\r\n\t\tif(index >= self.length):\r\n\t\t\traise self.InvalidIndexError(\"No element at index {0} in list of length {1}\".format(index, self.length))\r\n\t\telif(index < 0):\r\n\t\t\traise self.InvalidIndexError(\"Negative index of {0}\".format(index))\r\n\t\telse:\r\n\t\t\tnode = self.head\r\n\t\t\tfor i in range(index):\r\n\t\t\t\tnode = node.nextNode\r\n\t\t\treturn node.data\r\n\r\n\tdef insert(self, data):\r\n\t\tif(self.length == 0):\r\n\t\t\tself.head = self.Node(data)\r\n\t\t\tself.tail = self.head\r\n\t\telif(self.length == 1):\r\n\t\t\tself.head.nextNode = self.Node(data)\r\n\t\t\tself.tail = self.head.nextNode\r\n\t\telse:\r\n\t\t\tself.tail.nextNode = self.Node(data)\r\n\t\t\tself.tail = self.tail.nextNode\r\n\t\tself.length += 1\r\n\r\n\tdef search(self, data):\r\n\t\tif(self.length == 0):\r\n\t\t\treturn -1\r\n\t\telse:\r\n\t\t\tnode = self.head\r\n\t\t\ti = 0\r\n\t\t\twhile True:\r\n\t\t\t\tif(node.data == data):\r\n\t\t\t\t\treturn i\r\n\t\t\t\tif(node.nextNode == None):\r\n\t\t\t\t\treturn -1\r\n\t\t\t\tnode = node.nextNode\r\n\t\t\t\ti += 1\r\n\r\n\tdef delete(self, index=None):\r\n\t\tif(index == None):\r\n\t\t\tindex = self.length\r\n\r\n\t\tif(index >= self.length):\r\n\t\t\traise self.InvalidIndexError(\"No element at index {0} in list of length {1}\".format(index, self.length))\r\n\t\telif(index < 0):\r\n\t\t\traise self.InvalidIndexError(\"Negative index of {0}\".format(index))\r\n\t\telif(index == 0):\r\n\t\t\tself.head = self.head.nextNode\r\n\t\telse:\r\n\t\t\tprev = self.head\r\n\t\t\tfor i in range(index - 1):\r\n\t\t\t\tprev = prev.nextNode\r\n\t\t\tprev.nextNode = prev.nextNode.nextNode\r\n\t\tself.length -= 1\r\n\r\n\r\n\tclass InvalidIndexError(Exception):\r\n\t\tdef __init__(self, msg):\r\n\t\t\tself.msg = msg\r\n\t\tdef __str__(self):\r\n\t\t\treturn self.msg\r\n\r\n\tclass Node():\r\n\t\tdef __init__(self, data):\r\n\t\t\tself.data = data\r\n\t\t\tself.nextNode = None\r\n\r\nif __name__ == \"__main__\":\r\n\tmyList = SinglyLinkedList()\r\n\tmyList.insert(1)\r\n\tmyList.insert(2)\r\n\tmyList.insert(3)\r\n\tmyList.insert(4)\r\n\tmyList.delete(4)\r\n\tprint(myList)","sub_path":"data_structures/singlylinkedlist.py","file_name":"singlylinkedlist.py","file_ext":"py","file_size_in_byte":2317,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"424960923","text":"import json\nfrom flask import make_response, jsonify\nfrom flask_api import status\nfrom util import jsonDefault\n\n# from dcsmp.util.util import DateTimeEncoder, model_to_map,filter_result\n\n# def get_json(obj, fields = None):\n# if obj is None:\n# # Response does not contain any data\n# response = make_response()\n# response.status_code = status.HTTP_204_NO_CONTENT\n# else:\n # if type(obj) is list:\n # result = []\n # for item in obj:\n # if isinstance(item, Model):\n # i = model_to_map(item)\n # else:\n # i = item\n #\n # if type(i) is dict:\n # i = filter_result(i, fields)\n # result.append(i)\n # else:\n # if isinstance(obj, Model):\n # result = model_to_map(obj)\n # else:\n # result = obj\n #\n # if type(result) is dict:\n # result = filter_result(result, fields)\n # result = {\n # \"status\": status.HTTP_200_OK,\n # \"errorMessage\": \"\",\n # \"result\": result\n # }\n # response = make_response(json.dumps(result, default=jsonDefault))\n #\n # response.mimetype = 'application/json'\n # return response\n\n\ndef created(body=None):\n result = {\n \"status\": status.HTTP_201_CREATED,\n \"errorMessage\": \"\",\n }\n if body:\n result['result'] = body\n response = make_response(json.dumps(result, default=jsonDefault))\n\n response.status_code = status.HTTP_201_CREATED\n response.mimetype = 'application/json'\n return response\n\n\ndef bad_request(message):\n response = jsonify({\n \"status\": status.HTTP_400_BAD_REQUEST,\n \"errorMessage\": message\n })\n response.status_code = status.HTTP_400_BAD_REQUEST\n return response\n\n\ndef not_found(message=None):\n response = make_response()\n response.status_code = status.HTTP_404_NOT_FOUND\n return response\n\n\ndef duplicate(message):\n response = jsonify({\n \"status\": status.HTTP_409_CONFLICT,\n \"errorMessage\": message\n })\n response.status_code = status.HTTP_409_CONFLICT\n return response\n\n\ndef unsupported_media_type(message):\n response = jsonify({\n \"status\": status.HTTP_415_UNSUPPORTED_MEDIA_TYPE,\n \"errorMessage\": message\n })\n response.status_code = status.HTTP_415_UNSUPPORTED_MEDIA_TYPE\n return response\n\n\ndef ok(body=None):\n result = {\n \"status\": status.HTTP_200_OK,\n \"errorMessage\": \"\",\n }\n if body:\n result['result'] = body\n response = make_response(json.dumps(result, default=jsonDefault))\n\n response.status_code = status.HTTP_200_OK\n response.mimetype = 'application/json'\n return response\n","sub_path":"cckm/response.py","file_name":"response.py","file_ext":"py","file_size_in_byte":2803,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"620525593","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('w', '0008_auto_20151018_2204'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='archive',\n name='value',\n field=models.FloatField(default=0, null=True, blank=True),\n ),\n migrations.AlterField(\n model_name='rule',\n name='input_attribute',\n field=models.CharField(default=0, max_length=6, choices=[(b'VALUE', b'VALUE'), (b'STATUS', b'STATUS'), (b'NT', b'NT'), (b'IV', b'IV'), (b'W', b'WARNING'), (b'A', b'ALARM')]),\n ),\n migrations.AlterField(\n model_name='rule',\n name='output_attribute',\n field=models.CharField(default=0, max_length=6, choices=[(b'VALUE', b'VALUE'), (b'STATUS', b'STATUS'), (b'NT', b'NT'), (b'IV', b'IV'), (b'W', b'WARNING'), (b'A', b'ALARM')]),\n ),\n ]\n","sub_path":"growmat/w/migrations/0009_auto_20151018_2241.py","file_name":"0009_auto_20151018_2241.py","file_ext":"py","file_size_in_byte":1009,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"377063694","text":"# Python imports\nfrom setuptools import setup\n\n#Error in atexit._run_exitfuncs, TypeError: 'NoneType' object is not callable\nfrom multiprocessing import util\n\n# Project imports\nfrom notable import app\n\n# Attributes\nAUTHOR = 'John McFarlane'\nDESCRIPTION = 'A very simple note taking application'\nEMAIL = 'john.mcfarlane@rockfloat.com'\nNAME = 'Notable'\nPYPI = 'http://pypi.python.org/packages/source/N/Notable'\nURL = 'https://github.com/jmcfarlane/Notable'\nCLASSIFIERS = \"\"\"\nDevelopment Status :: 2 - Pre-Alpha\nIntended Audience :: Developers\nLicense :: OSI Approved :: MIT License\nOperating System :: OS Independent\nProgramming Language :: Python\nTopic :: Internet :: WWW/HTTP\nIntended Audience :: End Users/Desktop\nTopic :: Office/Business :: News/Diary\nTopic :: Security :: Cryptography\nTopic :: Utilities\n\"\"\"\n\nsetup(\n author = AUTHOR,\n author_email = EMAIL,\n classifiers = [c for c in CLASSIFIERS.split('\\n') if c],\n description = DESCRIPTION,\n download_url = '%s/Notable-%s.tar.gz' % (PYPI, app.version),\n include_package_data = True,\n name = NAME,\n packages = ['notable'],\n scripts = ['scripts/notable'],\n test_suite='nose.collector',\n url = URL,\n version = app.version\n)\n","sub_path":"pypi_install_script/Notable-0.4.2.tar/setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1212,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"153185783","text":"# coding: utf-8\r\n\r\n#this file is to scrape news title, article link and image link from some major news websites initially\r\n#next, insert scraped content into database and only keep the latest information\r\n#after that, use home made graph structure traversal algorithm to extract key info and remove similar contents\r\n#finally, send html emails including titles, links and images\r\n#for details of scraping, database and outlook manipulation, plz take the following link as a reference\r\n# https://github.com/tattooday/web-scraping/blob/master/Feeds%20from%20Database.py\r\n\r\nimport pandas as pd\r\nfrom bs4 import BeautifulSoup as bs \r\nimport marketanalysis as ma\r\nimport datetime as dt\r\nimport win32com.client as win32 \r\nimport sqlite3\r\nimport os\r\nimport re\r\nimport copy\r\nimport time\r\nos.chdir('d:/')\r\n\r\n#this is a home made special package for text mining\r\n#it is designed to extract key information and remove similar contents\r\n#for details of this graph traversal algorithm plz refer to the following link\r\n# https://github.com/tattooday/graph-theory/blob/master/Text%20Mining%20project/alternative%20bfs.py\r\nimport graph\r\n\r\n#main stuff\r\ndef main():\r\n \r\n ec=scrape('https://www.economist.com/middle-east-and-africa/',economist)\r\n aj=scrape('https://www.aljazeera.com/topics/regions/middleeast.html',aljazeera)\r\n tr=scrape('https://www.reuters.com/news/archive/middle-east',reuters) \r\n bc=scrape('https://www.bbc.co.uk/news/world/middle_east',bbc)\r\n ws=scrape('https://www.wsj.com/news/types/middle-east-news',wsj)\r\n ft=scrape('https://www.ft.com/world/mideast',financialtimes)\r\n bb=scrape('https://www.bloomberg.com/view/topics/middle-east',bloomberg)\r\n cn=scrape('https://edition.cnn.com/middle-east',cnn)\r\n fo=scrape('https://fortune.com/tag/middle-east/',fortune)\r\n \r\n #concat scraped data via append, can use pd.concat as an alternative\r\n #unlike the previous version, current version does not sort information by source\r\n df=ft\r\n for i in [aj,tr,bc,ws,cn,fo,ec,bb]:\r\n df=df.append(i)\r\n \r\n #CRUCIAL!!!\r\n #as we append dataframe together, we need to reset the index\r\n #otherwise, we would not be able to use reindex in database function call\r\n df.reset_index(inplace=True,drop=True)\r\n \r\n #first round, insert into database and remove outdated information\r\n df=database(df)\r\n \r\n #second round, use home made package to remove similar contents\r\n output=graph.remove_similar(df,graph.stopword)\r\n \r\n print(output)\r\n \r\n html='
\r\n #the issue with this method is that we have to scrape the website repeatedly to get images\r\n #or we can use < img src='data:image/jpg; base64, [remove the brackets and paste base64]'/>\r\n #but this is blocked by most email clients including outlook 2016\r\n for i in range(len(output)):\r\n \r\n html+=\"\"\"{}'.format(content))\n","sub_path":"tests/app/projects/test_utils.py","file_name":"test_utils.py","file_ext":"py","file_size_in_byte":351,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"283507902","text":"import os, sys\nsys.path.append('..')\nfrom PyQt5.QtCore import (\n Qt,\n QObject,\n QTimer,\n QByteArray,\n QDir,\n QFile,\n QFileInfo,\n QStandardPaths,\n QUrl,\n QMutex,\n QMutexLocker,\n QStandardPaths,\n pyqtSignal,\n pyqtSlot,\n QDateTime,\n QSettings,\n QVariant,\n QRegExp,\n QThread,\n QIODevice,\n QBuffer,\n QMessageLogger,\n qInstallMessageHandler,\n QCommandLineParser,\n QCommandLineOption,\n QJsonDocument,\n QJsonValue,\n QJsonParseError,\n QCoreApplication,\n)\nfrom PyQt5.QtNetwork import (\n QNetworkInterface,\n QHostAddress,\n QAbstractSocket,\n QUdpSocket,\n QSslConfiguration,\n QSsl,\n QSslCertificate,\n QSslKey,\n QSslSocket,\n )\nfrom PyQt5.QtWebSockets import (\n QWebSocketServer,\n QWebSocket,\n QWebSocketProtocol,\n )\nfrom PyQt5.QtSql import (\n QSqlDatabase,\n QSqlQuery\n )\n\n\nfrom PyQt5.QtSerialPort import (\n QSerialPort, QSerialPortInfo\n )\n\nfrom qtwebapp.httpserver import (\n g_session_store,\n HttpRequest,\n HttpResponse,\n HttpCookie,\n HttpSessionStore,\n HttpListener,\n StaticFileController,\n createSslConfig,\n )\nfrom qtwebapp.util import json_dumps, json_loads\nfrom qtwebapp.mime import MIMETYPE\n\n\nCONFIG = '../push_gateway.ini'\n\ndef getSettingsConfig(settings, settingGroup=''):\n ret = {}\n settings.beginGroup(settingGroup)\n if settingGroup == 'remote':\n ret['url'] = settings.value('url', defaultValue='')\n ret['retryTimes'] = int(settings.value('retryTimes', defaultValue='3'))\n ret['retryInterval'] = int(settings.value('retryInterval', defaultValue='60000'))\n ret['connectTimeout'] = int(settings.value('connectTimeout', defaultValue='5000'))\n\n elif settingGroup == 'local':\n ret['ipStart'] = settings.value('ipStart', defaultValue='192.168.1.101')\n ret['ipEnd'] = settings.value('ipEnd', defaultValue='192.168.1.130')\n ret['username'] = settings.value('username', defaultValue='a')\n ret['password'] = settings.value('password', defaultValue='aaaaaaa')\n ret['scanInterval'] = int(settings.value('scanInterval', defaultValue='3000'))\n ret['scanTimeout'] = int(settings.value('scanTimeout', defaultValue='1500'))\n # ret['pushInterval'] = int(settings.value('pushInterval', defaultValue='5000'))\n # ret['requestTimeout'] = int(settings.value('requestTimeout', defaultValue='2000'))\n elif settingGroup == 'hololens_device_portal_api':\n grps = settings.childGroups()\n for grp in grps:\n settings.beginGroup(grp)\n ret[grp] = {}\n for ks in settings.childKeys():\n ret[grp][ks] = settings.value(ks, defaultValue='')\n settings.endGroup()\n\n settings.endGroup()\n return ret\n\nclass TestServerController(StaticFileController):\n signalService = pyqtSignal(HttpRequest, HttpResponse)\n def __init__(self,\n parent=None,\n enableSession=False,\n unauthorizedPage='',\n loginPage='',\n **kwargs):\n super(TestServerController, self).__init__(parent=parent, **kwargs)\n self._parent = parent\n self.enableSession = enableSession\n self.unauthorizedPage = unauthorizedPage\n self.loginPage = loginPage\n\n\n\n\n def responseStaticPage(self, path, response):\n self.mutex.lock()\n filepath = self.docroot + path\n file = QFile(filepath)\n if file.open(QIODevice.ReadOnly):\n self.setContentType(filepath, response)\n # response.setHeader(b\"Cache-Control\", b\"max-age=\" + str(int(self.maxAge / 1000)).encode(encoding='utf-8'))\n response.setHeader(b'Cache-Control', b'no-cache')\n # response.setHeader(b'Access-Control-Allow-Origin', b'*')\n # response.setHeader(b'Access-Control-Allow-Methods', b'POST')\n response.write(file.readAll(), lastPart=True)\n file.close()\n self.mutex.unlock()\n else:\n self.mutex.unlock()\n response.setStatus(401, b'Unauthorized')\n msg = '401 forbidden: {} page error'.format(path)\n response.write(msg.encode(), lastPart=True)\n\n def handleRequest(self, path, request, response):\n ret = {}\n not200 = False\n if path == '/api/power/battery':\n ret = {\n \"AcOnline\": 1,\n \"BatteryPresent\": 1,\n \"Charging\": 0,\n \"DefaultAlert1\": 0,\n \"DefaultAlert2\": 1628,\n \"EstimatedTime\": 17295,\n \"MaximumCapacity\": 16275,\n \"RemainingCapacity\": 14749\n }\n elif path == '/api/holographic/mrc/status':\n ret = {\"IsRecording\" : False,\n \"ProcessStatus\" : {\n \"MrcProcess\" : \"Running\"\n }\n }\n elif path == '/api/holographic/mrc/file':\n if request.getMethod().decode().lower() == 'delete':\n response.setStatus(200, b'OK')\n response.write(b'', lastPart=True)\n return\n\n\n filename = request.getParameter(b'filename').decode()\n filename = QByteArray.fromBase64(filename.encode()).data().decode()\n fi = QFileInfo(filename)\n if fi.exists():\n path = fi.absoluteFilePath()\n f = QFile(path)\n ext = path[path.rindex('.'):]\n f.open(QIODevice.ReadOnly)\n ba = f.readAll()\n f.close()\n response.setHeader(b'Content-Type', MIMETYPE[ext].encode())\n # response.setHeader(b'Content-Length', str(len(ba)).encode())\n response.setStatus(200, b'OK')\n response.write(ba, lastPart=True)\n return\n else:\n not200 = True\n ret = {'error':'404 not found'}\n\n elif path == '/api/holographic/mrc/files':\n ret = {\"MrcRecordings\" : [\n {\n \"CreationTime\" : 131423339684371628.0,\n \"FileName\" : \"20170619_161928_HoloLens.mp4\",\n \"FileSize\" : 4713633\n },\n {\n \"CreationTime\" : 131423339609756226.0,\n \"FileName\" : \"20170619_161920_HoloLens.jpg\",\n \"FileSize\" : 301614\n },\n {\n \"CreationTime\" : 131423339870432136.0,\n \"FileName\" : \"20170619_161946_HoloLens.jpg\",\n \"FileSize\" : 357791\n }\n ]}\n\n elif path == '/api/holographic/os/webmanagement/settings/https':\n ret = {\"httpsRequired\" : False}\n\n elif path == '/api/holographic/mrc/thumbnail':\n filename = request.getParameter(b'filename').decode()\n filename = QByteArray.fromBase64(filename.encode()).data().decode()\n fi = QFileInfo('test_base64_json.json')\n l = json_loads(fi)\n ll = list(filter(lambda x:x['filename'] == filename, l))\n ba = b''\n if len(ll):\n ba = QByteArray.fromBase64(ll[0]['thumbnail'].encode()).data()\n response.setHeader(b'Content-Type', (MIMETYPE['.jpg'] + '').encode())\n response.setStatus(200, b'OK')\n response.write(ba, lastPart=True)\n return\n else:\n not200 = True\n response.setHeader(b'Content-Type', (MIMETYPE['.json'] + ';charset=utf-8').encode())\n response.setStatus(200, b'OK')\n if not200:\n response.setStatus(404, b'Not found')\n response.write(json_dumps(ret).encode(), lastPart=True)\n\n def service(self, request, response):\n global g_session_store\n path = request.getPath().decode(encoding='utf-8')\n self.handleRequest(path, request, response)\n\n\ndef main():\n app = QCoreApplication(sys.argv)\n controller = TestServerController(parent=app)\n # th0 = QThread()\n # th0.start()\n # th0 = QCoreApplication.instance().thread()\n httplistener = HttpListener(optParser=None,\n requestHandler=controller,\n parent=None,\n maxRequestSize=1600000000,\n maxMultiPartSize=1000000000,\n port=80,\n # host='192.168.1.129',\n )\n\n # httplistener.moveToThread(th0)\n httplistener.start.connect(httplistener.slotListen)\n httplistener.start.emit()\n sys.exit(app.exec_())\n\nif __name__ == '__main__':\n main()\n\n\n","sub_path":"test/test_holo_server.py","file_name":"test_holo_server.py","file_ext":"py","file_size_in_byte":9707,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"3933494","text":"#!/bin/python3\n\nimport numpy as np\nfrom field import field\nimport analyzer\n\n\n\ndef solve_sudoku(arr : np.array):\n \"\"\"solves a sudoku for a 9X9 array\"\"\"\n #init:\n ls = field_array_of(arr)\n changing = True\n finished = False\n \n #the actual solving algorithm:\n #tries to solve it by analyzing rows, columns and fields, starts guessing\n #if fixpoint is reached with other technique\n while not finished:\n while changing:\n analyzer.iterate(ls)\n changing = update_knowns(ls)\n if ready(ls):\n finished = True\n else:\n #the case where we have to guess\n a,b = analyzer.fewest_poss(ls)\n for val in ls[a,b].poss:\n arr = to_array(ls)\n arr[a,b] = val\n try:\n res = solve_sudoku(arr)\n except analyzer.NotSolvableException:\n print(\"one branch not solvable\")\n else:\n return res\n raise analyzer.NotSolvableException \n # output the result\n res = to_array(ls)\n return res\n\ndef to_array(ls):\n f = lambda li: list(map(lambda x : x.value, li))\n res = np.array(list(map(f, ls)))\n return res\n\ndef field_array_of(arr : np.array):\n ls = np.array([[None]*9]*9)\n\n #zeroes mean there was no entry\n for i in range(9):\n for j in range(9):\n ls[i,j] = field(arr[i,j],i,j)\n return ls\n\n\ndef update_knowns(ls):\n \"\"\"updates values where possible, and returns True if it was still able to\n change something, to use it for fixpoint iteration\"\"\"\n changing = False\n for i in ls:\n for j in i:\n if j.value == 0:\n if len(j.poss) == 0:\n raise analyzer.NotSolvableException\n elif len(j.poss)==1:\n if j.poss[0] == 0:\n raise analyzer.NotSolvableException\n j.value = j.poss[0]\n j.poss = []\n changing = True\n analyzer.iterate(ls)\n return changing\n\ndef ready(ls):\n\n for i in ls:\n for j in i:\n if j.value == 0:\n return False\n return True\n\n\n\n\n","sub_path":"sudoku/solver.py","file_name":"solver.py","file_ext":"py","file_size_in_byte":2239,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"324118541","text":"from django.db import models\nfrom django.conf import settings\n\n\nclass Category(models.Model):\n name = models.CharField(max_length=50)\n\n def __str__(self):\n return f\"{self.name}\"\n\n\nclass Tag(models.Model):\n name = models.CharField(max_length=30)\n\n def __str__(self):\n return f\"{self.name}\"\n\n\nclass Product(models.Model):\n title = models.CharField(max_length=50, null=False, blank=False)\n description = models.TextField()\n price = models.DecimalField(max_digits=20, decimal_places=2)\n stock = models.IntegerField()\n category = models.ForeignKey(\n to=Category,\n on_delete=models.CASCADE,\n blank=True\n )\n image = models.ImageField()\n tags = models.ManyToManyField(Tag, blank=True)\n\n @property\n def rating_list(self):\n rating_list = []\n objs = self.ratings.all()\n for obj in objs:\n rating_list.append({\n 'user': obj.user.username,\n 'rating': obj.rating,\n 'comment': obj.comment\n })\n return rating_list\n\n def __str__(self):\n return f\"{self.title} : {self.description}\"\n\n\nclass Wishlist(models.Model):\n user = models.ForeignKey(\n settings.AUTH_USER_MODEL,\n on_delete=models.CASCADE\n )\n products = models.ManyToManyField(Product, blank=True)\n\n def __str__(self):\n return f\"{self.user.username} : {self.products.all()}\"\n\n\nclass Cart(models.Model):\n user = models.ForeignKey(\n settings.AUTH_USER_MODEL,\n on_delete=models.CASCADE\n )\n products = models.ManyToManyField(Product, blank=True)\n\n def __str__(self):\n return f\"{self.user.username} : {self.products.all()}\"\n\n\nclass Orders(models.Model):\n user = models.ForeignKey(\n settings.AUTH_USER_MODEL,\n on_delete=models.CASCADE,\n related_name=\"orders\")\n total = models.FloatField()\n payment_id = models.CharField(max_length=50)\n order_id = models.CharField(max_length=50, unique=True)\n signature = models.CharField(max_length=250)\n date_ordered = models.DateTimeField(auto_now_add=True)\n\n @property\n def items(self):\n item_list = []\n objs = self.order_items.all()\n for obj in objs:\n item_list.append({\n \"item\": obj.product.title,\n \"id\": obj.product.id,\n \"price\": obj.product.price,\n \"image\": str(obj.product.image),\n \"quantity\": obj.qty,\n })\n return item_list\n\n def __str__(self):\n return f\"{self.order_id}\"\n\n\nclass OrderItem(models.Model):\n order = models.ForeignKey(\n Orders,\n on_delete=models.CASCADE,\n related_name=\"order_items\"\n )\n qty = models.IntegerField()\n product = models.ForeignKey(\n to=Product,\n on_delete=models.CASCADE,\n blank=True\n )\n\n\nRATING = (\n (1, 1),\n (2, 2),\n (3, 3),\n (4, 4),\n (5, 5),\n)\n\n\nclass Rating(models.Model):\n user = models.ForeignKey(\n settings.AUTH_USER_MODEL,\n on_delete=models.CASCADE\n )\n product = models.ForeignKey(\n Product,\n on_delete=models.CASCADE,\n related_name='ratings'\n )\n rating = models.PositiveIntegerField(choices=RATING, default=0)\n comment = models.CharField(max_length=300, default=\"\")\n\n def _str_(self):\n return f\"{self.product} - {self.rating}\"\n","sub_path":"backend/store/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":3399,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"}
+{"seq_id":"504758724","text":"# Copyright (c) 2012 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\nfrom buildbot.process.properties import WithProperties\n\nfrom master import master_config\nfrom master import master_utils\nfrom master.factory import remote_run_factory\n\nimport master_site_config\n\nActiveMaster = master_site_config.ChromiumWebkit\n\ndefaults = {}\n\nhelper = master_config.Helper(defaults)\nB = helper.Builder\nF = helper.Factory\nT = helper.Triggerable\n\nrevision_getter = master_utils.ConditionalProperty(\n lambda build: build.getProperty('revision'),\n WithProperties('%(revision)s'),\n 'master')\n\ndef m_remote_run_chromium_src(recipe, **kwargs):\n kwargs.setdefault('revision', revision_getter)\n return remote_run_factory.RemoteRunFactory(\n active_master=ActiveMaster,\n repository='https://chromium.googlesource.com/chromium/src.git',\n recipe=recipe,\n factory_properties={'path_config': 'kitchen'},\n use_gitiles=True,\n **kwargs)\n\ndefaults['category'] = 'layout'\n\n\n################################################################################\n## Release\n################################################################################\n\n#\n# Linux Rel Builder/Tester\n#\n\nB('WebKit Linux Precise', 'f_webkit_linux_rel', scheduler='global_scheduler')\nF('f_webkit_linux_rel', m_remote_run_chromium_src('chromium'))\n\nB('WebKit Linux Trusty', 'f_webkit_linux_rel_trusty',\n scheduler='global_scheduler')\nF('f_webkit_linux_rel_trusty', m_remote_run_chromium_src('chromium'))\n\nB('WebKit Linux Precise ASAN', 'f_webkit_linux_rel_asan',\n scheduler='global_scheduler', auto_reboot=True)\nF('f_webkit_linux_rel_asan', m_remote_run_chromium_src('chromium'))\n\nB('WebKit Linux Precise MSAN', 'f_webkit_linux_rel_msan',\n scheduler='global_scheduler', auto_reboot=True)\nF('f_webkit_linux_rel_msan', m_remote_run_chromium_src('chromium'))\n\nB('WebKit Linux Precise Leak', 'f_webkit_linux_leak_rel',\n scheduler='global_scheduler', category='layout')\nF('f_webkit_linux_leak_rel', m_remote_run_chromium_src('chromium'))\n\n\n################################################################################\n## Debug\n################################################################################\n\n#\n# Linux Dbg Webkit builders/testers\n#\n\nB('WebKit Linux Precise (dbg)', 'f_webkit_dbg_tests',\n scheduler='global_scheduler', auto_reboot=True)\nF('f_webkit_dbg_tests', m_remote_run_chromium_src('chromium'))\n\n\ndef Update(_config, _active_master, c):\n return helper.Update(c)\n","sub_path":"masters/master.chromium.webkit/master_linux_webkit_latest_cfg.py","file_name":"master_linux_webkit_latest_cfg.py","file_ext":"py","file_size_in_byte":2575,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"}
+{"seq_id":"47277008","text":"import multiprocessing\nimport os\nimport re\nimport threading\nimport shutil\nimport sys\nfrom typing import Optional\nimport uuid\n\nfrom v2.containers import BuildCfg\nfrom v2.logging_subprocess import call\nfrom v2.workspace import (\n CANNOT_BUILD,\n BENCHMARK_BRANCH_NAME, BENCHMARK_BRANCH_ROOT, BENCHMARK_ENV, BENCHMARK_ENV_BUILT,\n BUILD_LOG_ROOT, BUILD_IN_PROGRESS_ROOT,\n MUTATION_LOCK, REF_REPO_ROOT)\n\n_NAMESPACE_LOCK = threading.Lock()\n_CONDA_ENV_TEMPLATE = \"env_{n:0>2}\"\n_MAX_ACTIVE_ENVS = 50\n\n\nclass OutOfEnvsError(Exception):\n pass\n\n\nclass _Unbuildable:\n def __init__(self):\n self._known_unbuildable = None\n\n def _lazy_init(self):\n if self._known_unbuildable is None:\n with open(CANNOT_BUILD, \"at\") as f:\n pass\n\n with open(CANNOT_BUILD, \"rt\") as f:\n self._known_unbuildable = set(f.read().splitlines(keepends=False))\n\n def check(self, checkout: str) -> bool:\n self._lazy_init()\n return checkout in self._known_unbuildable\n\n def update(self, checkout: str) -> None:\n self._lazy_init()\n MUTATION_LOCK.get()\n if checkout not in self._known_unbuildable:\n with open(CANNOT_BUILD, \"at\") as f:\n f.write(f\"{checkout}\\n\")\n self._known_unbuildable.add(checkout)\n\n def reset_unbuildable(self, checkout: str) -> None:\n self._lazy_init()\n MUTATION_LOCK.get()\n if checkout in self._known_unbuildable:\n self._known_unbuildable.remove(checkout)\n with open(CANNOT_BUILD, \"wt\") as f:\n f.write(\"\\n\".join(self._known_unbuildable) + \"\\n\")\n\n_UnbuildableSingleton = _Unbuildable()\ncheck_unbuildable = _UnbuildableSingleton.check\nmark_unbuildable = _UnbuildableSingleton.update\nreset_unbuildable = _UnbuildableSingleton.reset_unbuildable\n\n\ndef make_conda_env(\n env_path: Optional[str] = None,\n build_cfg: BuildCfg = BuildCfg(),\n):\n MUTATION_LOCK.get()\n cleanup = (env_path is None)\n success = False\n try:\n with _NAMESPACE_LOCK:\n if env_path is None:\n active_envs = set(os.listdir(BUILD_IN_PROGRESS_ROOT))\n for i in range(_MAX_ACTIVE_ENVS):\n env_name = _CONDA_ENV_TEMPLATE.format(n=i)\n if env_name not in active_envs:\n break\n else:\n raise OutOfEnvsError(\"Failed to create env. Too many already exist.\")\n\n env_path = os.path.join(BUILD_IN_PROGRESS_ROOT, env_name)\n else:\n env_name = \"custom\"\n\n mkl_spec = f\"=={build_cfg.mkl_version}\" if build_cfg.mkl_version else \"\"\n call(\n f\"conda create --no-default-packages -y --prefix {env_path} python={build_cfg.python_version}\",\n shell=True,\n check=True,\n task_name=f\"Conda env creation: {env_name}\",\n log_dir=BUILD_LOG_ROOT,\n )\n\n call(\n f\"\"\"\n echo ADD_INTEL\n conda config --env --add channels intel\n\n echo MAIN_INSTALL\n conda install -y numpy ninja pyyaml mkl{mkl_spec} mkl-include setuptools cmake cffi hypothesis typing_extensions pybind11 ipython\n\n echo GLOG_INSTALL\n conda install -y -c conda-forge glog\n\n echo INSTALL_VALGRIND\n conda install -y -c conda-forge valgrind\n \"\"\",\n shell=True,\n check=True,\n task_name=f\"Conda env install: {env_name}\",\n conda_env=env_path,\n log_dir=BUILD_LOG_ROOT,\n )\n\n success = True\n return env_path\n\n finally:\n if cleanup and not success and env_path is not None and os.path.exists(env_path):\n shutil.rmtree(env_path)\n\n\ndef _build(\n repo_path: str,\n checkout: Optional[str],\n setup_mode: str,\n conda_env: str,\n build_cfg: BuildCfg,\n show_progress: bool,\n taskset_cores: Optional[str],\n nice: Optional[str],\n max_jobs: Optional[int],\n) -> int:\n assert setup_mode in (\"develop\", \"install\")\n\n no_fbgemm = '-c submodule.\"third_party/fbgemm\".update=none'\n no_xnnpack = '-c submodule.\"third_party/XNNPACK\".update=none'\n no_nervanagpu = '-c submodule.\"third_party/nervanagpu\".update=none'\n call(\n f\"\"\"\n retry () {{ $* || (sleep 1 && $*) || (sleep 2 && $*); }}\n\n git checkout .\n git clean -fd\n git checkout .\n git checkout {checkout}\n git clean -fd\n\n # `git submodule sync` doesn't sync submodule submodules, which can\n # cause build failures. So instead we just start over.\n rm -rf third_party/*\n git checkout third_party\n retry git submodule sync\n\n # History for XNNPack has changed, so this will fail in February/March\n retry git {no_fbgemm} {no_xnnpack} {no_nervanagpu} submodule update --init --recursive\n \"\"\",\n shell=True,\n cwd=repo_path,\n check=True,\n conda_env=conda_env,\n task_name=\"(pre) Build PyTorch\",\n log_dir=BUILD_LOG_ROOT,\n )\n\n call(\n f\"\"\"\n retry () {{ $* || (sleep 1 && $*) || (sleep 2 && $*); }}\n retry git submodule update --init --recursive\n \"\"\",\n shell=True,\n cwd=repo_path,\n check=False,\n conda_env=conda_env,\n task_name=\"(pre) Build PyTorch\",\n log_dir=BUILD_LOG_ROOT,\n )\n\n progress_pattern = re.compile(r\"^\\[[0-9]+/[0-9]+\\]\\s.+$\")\n def per_line_fn(l):\n if progress_pattern.search(l):\n print(f\"\\r{l.strip()[:120]:<120}\", end=\"\")\n sys.stdout.flush()\n\n if \"BUILD_DONE\" in l:\n print(\"\\r\")\n\n taskset_str = f\"taskset --cpu-list {taskset_cores} \" if taskset_cores else \"\"\n nice_str = f\"nice -n {nice} \" if nice is not None else \"\"\n retcode = call(\n f\"\"\"\n # CCACHE variables are generally in `.bashrc`\n source ~/.bashrc\n which c++ | awk '{{print \"which c++: \"$1}}'\n\n {taskset_str}{nice_str}python -u setup.py clean\n {taskset_str}{nice_str}python -u setup.py {setup_mode}\n echo BUILD_DONE\n \"\"\",\n shell=True,\n cwd=repo_path,\n env={\n \"USE_DISTRIBUTED\": \"0\",\n \"BUILD_TEST\": build_cfg.build_tests,\n \"USE_CUDA\": \"0\",\n \"USE_FBGEMM\": \"0\",\n \"USE_NNPACK\": \"0\",\n \"USE_QNNPACK\": \"0\",\n \"USE_PYTORCH_QNNPACK\": \"0\",\n \"USE_XNNPACK\": \"0\",\n \"BUILD_CAFFE2_OPS\": \"0\",\n \"REL_WITH_DEB_INFO\": \"1\",\n \"MKL_THREADING_LAYER\": \"GNU\",\n \"USE_NUMA\": \"0\",\n \"MAX_JOBS\": \"\" if max_jobs is None else str(max_jobs),\n \"CFLAGS\": f\"-Wno-error=stringop-truncation\",\n },\n per_line_fn=per_line_fn if show_progress else None,\n conda_env=conda_env,\n task_name=\"Build PyTorch\",\n log_dir=BUILD_LOG_ROOT,\n )\n\n if not retcode:\n retcode = call(\n 'python -c \"import torch\"',\n shell=True,\n env={\n \"MKL_THREADING_LAYER\": \"GNU\",\n },\n conda_env=conda_env,\n task_name=\"Test PyTorch\",\n log_dir=BUILD_LOG_ROOT,\n )\n\n if retcode:\n mark_unbuildable(checkout)\n shutil.rmtree(conda_env)\n\n return retcode\n\n\ndef build_benchmark_env():\n MUTATION_LOCK.get()\n if os.path.exists(BENCHMARK_ENV_BUILT):\n return\n\n if not os.path.exists(BENCHMARK_ENV):\n shutil.rmtree(BENCHMARK_ENV, ignore_errors=True)\n\n make_conda_env(env_path=BENCHMARK_ENV, build_cfg=BuildCfg())\n call(\"pip install yattag\", shell=True, conda_env=BENCHMARK_ENV)\n\n # By default, build will try to take over all cores. However, this can\n # lead to OOM during some of the memory-intensive parts of compilation.\n max_jobs = max(int(multiprocessing.cpu_count() * 0.9), 1)\n\n retcode = _build(\n BENCHMARK_BRANCH_ROOT,\n BENCHMARK_BRANCH_NAME,\n \"develop\",\n BENCHMARK_ENV,\n build_cfg=BuildCfg(),\n show_progress=True,\n taskset_cores=None,\n nice=None,\n max_jobs=max_jobs,\n )\n\n assert not retcode\n with open(BENCHMARK_ENV_BUILT, \"wt\") as f:\n pass\n\n\ndef build_clean(\n checkout,\n build_cfg: BuildCfg,\n show_progress: bool = True,\n taskset_cores: Optional[str] = None,\n nice: Optional[str] = None,\n max_jobs: Optional[int] = None,\n) -> Optional[str]:\n MUTATION_LOCK.get()\n if check_unbuildable(checkout):\n print(f\"{checkout} is known to be unbuildable.\")\n return\n\n conda_env = None\n retcode = 1\n try:\n repo_path = os.path.join(BUILD_IN_PROGRESS_ROOT, f\"pytorch_{uuid.uuid4()}\")\n shutil.copytree(REF_REPO_ROOT, repo_path)\n conda_env = make_conda_env(build_cfg=build_cfg)\n retcode = _build(\n repo_path,\n checkout,\n \"install\",\n conda_env,\n build_cfg,\n show_progress,\n taskset_cores,\n nice,\n max_jobs,\n )\n\n return None if retcode else conda_env\n\n except KeyboardInterrupt:\n print(f\"Build stopped: {checkout}\")\n raise\n\n finally:\n if retcode and conda_env is not None and os.path.exists(conda_env):\n shutil.rmtree(conda_env)\n\n if os.path.exists(repo_path):\n shutil.rmtree(repo_path)\n","sub_path":"v2/build.py","file_name":"build.py","file_ext":"py","file_size_in_byte":9455,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"}
+{"seq_id":"74731744","text":"import numpy as np\nfrom matplotlib import pyplot as plt\n\nx = np.linspace(-3, 3, 200)\n\nfig, ax = plt.subplots()\n\nfor a in range(1, 11):\n y = x*(x + 2)*(x - 2)/a\n ax.plot(x, y)\n fig.savefig('мой график' + str(a))\n\nplt.show()","sub_path":"python/modeling/plottrial.py","file_name":"plottrial.py","file_ext":"py","file_size_in_byte":240,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"}
+{"seq_id":"188684122","text":"#Taher Dohadwala\n\n#imports\nimport speech_recognition as sr\nimport pyaudio,os,time\nfrom gtts import gTTS\nimport requests,json\nimport wolframalpha\n#------------------------------------------------------------------\n# Rules for adding new commands:\n# -Create a new method in AI Functions\n# -Add the keyword to the variable key_word in the __init__\n# -Add any api keys to the __init__\n# -Add the keyword and method to the commands dictionary\n\n\n#------------------------------------------------------\n#Things to do:\n# 1. use espeak as the voice\n# 2. add description from the weather api\n# 3. add a run() to constantly run the program waiting for keyword (NAME)\n# 4. use amazon cloud for what is command\n# 5. add spotify api\n# 6. add various other command...\nclass speech_program():\n \n\n #-----------------------------------------------------------------\n #init()\n def __init__(self):\n self.key_words = [\"weather\",\"what is\",\"end\"]\n self.weather_key = \"APPID=714964c165954619971a23be696deb1a\"\n self.wolfram_key = \"8A86L7-EQW9P7T464\"\n \n #--------------------------------------------------------------------\n #AI functions-- 1 function = 1 command\n\n def weather(self):\n if \"in\" in self.save_audio:\n self.location = self.save_audio[self.save_audio.find(\"in\")+3:]\n else:\n self.location = \"Naperville\"\n response = requests.get(\"http://api.openweathermap.org/data/2.5/weather?q=\"+self.location+\",us&units=imperial&\"+self.weather_key).json() \n a2 = \"The Temperature in\"+self.location+\" is \"+str(response[\"main\"][\"temp\"])+\" degrees fahrenheit\"\n tts = gTTS(text = a2)\n tts.save(\"sample2.mp3\")\n\n os.system(\"sample2.mp3\")\n\n\n def what_is(self):\n client = wolframalpha.Client(self.wolfram_key)\n res = client.query(self.save_audio)\n a2 = self.save_audio+\" the answer is \"+str((next(res.results).text))\n tts = gTTS(text = a2)\n tts.save(\"sample2.mp3\")\n\n os.system(\"sample2.mp3\")\n\n\n def end(self):\n tts = gTTS(text = \"System shutting down\")\n tts.save(\"sample2.mp3\")\n\n os.system(\"sample2.mp3\")\n #--------------------------------------------------------------------\n # AI brain\n def listen(self): \n with sr.Microphone() as source:\n print(\"Say something!\")\n audio = sr.Recognizer().listen(source)\n while audio == None:\n audio = sr.Recognizer().listen(source)\n\n try:\n self.save_audio = sr.Recognizer().recognize_google(audio)\n print (\"You said: \"+ self.save_audio)\n\n except sr.UnknownValueError:\n print(\"Google Speech Recognition could not understand audio\")\n speech_program.listen(self)\n\n except sr.RequestError as e:\n\n print(\"Could not request results from Google Speech Recognition service; {0}\".format(e))\n\n def find_keyword(self):\n #Checks if a key_word is in user said sentence\n for x in range(len(self.key_words)):\n if self.key_words[x] in self.save_audio:\n self.key_word = self.key_words[x]\n print (\"keyword finder worked\")\n break\n ############################# \n commands = {\"weather\":weather,\n \"what is\":what_is,\n \"end\":end}\n #############################\n \n def compute(self):\n\n # using key_word - made in find_keyword\n # matches the key_word to respective key and executes the key value(AI function)\n for x in self.commands:\n if self.key_word in self.save_audio:\n self.commands.get(self.key_word)(self)\n print (\"reached command\")\n break\n#---------------------------------------------------------------------------------------\n\n\ntester = speech_program()\n\ntester.listen()\ntester.find_keyword()\ntester.compute()\ninput\n \n \n","sub_path":"AI-Project/Project Testing/speech_program3.0.py","file_name":"speech_program3.0.py","file_ext":"py","file_size_in_byte":3944,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"}
+{"seq_id":"561360875","text":"# Import everything needed to edit video clips \nfrom moviepy.editor import *\n\nimg = ['output-raspberry-3.jpg']\nclips = [ImageClip(m).set_duration(2)\n for m in img]\nclip1 = concatenate_videoclips(clips, method=\"compose\")\n# loading video dsa gfg intro video \n#clip1 = VideoFileClip(\"output-raspberry-3.mp4\")\nclip2 = VideoFileClip(\"test1.mp4\")\nclip3 = VideoFileClip(\"test2.mp4\")\n#clip4 = VideoFileClip(\"raspberry-influx-temperature-ok-recording-data.mp4\")\n\n\n# speedup clip2\n#clip2_speed= clip2.fx(vfx.speedx, 2)\nclip3_speed= clip3.fx(vfx.speedx, 2)\n\nfinal = concatenate_videoclips([clip1, clip2, clip3_speed], method=\"compose\") \n\n#concat_clip = concatenate_videoclips(clips, method=\"compose\")\nfinal.write_videofile(\"output-rasp-3.mp4\", fps=24)\n","sub_path":"concatenate.py","file_name":"concatenate.py","file_ext":"py","file_size_in_byte":746,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"}
+{"seq_id":"30247660","text":"\"\"\" Assignment two in the Cteate a Blockchain with Python course. \"\"\"\n\n\nNAMES = list(('Aegir', 'Rosa', 'Hafrun', 'Hronn', 'Oli', 'Thorvaldur', 'Ada'))\n\n\ndef names_is_empty():\n \"\"\" Returns if the NAMES list is empty or not. \"\"\"\n return len(NAMES) < 1\n\n\nprint('Names longer than 5 - contain \\'n\\' or \\'N\\'?')\n\nfor name in NAMES:\n is_n = False\n if len(name) > 5:\n temp = name.lower()\n for letter in temp:\n if letter == 'n':\n is_n = True\n if is_n:\n print('YES - %s' % name)\n else:\n print('NO - %s' % name)\n\nprint('-' * 40)\n\nwhile not names_is_empty():\n print(NAMES.pop())\n","sub_path":"udemy_blockchain/assignment02/two.py","file_name":"two.py","file_ext":"py","file_size_in_byte":659,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"}
+{"seq_id":"15185487","text":"#!/usr/bin/python\n\nimport os\nimport gtk\nimport pygtk\n\nclass ImageView:\n def __init__(self):\n #self.aspect_frame = gtk.Frame(None)\n\n self.scrolledwindow = gtk.ScrolledWindow()\n self.scrolledwindow.set_policy(gtk.POLICY_NEVER,gtk.POLICY_NEVER)\n\n self.image = gtk.Image()\n self.pixbuf = self.image.get_pixbuf()\n self.image.connect('expose-event', self.on_image_resize, self.scrolledwindow)\n\n self.scrolledwindow.add(self.image)\n\n #self.aspect_frame.add(self.scrolledwindow)\n\n def on_image_resize(self, widget, event, window):\n allocation = self.scrolledwindow.get_allocation()\n\n x, y = self.scale(self.pixbuf.get_width(), self.pixbuf.get_height(), allocation.width, allocation.height)\n pixbuf = self.pixbuf.scale_simple(x, y, gtk.gdk.INTERP_BILINEAR)\n self.image.set_from_pixbuf(pixbuf)\n\n def change_path(self, path):\n self.directory = path\n self.photos = os.listdir(path)\n self.photos.sort()\n self.position = 0\n self.change_photo(None, 0)\n\n def change_photo(self, button, direction):\n self.position = (self.position + direction) % len(self.photos)\n\n self.image.set_from_file(os.path.join(self.directory, self.photos[self.position]))\n self.pixbuf = self.image.get_pixbuf()\n #ratio = self.find_ratio(self.image.get_allocation().width, self.image.get_allocation().height)\n #self.aspect_frame.set(0.5, 0.5, ratio, False)\n\n def get_widget(self):\n return self.scrolledwindow#self.aspect_frame\n\n def scale(self, w, h, x, y, maximum=True):\n nw = y * w / h\n nh = x * h / w\n if maximum ^ (nw >= x):\n return nw or 1, y\n return x, nh or 1\n\n def find_ratio(self, a, b):\n if b == 0:\n return a\n return self.find_ratio(b, a % b)\n\n\n\nif __name__ == '__main__':\n window = gtk.Window()\n\n photo = ImageView()\n photo.change_path('/home/ben/Pictures/Testimages')\n\n window.add(photo.get_widget())\n window.show_all()\n gtk.main()\n","sub_path":"ImageView.py","file_name":"ImageView.py","file_ext":"py","file_size_in_byte":2067,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"}
+{"seq_id":"66841460","text":"from django.conf.urls import patterns, include, url\nfrom Scheduler.forms import SchedulesForm\nfrom Indigo7.views import SimpleViewListView, SimpleViewDetailView, SimpleViewCreateView, SimpleViewDeleteView, SimpleViewUpdateView\nfrom Scheduler.models import Schedules\n\n\nurlpatterns = patterns('Scheduler.views', \n url(r'^scheduleddays/', include('Scheduler.urlconf.scheduleddays')),\n \n url(r'^$',\n SimpleViewListView.as_view(model=Schedules), \n name='schedules_list'),\n \n url(r'^create/$',\n SimpleViewCreateView.as_view(model=Schedules, \n form_class=SchedulesForm, \n success_url='schedules_list'), \n name='create_schedules'), \n \n url(r'^update/(?P\" + i[0] + \"\\n\"\n text = text + \"\\nUse #notename to view the note\"\n\n update.message.reply_text(\n text=text, parse_mode=\"HTML\", disable_web_page_preview=True)\n return\n except:\n update.message.reply_text(\n text=\"Notes not available..\", parse_mode=\"HTML\", disable_web_page_preview=True)\n return\n\n try:\n text_2 = res[2]\n except:\n text_2 = \"\"\n\n m = extract.sudocheck(update, context)\n if m == 2:\n return\n # ss\n if text_1 == 'remove':\n if text_2 != \"\":\n chat_idd = str(chat_id)[1:]\n user_id = str(msg.from_user.id)\n\n u = push_note(chat_id=chat_idd, note_name=text_2, pop=1)\n if u == 1:\n text = \"Note #\" + text_2 + \" deleted !\"\n elif u == -2:\n text = \"Noet #\" + text_2 + \" does not exist !\"\n else:\n text = \"Error !\"\n\n elif text_1 == 'set':\n try:\n text_2 = res[2]\n except:\n update.message.reply_text(\"Note Name & Content not provided !\")\n return\n try:\n text_3 = res[3]\n except:\n update.message.reply_text(\"Note content not provided !\")\n return\n\n chat_idd = str(chat_id)\n chat_idd = chat_idd[1:]\n user_id = str(msg.from_user.id)\n\n push_note(chat_id=chat_idd, note_name=text_2,\n note=text_3, set_by=user_id)\n text = 'Note #' + str(text_2) + ' - \\n\"' + str(text_3) + '\"'\n\n else:\n text = \"Wrong format !\"\n\n update.message.reply_text(text)\n","sub_path":"modules/notes.py","file_name":"notes.py","file_ext":"py","file_size_in_byte":2894,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"188018220","text":"'''\nGiven a collection of distinct integers, return all possible permutations(排列).\n\nExample:\nInput: [1,2,3]\nOutput:\n[\n [1,2,3],\n [1,3,2],\n [2,1,3],\n [2,3,1],\n [3,1,2],\n [3,2,1]\n]\n'''\n\nclass Solution:\n # 耶!一次过!\n # 97.59%\n def permute1(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: List[List[int]]\n \"\"\"\n if len(nums) == 0: return []\n if len(nums) == 1: return [nums]\n if len(nums) == 2: return [[nums[0], nums[1]],[nums[1], nums[0]]]\n res = []\n for i in range(len(nums)):\n cur = nums[i]\n remain = nums.copy()\n remain.remove(cur)\n lastRes = self.permute(remain)\n for r in lastRes:\n res.append([cur]+r)\n\n return list(set(res))\n\n # 99.81%\n def permute2(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: List[List[int]]\n \"\"\"\n from itertools import permutations\n return list(permutations(nums, len(nums)))\n\n # 97.59%\n def permute(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: List[List[int]]\n \"\"\"\n if not nums:\n return []\n\n nums.sort()\n res = [nums[:]]\n n = len(nums)\n i = n - 1\n # 其实就是排序完求全排列\n '''\n 1.首先从最尾端开始往前寻找两个相邻元素,令第一元素为i-1,第二元素为i,且满足i-1 0:\n if nums[i - 1] < nums[i]: # 从后往前循环,找到第一个相邻的递增的位置,i是右边大的那一个,i-1则是左边较小的那个数\n j = n - 1\n while nums[j] < nums[i - 1]: # 再找到从后向前找到要交换的位置\n j -= 1\n nums[i - 1], nums[j] = nums[j], nums[i - 1]\n nums[i:] = sorted(nums[i:]) # 把之后的序列排序\n res.append(nums[:])\n i = n - 1\n else:\n i -= 1\n\n return res\n\nso = Solution()\nnums = [1,2,3,4,5]\nprint(so.permute(nums))\n\n","sub_path":"Algorithm01-50/46_Permutations.py","file_name":"46_Permutations.py","file_ext":"py","file_size_in_byte":2337,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"468387605","text":"from p21 import divisors\n\ndef abundant(n):\n return n < sum(divisors(n))\n\nlimit = 28123 \nabundants = filter(abundant, range(1, limit+1))\n\ndef sumsOfAbundants():\n sums = [a+b for a in abundants\n for b in abundants\n if (b >= a) and (b <= limit - a)]\n return list(set(sums))\n \ndef solution():\n s = sum(range(1, limit+1))\n t = sum(sumsOfAbundants())\n return s - t\n\nif __name__ == \"__main__\":\n print(solution())\n \n","sub_path":"src/p23.py","file_name":"p23.py","file_ext":"py","file_size_in_byte":467,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"10401920","text":"\"\"\"\nSingleton class containing all known personality traits/attributes.\n\"\"\"\n\n\nclass Personality(object):\n def __init__(self):\n \"\"\"\n Describes the five traits of the big five personality model.\n A value of 0 denotes e.g. a very introverted person, a value of 100 e.g. a very extroverted person.\n \"\"\"\n self.openness = 0\n self.conscientiousness = 0\n self.extraversion = 0\n self.agreeableness = 0\n self.neuroticism = 0\n self.dict_all_traits = {\"openness\": self.openness, \"conscientiousness\": self.conscientiousness,\n \"extraversion\": self.extraversion, \"agreeableness\": self.agreeableness,\n \"neuroticism\": self.neuroticism}\n\n # Saves the amount of changes/updates that were made to a trait.\n # This is a measure for how sure we are of that trait's prediction.\n self.changes_made = {\"openness\": 0, \"conscientiousness\": 0, \"extraversion\": 0, \"agreeableness\": 0,\n \"neuroticism\": 0}\n\n def update_dict(self):\n self.dict_all_traits = {\"openness\": self.openness, \"conscientiousness\": self.conscientiousness,\n \"extraversion\": self.extraversion, \"agreeableness\": self.agreeableness,\n \"neuroticism\": self.neuroticism}\n\n","sub_path":"PersonalInformationCollection/Personality.py","file_name":"Personality.py","file_ext":"py","file_size_in_byte":1360,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"12100188","text":"#!/usr/bin/env python3\n\n# 2020暑期排位5K_生日谜题\n# https://codeforces.com/group/H9K9zY8tcT/contest/286081/problem/K\n# bit-operation? 不适合python? 也能做\n# 还能用dfs来做!?\n\nt = input()\nl = list(map(int,input().split())) \n\nnob = 20 #as ai<10^5\nn = len(l)\ncounts = [sum([(1< 313 h-1 Mpc z = 0.087 -> 257 h-1 Mpc\n\n\n################################################################################\n#\n# OPEN FILES\n#\n################################################################################\n\n\ninfile = Table.read(in_filename, format='ascii.commented_header')\nmaskfile = Table.read(mask_filename, format='ascii.commented_header')\n\n\n################################################################################\n#\n# FILTER GALAXIES\n#\n################################################################################\n\n\n'''\ncoord_min_table, mask, ngrid = filter_galaxies(infile, maskfile, min_dist, max_dist, survey_name, True)\n\ntemp_outfile = open(\"filter_galaxies_output.pickle\", 'wb')\npickle.dump((coord_min_table, mask, ngrid), temp_outfile)\ntemp_outfile.close()\n'''\n\n\n\n################################################################################\n#\n# FIND VOIDS\n#\n################################################################################\n\n\ntemp_infile = open(\"filter_galaxies_output.pickle\", 'rb')\ncoord_min_table, mask, ngrid = pickle.load(temp_infile)\ntemp_infile.close()\n\n\nfind_voids(ngrid, min_dist, max_dist, coord_min_table, mask, out1_filename, out2_filename, survey_name)\n","sub_path":"SDSS_VoidFinder.py","file_name":"SDSS_VoidFinder.py","file_ext":"py","file_size_in_byte":2702,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"526733096","text":"from math import sqrt\nfor n in range(1, 60):\n r_org = 2.0\n r = r_org\n for i in range(n):\n r = sqrt(r)\n for i in range(n):\n r = r ** 2\n print ('With {} times sqrt and then {} times **2, the number {} becomes: {:.16f}'.format(n,n,r_org,r))\n\n\n '''\n What this code is doing is take the square root of two and then squaring is again. As it does this, it goes through a for loop square rooting and resquaring up to the specified number of times.\n However, the more times we square, the smaller the round error gets and that's why resquaring doesn't arrive back at our original value.\n '''\n","sub_path":"homework/3/round_off_errors.py","file_name":"round_off_errors.py","file_ext":"py","file_size_in_byte":625,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"350435438","text":"import sys\nD = {}\ninFile = open(sys.argv[1])\nouFile = open(sys.argv[1]+'.stop','w')\nouFile2 = open(sys.argv[1]+'.stop2','w')\nouFile3 = open(sys.argv[1]+'.not.stop','w')\nfor line in inFile:\n fields = line.split(',')\n pep = fields[2]\n if pep[-1]=='K' or pep[-1]=='R':\n uniprot=fields[9].split('|')[1]\n ouFile3.write(pep+'\\t'+uniprot+'\\n')\n elif pep==' Peptide':\n D.setdefault(fields[2], 0)\n D[fields[2]]+=1\ninFile.close()\n\n'''\ndef protein():\n D2 = {}\n inFile = open('/netshare1/home1/people/hansun/StopGainProteomics/2.uniprot/human_uniprot_sprot.fa')\n while True:\n line1 = inFile.readline().strip()\n line2 = inFile.readline().strip()\n if line1:\n fields = line1.split('|')\n name = fields[1]\n D2.setdefault(line2,[])\n D2[line2].append(name)\n else:\n break\n inFile.close()\n return D2\n\nD2=protein()\n\nfor k in D:\n ouFile.write(k+'\\t'+str(D[k])+'\\n')\n ouFile2.write(k+'\\t')\n for x in D2:\n if x[-len(k):]==k:\n ouFile2.write('\\t'.join(D2[x])+'\\t')\n ouFile2.write('\\n')\nouFile2.close()\n'''\n","sub_path":"Project/StopGainProteomics/8.omssa/1-stop.py","file_name":"1-stop.py","file_ext":"py","file_size_in_byte":1147,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"368101664","text":"# Program de citit ziare online\n# de pe mai multe surse precum\n# Adevarul\n# Program realizat de catre:\n# Olariu Alexandru-Razvan\n# Contact: ufcolonel@gmail.com\n\n\nimport Crawler,Scraper\nfrom datetime import datetime\nfrom tkinter import *\nfrom tkinter import ttk\n\n\n\ndef save_status():\n global alegeri\n global interest\n alegeri = [option1.get(),option2.get(),option3.get(),\n option4.get(),option5.get(),option6.get(),\n option7.get(),option8.get(),option9.get()]\n interest = [search.get()]\n \ndef check():\n for source in range(len(alegeri)):\n https = alegeri[source]\n if https.startswith(\"https://\"):\n try:\n https = Crawler.crawl(https,interest)\n except Exception as error:\n with open(\"Reports.txt\",\"w\") as f:\n f.write(str(error))\n finally:\n interest = Scraper.scrape(https)\n\n\ndef save():\n alegeri = [option1.get(),option2.get(),option3.get(),\n option4.get(),option5.get(),option6.get(),\n option7.get(),option8.get(),option9.get()]\n for source in range(len(alegeri)):\n if alegeri[source].startswith(\"https://\"):\n header = f\"Date din sursa {alegeri[source]} :\\n\"\n with open(\"Salvate.txt\",\"a\") as g:\n g.write(datetime.now().strftime(\"\\t%d/%m/%Y <---> %H:%M:%S\\n\"))\n g.write(header)\n g.write(interest)\n\n \n\n# Graphical User Interface initialization - FINISHED\napp = Tk()\napp.title('Ziare online de Olariu Alexandru-Razvan')\napp.geometry('900x500')\napp.resizable(False,False)\n\n\n# Search\nttk.Label(app,text=\"Căutare după termeni:\").place(x=10,y=5)\nsearch = ttk.Entry(app)\nsearch.place(x=150,y=5,width=300,height=23)\n\n# Online News Variables\noption1, option2, option3 = StringVar(), StringVar(), StringVar()\noption4, option5, option6 = StringVar(), StringVar(), StringVar() \noption7, option8, option9 = StringVar(), StringVar(), StringVar() \n\n# Online News\n\nttk.Checkbutton(app,text=\"Ziarul Libertatea\",variable=option1,\n onvalue='https://www.libertatea.ro/',\n command=save_status).place(x=10,y=50)\nttk.Checkbutton(app,text=\"Ziarul Financiar\",variable=option2,\n onvalue='https://www.zf.ro/',\n command=save_status).place(x=10,y=100)\nttk.Checkbutton(app,text=\"Mediafax\",variable=option3,\n onvalue='https://www.mediafax.ro/',\n command=save_status).place(x=160,y=50)\nttk.Checkbutton(app,text=\"Evenimentul Zilei\",variable=option4,\n onvalue='https://evz.ro/',\n command=save_status).place(x=160,y=100)\nttk.Checkbutton(app,text=\"Ziarul Adevarul\",variable=option5,\n onvalue='https://adevarul.ro/cauta',\n command=save_status).place(x=310,y=50)\nttk.Checkbutton(app,text=\"Jurnalul Zilei\",variable=option6,\n onvalue='https://jurnalulnational.ro/',\n command=save_status).place(x=310,y=100)\nttk.Checkbutton(app,text=\"Digi24\",variable=option7,\n onvalue='https://www.digi24.ro/',\n command=save_status).place(x=470,y=50)\nttk.Checkbutton(app,text=\"Realitatea.NET\",variable=option8,\n onvalue='https://www.realitatea.net/',\n command=save_status).place(x=470,y=100)\n\n\n# Buttons\ncheck_button = ttk.Button(app,text=\"Caută\",command=check)\ncheck_button.place(x=470,y=4)\nsave_button = ttk.Button(app,text=\"Salvare\",command=save)\nsave_button.place(x=570,y=4)\n\n\n\n# Output\noutput = ttk.Notebook(app)\nziar1, ziar2, ziar3 = Text(output), Text(output), Text(output)\nziar4, ziar5, ziar6 = Text(output), Text(output), Text(output)\nziar7, ziar8, ziar9 = Text(output), Text(output), Text(output) \noutput.add(ziar1, text=\"Ziarul Libertatea\")\noutput.add(ziar2, text=\"Ziarul Financiar\")\noutput.add(ziar3, text=\"Gazeta Sporturilor\")\noutput.add(ziar4, text=\"Romania Libera\")\noutput.add(ziar5, text=\"Ziarul Adevarul\")\noutput.add(ziar6, text=\"Jurnalul Zilei\")\noutput.add(ziar7, text=\"Digi24\")\noutput.add(ziar8, text=\"Realitatea.NET\")\noutput.add(ziar9, text=\"Ajutor\")\noutput.place(x=0,y=140,width=710,height=360)\n\n\n# Progress TODO\nprogress_bar = ttk.Progressbar(app,orient=HORIZONTAL,length=185,mode='determinate')\nprogress_bar.place(x=711,y=477)\n\n# ADS\n\nad1_image = PhotoImage(file=\"ad1.png\")\nad1 = Label(app,image=ad1_image)\nad1.place(x=707,y=0)\nttk.Label(app,text=\"Reclama TA, AICI!\").place(x=765,y=231)\nad2_image = PhotoImage(file=\"ad2.png\")\nad2 = Label(app,image=ad2_image)\nad2.place(x=707,y=247)\n\n\n\napp.mainloop()","sub_path":"GUInterface.py","file_name":"GUInterface.py","file_ext":"py","file_size_in_byte":4710,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"214647893","text":"import copy\nimport io\nimport sys\nimport warnings\nfrom typing import (\n List, Set, Dict, Any, Optional, Union, Tuple\n\n)\n\nfrom gelidum.collections import frozendict, frozenlist, frozenzet\nfrom gelidum.exceptions import FrozenException\nfrom gelidum.frozen import make_frozen_class, FrozenBase\nfrom gelidum.typing import OnFreezeFuncType, OnUpdateFuncType, T, FrozenType, FrozenList\nfrom gelidum.utils import isbuiltin\n\n\ndef freeze(\n obj: T,\n on_update: Union[str, OnUpdateFuncType] = \"exception\",\n on_freeze: Union[str, OnFreezeFuncType] = \"copy\",\n inplace: Optional[bool] = None,\n ) -> FrozenType:\n\n # inplace argument will be removed from freeze in the next major version (0.5.0)\n if isinstance(inplace, bool):\n warnings.warn(\n DeprecationWarning(\n \"Use of inplace is deprecated and will be removed in next major version (0.5.0)\"\n )\n )\n if inplace:\n on_freeze_func: OnFreezeFuncType = __on_freeze_func(on_freeze=\"inplace\")\n else:\n on_freeze_func: OnFreezeFuncType = __on_freeze_func(on_freeze=\"copy\")\n\n else:\n on_freeze_func: OnFreezeFuncType = __on_freeze_func(on_freeze=on_freeze)\n\n on_update_func: OnUpdateFuncType = __on_update_func(on_update=on_update)\n\n return __freeze(obj=obj, on_update=on_update_func, on_freeze=on_freeze_func)\n\n\ndef __freeze(obj: Any, on_update: OnUpdateFuncType,\n on_freeze: OnFreezeFuncType) -> Any:\n\n if isbuiltin(obj):\n return obj\n\n if isinstance(obj, FrozenBase):\n return obj\n\n class_name = type(obj).__name__\n freeze_func_name = f\"__freeze_{class_name}\"\n this_module = sys.modules[__name__]\n if hasattr(this_module, freeze_func_name):\n freeze_func = getattr(this_module, freeze_func_name)\n return freeze_func(obj, on_update=on_update, on_freeze=on_freeze)\n\n if isinstance(obj, object):\n return __freeze_object(obj, on_update=on_update, on_freeze=on_freeze)\n\n # Actually, this code is unreachable\n raise ValueError(f\"object of type {obj.__class__} not frozen\") # pragma: no cover\n\n\ndef __freeze_bytearray(obj: bytearray, *args, **kwargs) -> bytes: # noqa\n return bytes(obj)\n\n\ndef __freeze_dict(obj: Dict, on_update: OnUpdateFuncType,\n on_freeze: OnFreezeFuncType) -> frozendict:\n def freeze_func(item: Any) -> FrozenType:\n return freeze(item, on_update=on_update, on_freeze=on_freeze)\n return frozendict(obj, freeze_func=freeze_func)\n\n\ndef __freeze_list(obj: List, on_update: OnUpdateFuncType,\n on_freeze: OnFreezeFuncType) -> FrozenList:\n def freeze_func(item: Any) -> FrozenType:\n return freeze(item, on_update=on_update, on_freeze=on_freeze)\n return frozenlist(obj, freeze_func=freeze_func)\n\n\ndef __freeze_tuple(obj: Tuple, on_update: OnUpdateFuncType,\n on_freeze: OnFreezeFuncType) -> Tuple:\n return tuple(freeze(item, on_update=on_update, on_freeze=on_freeze)\n for item in obj)\n\n\ndef __freeze_set(obj: Set, on_update: OnUpdateFuncType,\n on_freeze: OnFreezeFuncType) -> frozenzet:\n def freeze_func(item: Any) -> FrozenType:\n return freeze(item, on_update=on_update, on_freeze=on_freeze)\n return frozenzet(obj, freeze_func=freeze_func)\n\n\ndef __freeze_TextIOWrapper(*args, **kwargs) -> None: # noqa\n raise io.UnsupportedOperation(\"Text file handlers can't be frozen\")\n\n\ndef __freeze_BufferedWriter(*args, **kwargs) -> None: # noqa\n raise io.UnsupportedOperation(\"Binary file handlers can't be frozen\")\n\n\ndef __freeze_object(obj: object, on_update: OnUpdateFuncType,\n on_freeze: OnFreezeFuncType) -> FrozenBase:\n\n if hasattr(obj.__class__, \"__slots__\"):\n raise FrozenException(\"gelidum does not support classes with __slots__\")\n\n frozen_obj = on_freeze(obj)\n for attr, value in frozen_obj.__dict__.items():\n attr_value = getattr(frozen_obj, attr)\n setattr(frozen_obj, attr, freeze(attr_value, on_update=on_update, on_freeze=on_freeze))\n\n frozen_class = make_frozen_class(\n klass=obj.__class__,\n attrs=list(obj.__dict__.keys()),\n on_update=on_update\n )\n frozen_obj.__class__ = frozen_class\n return frozen_obj\n\n\ndef __on_freeze_func(on_freeze: Union[str, OnFreezeFuncType]) -> OnFreezeFuncType:\n if isinstance(on_freeze, str):\n if on_freeze == \"copy\":\n return lambda obj: copy.deepcopy(obj)\n elif on_freeze == \"inplace\":\n return lambda obj: obj\n else:\n raise AttributeError(\n f\"Invalid value for on_freeze parameter, '{on_freeze}' found, \"\n f\"only 'copy' and 'inplace' are valid options if passed a string\"\n )\n\n elif callable(on_freeze):\n return on_freeze\n\n else:\n raise AttributeError(\n f\"Invalid value for on_freeze parameter, '{on_freeze}' found, \"\n f\"only 'copy', 'inplace' or a function are valid options\"\n )\n\n\ndef __on_update_exception(\n frozen_obj: FrozenBase, message: str, *args, **kwargs # noqa\n) -> None:\n raise FrozenException(message)\n\n\ndef __on_update_warning(\n frozen_obj: FrozenBase, message: str, *args, **kwargs # noqa\n) -> None:\n warnings.warn(message)\n\n\ndef __on_update_func(on_update: OnUpdateFuncType) -> OnUpdateFuncType:\n if isinstance(on_update, str):\n if on_update == \"exception\":\n return __on_update_exception\n elif on_update == \"warning\":\n return __on_update_warning\n elif on_update == \"nothing\":\n return lambda message, *args, **kwargs: None\n else:\n raise AttributeError(\n f\"Invalid value for on_update parameter, '{on_update}' found, \"\n f\"only 'exception', 'warning', and 'nothing' are valid options \"\n f\"if passed a string\"\n )\n\n elif callable(on_update):\n return on_update\n\n else:\n raise AttributeError(\n f\"Invalid value for on_update parameter, '{on_update}' found, \"\n f\"only 'exception', 'warning', 'nothing' or a function are \"\n f\"valid options\"\n )\n","sub_path":"gelidum/freeze.py","file_name":"freeze.py","file_ext":"py","file_size_in_byte":6215,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"415403875","text":"import unittest\n\nfrom pesel_generator.exception import PeselGeneratorException\nfrom pesel_generator.pesel import population_description, pesel_generator\n\n\nclass TestPesel(unittest.TestCase):\n\n def test_population_description0(self):\n try:\n with population_description.PopulationDescription() as ep:\n ep.define_population_range(age_start=24, age_stop=25, female_quantity=1200, male_quantity=2400)\n ep.define_population_range(age_start=26, age_stop=28, female_quantity=1200, male_quantity=2400)\n ep.define_population_range(age_start=25, age_stop=26, female_quantity=1200, male_quantity=2400)\n except PeselGeneratorException:\n self.fail()\n\n def test_population_description1(self):\n with self.assertRaises(Exception):\n with population_description.PopulationDescription() as ep:\n ep.define_population_range(age_start=24, age_stop=25, female_quantity=1200, male_quantity=2400)\n ep.define_population_range(age_start=26, age_stop=28, female_quantity=1200, male_quantity=2400)\n ep.define_population_range(age_start=25, age_stop=27, female_quantity=1200, male_quantity=2400)\n\n def test_pesel_generator0(self):\n try:\n with population_description.PopulationDescription() as ep:\n ep.define_population_range(age_start=24, age_stop=25, female_quantity=1200, male_quantity=2400)\n ep.define_population_range(age_start=26, age_stop=28, female_quantity=1200, male_quantity=2400)\n with pesel_generator.PeselGenerator() as pg:\n _ = pg.gen_based_on_desc(ep)\n\n except PeselGeneratorException:\n self.fail()\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"pesel_generator/tests/pesel_generator_tests.py","file_name":"pesel_generator_tests.py","file_ext":"py","file_size_in_byte":1785,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"653692996","text":"# -*- coding: utf-8 -*-\n\"\"\"\n功能:本程序旨在实验出少女前线重装部队芯片强化的最优情况\n作者:史学超\n更新日期:2018.10.23\n现在已经实现了一个简单的模型,需要的是不断额往芯片池中添加芯片\n代码变得更简洁,生成的方案更多,也更加容易出现重复的地方\n\"\"\"\n# TODO(sxc): 添加图像化界面\n# TODO(sxc): 芯片数据要能够附带属性\n# TODO(sxc): 重复结果的删除和结果的评分\n\nimport copy\n\n\nclass Stack:\n def __init__(self):\n self.items = []\n\n def push(self, item):\n self.items.append(item)\n\n def pop(self):\n return self.items.pop()\n\n def clear(self):\n del self.items[:]\n\n def empty(self):\n return self.size() == 0\n\n def size(self):\n return len(self.items)\n\n def top(self):\n return self.items[self.size() - 1]\n\n # this is my add\n def show(self):\n i = 1\n for x in self.items:\n print('this is stack no ', i)\n show_map(x)\n i = i + 1\n\n def rest(self):\n tmp = ''\n for x in self.items:\n tmp = tmp + str(x)\n return tmp\n\n\nmap_stack = Stack() # 用来记录安装芯片后的地图信息\nchip_stack = Stack() # 用来记录安装的芯片在芯片池中的起始序号\nChipList33 = [(0, 0), (0, 1), (0, 2),\n (1, 0), (1, 1), (1, 2)]\nChipList222 = [(0, 0), (0, 1),\n (1, 0), (1, 1),\n (2, 0), (2, 1)]\nChipList6 = [(0, 0), (0, 1), (0, 2), (0, 3), (0, 4), (0, 5)]\nMapLList = [[0, 0, 0, 0, 0, 0],\n [0, 0, 0, 0, 0, 0],\n [0, 0, 0, 0, 0, 0],\n [0, 0, 0, 0, 0, 0],\n [0, 0, 0, 0, 0, 0],\n [0, 0, 0, 0, 0, 0]]\nChipPoolLists = [ChipList33, ChipList222, ChipList6]\nresultList = []\n\n\n# 用于在屏幕上输出map的图像结果\ndef show_map(ml):\n _MapLList = ml\n _resStr = 'MapInfo:\\n'\n _MapWidth = len(MapLList)\n for x in range(_MapWidth):\n for y in range(_MapWidth):\n if _MapLList[x][y] == -1:\n _resStr = _resStr + 'x' + ' '\n else:\n _resStr = _resStr + str(_MapLList[x][y]) + ' '\n _resStr = _resStr + '\\n'\n print(_resStr)\n\n\n# 传入的_chip是上一次插入的芯片序号\n# 返回0的时候是按预设情况正常结束,返回1则异常结束\ndef install_chip(_map, _chip=-1):\n # 用于中断递归的条件\n if map_stack.empty() and _chip == (len(ChipPoolLists) - 1):\n print('install_chip is completed')\n return\n if _chip != (len(ChipPoolLists) - 1):\n chip_start = _chip + 1 # 记录本次芯片池中插入芯片的起始序号\n maplist = copy.deepcopy(_map)\n _MapWidth = len(maplist)\n for x in range(_MapWidth):\n for y in range(_MapWidth):\n # 这边得到的xy是可插入点坐标,下面的Z是芯片池中的序号\n if maplist[x][y] == 0:\n for z in range(chip_start, len(ChipPoolLists)):\n chiplist = ChipPoolLists[z]\n if insertable(x, y, chiplist, maplist):\n no = map_stack.size() + 1 # 插入位置的数值,用来区分插入的是哪一块芯片\n maplist = inserted(x, y, chiplist, no, maplist)\n chip_start = 0\n # show_map(maplist)\n map_stack.push(maplist)\n chip_stack.push(z)\n # print('==============================')\n break\n else:\n continue\n if isfulll(maplist):\n resultList.append(maplist)\n show_map(maplist)\n print(\"+++有一个方案已经完成+++\")\n map_stack.pop()\n tempc1 = chip_stack.pop()\n if map_stack.empty():\n tempm1 = copy.deepcopy(MapLList)\n else:\n tempm1 = map_stack.top()\n install_chip(tempm1, tempc1)\n return\n\n\ndef insertable(x, y, _chiplist, _map):\n maplist = _map\n for g in _chiplist:\n xp, yp = g\n try:\n gno = maplist[x + xp][y + yp]\n except IndexError:\n return False\n finally:\n pass\n if gno != 0:\n return False\n return True\n\n\n# xy表示插入起始点的坐标,_chiplist表示插入芯片的类型,\n# _no表示插入的数值, _map表示被插入的map信息\n# 返回插入后的map\ndef inserted(x, y, _chiplist, _no, _map):\n maplist = copy.deepcopy(_map)\n for g in _chiplist:\n xp, yp = g\n maplist[x + xp][y + yp] = _no\n # print(_chiplist, 'is inserted')\n return maplist\n\n\ndef rotate90(_chiplist, _time=0):\n list_ = []\n if _time == 0:\n list_ = _chiplist\n return list_\n else:\n for t in range(_time):\n list_ = []\n for item in _chiplist:\n x, y = item\n list_.append((-y, x))\n return list_\n\n\n# 返回False为不满,代表未完全插入;返回True为满,已经完全插入.\ndef isfulll(_map):\n maplist = copy.deepcopy(_map)\n _MapWidth = len(maplist)\n for x in range(_MapWidth):\n for y in range(_MapWidth):\n if _map[x][y] == 0:\n return False\n return True\n\n\ndef writefile(_aimlist):\n with open('./resultList.txt', 'w') as f:\n for x in _aimlist:\n f.write(str(x) + '\\n==========\\n')\n print('已经成功写入文件到', 'resultList.txt')\n\n\nif __name__ == \"__main__\":\n # map_stack.push(copy.deepcopy(MapLList))\n # install_chip(MapLList, -1)\n # writefile(resultList)\n print('总共有', len(resultList), '个结果')\n print('the process is end')\n","sub_path":"GFHC0.py","file_name":"GFHC0.py","file_ext":"py","file_size_in_byte":5741,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"}
+{"seq_id":"153364219","text":"# 学习使用lambda函数\n# 原定义的函数\n\"\"\"\ndef true():\n return True\ntrue()\n\"\"\"\n\n# 进一步简化,两行写为一行\n\"\"\"\ndef true():return True\ntrue()\n\"\"\"\n# 如果把函数名也简化了,那就是用lambda\n\"\"\"\ndef true():return True\nlambda : True\n\"\"\"\n\"\"\"\nlambda函数:\ndef add(x,y):\n return x+y\nadd(3,5)\n8\nlambda x,y: x+y\n| ⏳ | '.format(\n data.feedidfromroute[route], data.stopidfromstop[stop + 'N'],\n route))\n for i in range(width - len(row)):\n print('')\n print(' |
| {} | '.format(tdclass, route))\n for i in range(width - len(row)):\n print('')\n print(' |
| ⏳ | '.format(\n data.feedidfromroute[route], data.stopidfromstop[stop + 'S'],\n route))\n for i in range(width - len(row)):\n print('')\n print(' |
| No. | \")\n\t\tresult.write(\"Team Name | \")\n\t\tresult.write(\"Institute Name | \")\n\t\tresult.write(\"Member1 | \")\n\t\tresult.write(\"Member2 | \")\n\t\tresult.write(\"Member3 | \")\n\t\tresult.write(\"Rating | \")\n\t\tresult.write(\"Location | \")\n\t\tresult.write(\"
|---|---|---|---|---|---|---|---|
| {} | \" .format(i + 1))\n\t\t\tresult.write(\"{} | \" .format(teamList[i]['name']))\n\t\t\tresult.write(\"{} | \" .format(teamList[i]['inst']))\n\t\t\tresult.write(\"{} | \" .format(teamList[i]['mem1']['han']))\n\t\t\tresult.write(\"{} | \" .format(teamList[i]['mem2']['han']))\n\t\t\tresult.write(\"{} | \" .format(teamList[i]['mem3']['han']))\n\t\t\tresult.write(\"{} | \" .format(teamList[i]['rat']))\n\t\t\tresult.write(\"{} | \" .format(teamList[i]['loc']))\n\n\t\t# close table\n\t\tresult.write(\" \")\n\t\tresult.write(\"
оплачено
'\n\t\t\telif self.check:\n\t\t\t\treturn 'отлконено
'\n\t\t\telse:\n\t\t\t\treturn 'error'","sub_path":"cupons/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":6727,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"233529986","text":"# dummy easy\nanswer = 0\nanswered = False\nfor a in range(1,999):\n for b in range(2, 999):\n c = (a*a + b*b)**(1/2)\n if c - int(c) == 0:\n if a + b + c == 1000:\n answer = a*b*c\n answered = True\n else:\n continue\n else:\n continue\n if answered == True:\n break\nprint(int(answer))\n","sub_path":"005 Percent Problems/009 - Special Pythagorean Triplet/Problem9.py","file_name":"Problem9.py","file_ext":"py","file_size_in_byte":384,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"83839814","text":"import os\nimport shutil\nimport random\n\n# CSV_DIR_NO_TRANSPORT = 'C:\\\\Users\\\\ucfaalf\\\\Documents\\\\Projects\\\\AcousticAnalysis\\\\2013Random\\\\LabelsCSV\\\\'\n# CSV_DIR_TRANSPORT24kHz = 'C:\\\\Users\\\\ucfaalf\\\\Documents\\\\Projects\\\\AcousticAnalysis\\\\2013Random\\\\LabelsCSVTransport\\\\24000HzSR'\n# CSV_DIR_TRANSPORT41kHz = 'C:\\\\Users\\\\ucfaalf\\\\Documents\\\\Projects\\\\AcousticAnalysis\\\\2013Random\\\\LabelsCSVTransport\\\\41000HzSR'\nCSV_DIR = 'C:\\\\Users\\\\ucfaalf\\\\Dropbox\\\\EngD\\\\Data\\\\2013Random\\\\allLabelFiles\\\\'\nWAV_DIR = 'C:\\\\Users\\\\ucfaalf\\\\Documents\\\\Projects\\\\AcousticAnalysis\\\\2013Random\\\\Amalgamated_Files_24kHz\\\\'\n\n# csvFilesNoTransport = os.listdir(CSV_DIR_NO_TRANSPORT)\n# csvFilesTransport24kHz = os.listdir(CSV_DIR_TRANSPORT24kHz)\n# csvFilesTransport41kHz = os.listdir(CSV_DIR_TRANSPORT41kHz)\ncsvFiles = os.listdir(CSV_DIR)\nwavFiles = os.listdir(WAV_DIR)\n\ncsvFileList = []\nfor csvFile in csvFiles:\n csvBaseName = csvFile[:-14]\n csvFileList.append(csvBaseName)\n\n# for csvFile in csvFilesTransport24kHz:\n# csvBaseName = csvFile[:-14]\n# csvFileList.append(csvBaseName) \n\n# for csvFile in csvFilesTransport41kHz:\n# csvBaseName = csvFile[:-14]\n# csvFileList.append(csvBaseName)\n\n# csvFileList = list(set(csvFileList)) \n \nwavFileList = []\nfor wavFile in wavFiles:\n wavBaseName = wavFile[:-4]\n wavFileList.append(wavBaseName)\n\nwavWOcsv = []\nfor wav in wavFileList:\n if wav not in csvFileList:\n wavWOcsv.append(wav)\n\n# verityLabels = []\n# for wavFile in wavWOcsv:\n# if wavFile[:3] in ('BR4', 'BR2', 'IG6', 'E29', 'TW7', 'BR6', 'DA5', 'RM4', 'CM1'):\n# verityLabels.append(wavFile)\n\n# print verityLabels\n\n# dirExtention = \"C:\\\\Users\\\\ucfaalf\\\\Documents\\\\Projects\\\\AcousticAnalysis\\\\2013Random\\\\Amalgamated_Files_24kHz\\\\\"\n# fileList = os.listdir(WAV_DIR)\n\n# wav2FileList = []\n# for wav2 in fileList:\n# \twav2BaseName = wav2[:-4]\n# \twav2FileList.append(wav2BaseName)\n\n# fileListBlankFiles = [x for x in wavFileList if x in verityLabels]\n\n# print len(fileListBlankFiles)\n\n# randomNumbers = random.sample(xrange(len(fileListBlankFiles)), 2)\n# fileSelection = [fileListBlankFiles[i] for i in randomNumbers]\n\n# print fileSelection\n\n# for i in fileSelection:\n# shutil.copy(WAV_DIR + i + '.wav', \"C:\\\\Users\\\\ucfaalf\\\\Dropbox\\\\EngD\\\\Projects\\\\Chapter 3\\\\goldenTestSet\\\\21VerityFiles\")\n\nprint(wavWOcsv)","sub_path":"findBlankSoundFiles.py","file_name":"findBlankSoundFiles.py","file_ext":"py","file_size_in_byte":2325,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"414951853","text":"from bs4 import BeautifulSoup\nimport requests\nimport time\nimport os\n\ndef find_recipes():\n site = 'https://akispetretzikis.com'\n\n # create new directory for the results\n parent_dir = os.path.dirname(os.path.abspath(\"top_level_file.txt\"))\n # Directory\n directory = \"Matched Recipes\"\n\n # Path\n path = os.path.join(parent_dir, directory)\n os.mkdir(path)\n\n print(\"Search the website: akispetretzikis.com \")\n print(\"Insert up to 3 ingredients to filter all the recipes\")\n\n\n first_ing=''\n second_ing=''\n while(first_ing == ''):\n first_ing= input(\"Give the main ingredient\\n>\")\n while( second_ing == ''):\n second_ing = input('Give a second ingredient\\n>')\n third_ing = input('Give a third ingredient if you want\\n>')\n\n\n search = '&search=' + first_ing\n page_counter = 1\n\n\n\n url='https://akispetretzikis.com/en/search?from=admin' + '&page=' +str(page_counter)+ search+'&utf8=%E2%9C%93'\n html_text=requests.get(url).text\n soup = BeautifulSoup(html_text,'lxml')\n show_more_button=soup.find(id='next_page_link')\n\n\n\n\n if(show_more_button is None):\n print('NO RESULTS FOR THE MAIN INGREDIENT')\n\n while (show_more_button is not None):\n print(\"searching...\")\n texts = soup.find_all('div', class_='texts')\n for i in texts:\n\n recipe_url = site + i.find('a')['href']\n\n recipe_html_text = requests.get(recipe_url).text\n recipe_soup = BeautifulSoup(recipe_html_text, 'lxml')\n\n ingredients = recipe_soup.find('div', class_='text ingredients-list')\n # check if other ingrediens are in the same recipe\n if (((second_ing and third_ing) in str(ingredients)) is True):\n\n # Recipe url\n\n # for the Recipe Name\n recipe_name = i.find('h4').text\n\n # Time till we feast\n hands_on_time = recipe_soup.find('ul', class_='new-times').find('h5').text\n\n # print ingredients\n ing_li = ingredients.find_all(['li', 'p'])\n\n\n # for the recipe method\n method_box = recipe_soup.find('div', class_='method')\n method_list = method_box.find_all('li')\n\n\n\n # select directory\n\n\n\n with open(f'Matched Recipes/{recipe_name}.txt', 'w',encoding='utf-8') as f:\n f.write('Recipe link\\t'+recipe_url+'\\n')\n f.write(recipe_name+'\\n')\n f.write('Hands on Time:\\t'+hands_on_time+'\\n')\n f.write('Ingredients\\n')\n for i in ing_li:\n f.write(i.text+'\\n')\n #f.write('\\n')\n f.write('Ingredients\\n')\n for m in method_list:\n f.write(m.text+'\\n')\n\n\n else:\n continue\n\n\n\n\n\n\n #increments for while\n page_counter = page_counter + 1\n url = 'https://akispetretzikis.com/en/search?from=admin' + '&page=' + str(page_counter) + search + '&utf8=%E2%9C%93'\n html_text = requests.get(url).text\n soup = BeautifulSoup(html_text, 'lxml')\n show_more_button = soup.find(id='next_page_link')\n\n\n\nif __name__ == '__main__':\n\n\n\n find_recipes()\n print(\"Search completed\\nOpen Matched Recipes directory to view them\")\n print(\"Goodbye!\")\n","sub_path":"recipe_scrapper.py","file_name":"recipe_scrapper.py","file_ext":"py","file_size_in_byte":3377,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"575921333","text":"import os\nimport datetime\nimport sqlite3\n\n\nfrom sqlite3 import Error\nfrom .get_json import get_json_data\nfrom .settings import BASE_DIR\n\n\ndef decdeg2dms(dd):\n negative = dd < 0\n dd = abs(dd)\n minutes, seconds = divmod(dd*3600, 60)\n degrees, minutes = divmod(minutes, 60)\n if negative:\n if degrees > 0:\n degrees = -degrees\n elif minutes > 0:\n minutes = -minutes\n else:\n seconds = -seconds\n return (degrees, minutes, seconds)\n\n\ndef create_db_connection():\n conn = None\n\n try:\n conn = sqlite3.connect(os.path.join(BASE_DIR, 'db.sqlite3'))\n except Error as e:\n print(e)\n\n return conn\n\n\ndef insert_eathquake(sql_object):\n conn = create_db_connection()\n\n sql_string = '''INSERT INTO quake_eathquake(session_id,src,id_eathquake,version\n ,eathquake_time,lat,lng,mag,depth,nst,region,data_source,create_date,url,lat_deg,lng_deg)\n VALUES(?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?)'''\n\n try:\n with conn:\n sql_cursor = conn.cursor()\n sql_cursor.execute(sql_string, sql_object)\n except Error as e:\n print(e)\n\n # return sql_cursor.lastrowid\n\n\ndef create_row(feature, session_id):\n\n properties = feature['properties']\n geometry = feature['geometry']\n coordinates = geometry['coordinates']\n\n local_date_time = datetime.datetime.strptime(\n '1970-01-01' + ' 0:0:0.0', \"%Y-%m-%d %H:%M:%S.%f\")\n local_date_time = local_date_time + \\\n datetime.timedelta(milliseconds=properties['time'])\n\n lat_deg = decdeg2dms(coordinates[0])\n lng_deg = decdeg2dms(coordinates[1])\n\n sql_object = (session_id,\n properties['sources'],\n feature['id'],\n '1',\n local_date_time,\n coordinates[0],\n coordinates[1],\n properties['mag'],\n coordinates[2],\n properties['nst'],\n properties['place'],\n 'usgs-gov',\n datetime.datetime.now(),\n properties['url'],\n str(lat_deg[0]).replace('.0','') + \"°\" + str(lat_deg[1]).replace('.0','') +\n \"'\" + str(round(lat_deg[2], 0)).replace('.0','') + \"''\",\n str(lng_deg[0]).replace('.0','') + \"°\" + str(lng_deg[1]).replace('.0','') +\n \"'\" + str(round(lng_deg[2], 0)).replace('.0','') + \"''\",\n )\n\n row_id = insert_eathquake(sql_object)\n","sub_path":"quake/quake/sql_db.py","file_name":"sql_db.py","file_ext":"py","file_size_in_byte":2550,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"621355142","text":"import sys\nimport unicodedata\n\n\ndef buscar(*palavras_chave):\n \"\"\" Busca por caracteres que contenham a palavra chave em seu nome.\n Ex:\n >>> from exercicios.buscador import buscar\n >>> for caracter, nome in sorted(buscar('BLACK', 'suit')):\n ... print(caracter, nome)\n ...\n ♠ BLACK SPADE SUIT\n ♣ BLACK CLUB SUIT\n ♥ BLACK HEART SUIT\n ♦ BLACK DIAMOND SUIT\n >>> for caracter, nome in sorted(buscar('suit')):\n ... print(caracter, nome)\n ...\n ♠ BLACK SPADE SUIT\n ♡ WHITE HEART SUIT\n ♢ WHITE DIAMOND SUIT\n ♣ BLACK CLUB SUIT\n ♤ WHITE SPADE SUIT\n ♥ BLACK HEART SUIT\n ♦ BLACK DIAMOND SUIT\n ♧ WHITE CLUB SUIT\n 🕴 MAN IN BUSINESS SUIT LEVITATING\n >>> dict(buscar('BlAcK', 'suit', 'ClUb'))\n {'♣': 'BLACK CLUB SUIT'}\n >>> for caracter, nome in sorted(buscar('chess', 'king')):\n ... print(caracter, nome)\n ...\n ♔ WHITE CHESS KING\n ♚ BLACK CHESS KING\n :param palavras_chave: tupla de strings\n :return: generator onde cada elemento é uma tupla. O primeiro elemento da\n tupla é o caracter e o segundo é seu nome. Assim ele pode ser utilizado no\n construtor de um dicionário\n \"\"\"\n limite = 0\n max_unicode_value = sys.maxunicode\n palavras_upper = [palavra.upper() for palavra in palavras_chave]\n while limite < max_unicode_value:\n caracter = chr(limite)\n try:\n unicode_name_upper = unicodedata.name(caracter).upper().split()\n except ValueError:\n pass\n else:\n if all(\n palavra in unicode_name_upper\n for palavra in palavras_upper\n ):\n yield (caracter, \" \".join(unicode_name_upper))\n finally:\n limite += 1\n\n\nif __name__ == '__main__':\n print(dict(buscar('white', 'suit')))\n","sub_path":"exercicios/buscador.py","file_name":"buscador.py","file_ext":"py","file_size_in_byte":1858,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"278392700","text":"#!/usr/bin/env python\n\nimport numpy as np\nimport os,argparse\nimport glob\n\ndef read_star_by_line(STARFILE,MRCFILE,OFILE):\n this_file = open(STARFILE,'r')\n these_lines=this_file.readlines()\n for line in these_lines:\n #For lines that don't begin with the \"new line\" character\n if len(line) > 1:\n this_line = line.split()\n if not this_line[0][0].isnumeric():\n continue\n else:\n this_x = this_line[0]\n this_y = this_line[1]\n ofile.write(this_x+'\\t'+this_y+'\\t')\n ofile.write(MRCFILE+'\\n')\n #If line begins with \"newline\" character \n else:\n continue\n\n\n#Command Line Options\nparser = argparse.ArgumentParser()\nparser.add_argument(\"-sf\",help='Full path to the folder containing the .star files that will be imported to cryosparc',required=True)\nparser.add_argument(\"-mf\",help='Full path to folder containing micrographs to which the .star files should be linked',required=True)\nparser.add_argument(\"-o\",help='Path & name for the output (combined) .star file for importing to cryosparc')\nargs = parser.parse_args()\n\n#Get list of star files before we generate the new star file\nstar_list = glob.glob(os.path.join(args.sf,\"*.star\"))\n\n#Write header information for concatenated .star file\nofile = open(args.o,'w')\nofile.write(\"\\ndata_\\n\\nloop_\\n_rlnCoordinateX #1\\n_rlnCoordinateY #2\\n_rlnMicrographName #3\\n\")\n\n#Add an MRC file to each star's particle line\nfor STARFILE in star_list:\n STARBASE = os.path.split(STARFILE)[1]\n MRCBASE = STARBASE[:-5] +\".mrc\"\n MRCFULL = os.path.join(args.mf,MRCBASE)\n read_star_by_line(STARFILE,MRCFULL,ofile)\n\n#Close and write the new .star file\nofile.close()\n","sub_path":"cryolo_to_cryosparc/convert_stars_to_csparc.py","file_name":"convert_stars_to_csparc.py","file_ext":"py","file_size_in_byte":1764,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"146124252","text":"# -*- coding:utf-8 -*-\n\nimport re\nimport pandas as pd\nfrom nltk.corpus import stopwords\nfrom bs4 import BeautifulSoup\nimport nltk.data\n\ntrain = pd.read_csv(\"./IMDB/labeledTrainData.tsv\", delimiter = \"\\t\")\ntest = pd.read_csv(\"./IMDB/testData.tsv\", delimiter = \"\\t\")\n# print(train.head())\n# print(test.head())\n\nunlabeled_train = pd.read_csv(\"./IMDB/unlabeledTrainData.tsv\",\n\tdelimiter = \"\\t\", quoting = 3)\n\ntokenizer = nltk.data.load(\"tokenizers/punkt/english.pickle\")\n\ndef review_to_text(review, remove_stopwords):\n\t\"\"\"\n\treview: type str\n\tremove_stopwords: type boolean\n\t\"\"\"\n\traw_text = BeautifulSoup(review, \"html\").get_text()\n\tletters = re.sub(\"[^a-zA-Z]\", \" \", raw_text)\n\twords = letters.lower().split()\n\tif remove_stopwords:\n\t\tstop_words = set(stopwords.words(\"english\"))\n\t\twords = [w for w in words if w not in stop_words]\n\treturn words\n\ndef review_to_sentences(review, tokenizer):\n\traw_sentences = tokenizer.tokenize(review.strip())\n\tsentences = []\n\tfor raw_sentence in raw_sentences:\n\t\tif len(raw_sentence) > 0:\n\t\t\tsentences.append(review_to_text(raw_sentence, False))\n\treturn sentences\n\ncorpora = []\nfor review in unlabeled_train[\"review\"]:\n\tcorpora += review_to_sentences(review.decode(\"utf-8\"), tokenizer)\n\nnum_features = 300 # word vectors dimension\nmin_word_count = 20\nnum_workers = 2\ncontext = 10\ndownsampling = 1e-3\n\nfrom gensim.models import word2vec\nprint(\"Training word2vec model...\")\n\nmodel = word2vec.Word2Vec(corpora, workers = num_workers,\n\tsize = num_features, min_count = min_word_count,\n\twindow = context, sample = downsampling)\n\nmodel.init_sims(replace = True)\n\nmodel.save(\"./IMDB/model.sav\")\n\nfrom gensim.models import Word2Vec\nmodel = Word2Vec.load(\"./IMDB/model.sav\")\nmodel.most_similar(\"man\")","sub_path":"kaggle_competitions/IMDB_egs2.py","file_name":"IMDB_egs2.py","file_ext":"py","file_size_in_byte":1721,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"537598861","text":"'''\nGiven an unsorted array of integers, find the length of the longest consecutive\nelements sequence.\n\nFor example, given [100,4,200,1,3,2], the longest consecutive elements sequence\nshould be [1,2,3,4]. It's length is 4\n\nYour algorithm shouuld run in O(n) complexity\n'''\n\n\n'''\nApproach #1 Brute Force [Time Limit Exceeded]\nIntuition\n\nBecause a sequence could start at any number in `nums`, we can exhaust the entire search \nspace by building as long a sequence as possible from every number.\n\nAlgorithm\n\nThe brute force algorithm does not do anything clever - it just considers each number \nin nums, attempting to count as high as possible from that number using \nonly numbers in nums. After it counts too high (i.e. currentNum refers to a number that\nnums does not contain), it records the length of the sequence if it is larger than the\ncurrent best. The algorithm is necessarily optimal because it explores every possibility.\n\nComplexity Anaylsis:\n\nTime complexity O(n^3):\n The outer loop runs exactly n times, and because currentNum increments by 1 during\n each iteration of the `while` loop, it runs in O(n) time. Then, on each iteration of the\n `while` loop, an O(n) lookup in the array is performed. Therefore, this brute force\n algorithm is really three nested O(n) loops, which compound multiplicatively to a cubic\n runtime.\n'''\n\nclass Solution:\n\n def longest_consecutive(self, nums):\n longest_streak = 0\n\n for num in nums:\n current_num = num\n current_streak = 1\n\n while current_num + 1 in nums:\n current_num += 1\n current_streak += 1\n\n longest_streak = max(longest_streak, current_streak)\n\n return longest_streak\n\n '''\n Approach # Sorting [Accepted]\n Intuition\n\n If we can iterate over the numbers in ascending order, then it will be easy to find sequences\n of consecutive numbers. To do so, we can sort the array.\n\n Algorithm\n\n Before we do anything, we check for the base case input of the empty array. The longest \n sequence in an empty array is, of course, 0, so we can simply return that. For \n all other cases, we sort nums and consider each number after the first (because\n we need to compare each number to its previous number). If the current number and \n the previous are equal, then our current sequence is neither extended nor broken, so\n we simply move on to the next number. If they are unequal, then we must check whether \n the current number extends the sequence (i.e. nums[i] == nums[i-1] + 1). If it does,\n then we add to our current count and continue. Otherwise, the sequence is broken, \n so we record our current sequence and reset it to 1 (to include the number that broke \n the sequence). It is possible that the last element of nums is part of the longest sequence,\n so we return the maximum of the current sequence and the longest one.\n\n Sorting Example\n\n [9,1,4,7,3,-1,0,5,8,-1,6] \n |\n \\/\n\n [-1,-1,0,1,3,4,5,6,7,8,9]\n\n Here, an example array is sorted before the linear scan identifies all consecutive sequences.\n The longest sequence is colored in red.\n\n Complexity Analysis:\n Time complexity O(nlgn):\n The main `for` loop does constant work n times, so the algorithm's time complexity is \n dominated by the invocation of sort, which will run O(nlgn) time\n Space complexity O(1) (or O(n))\n For the implementations provided here, the space complexity is constant beacse we sort\n the input array in place. If we are not allowed to modify the input array, we must\n spend liear space to store a sorted copy.\n '''\n\nclass Solution:\n\n\n def longest_consecutive_2(self, nums):\n if not nums:\n return 0\n\n nums.sort()\n\n longest_streak = 1\n current_streak = 1\n\n for i in range(1, len(nums)):\n if nums[i] != nums[i - 1]:\n if nums[i] == nums[i - 1] + 1:\n current_streak += 1\n else:\n longest_streak = max(longest_streak, current_streak)\n current_streak = 1\n\n return max(longest_streak, current_streak)\n\n'''\nApproach #3 HashSet and intelligent sequence building [accepted]\n\nIntuition:\n It turns out that our initial brute force solution was on the right track, but missing\n a few optimizations necessary to reach O(n) time complexity.\n\nAlgorithm:\n The optimised algorithm contains only two changes from the brute forec approach: the \n numbers are stored in a `HashSet` (or `Set` in Python) to allow O(1) lookups, and \n we only attempt to build sequences from numbers that are not already part of a longer\n sequence. This is accomplished by first ensuring that the number that would immediately \n precede the current number in a sequence is not present, as that number would\n necessarily be part of a longer sequence.\n'''\n\nclass Solution:\n\n\n def longest_consecutive_3(self, nums):\n longest_streak = 0\n num_set = set(nums)\n\n for num in num_set:\n if num - 1 not in num_set:\n current_num = num\n current_streak = 1\n\n while current_num + 1 in num_set:\n current_num += 1\n current_streak += 1\n\n longest_streak = max(longest_streak, current_streak)\n\n return longest_streak\n","sub_path":"gs/longest_consecutive_sequence.py","file_name":"longest_consecutive_sequence.py","file_ext":"py","file_size_in_byte":5490,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"520312571","text":"import FWCore.ParameterSet.Config as cms\n\nprocess = cms.Process('RECODQM')\n\n# minimum of logs\nMessageLogger = cms.Service(\"MessageLogger\",\n statistics = cms.untracked.vstring(),\n destinations = cms.untracked.vstring('cerr'),\n cerr = cms.untracked.PSet(\n threshold = cms.untracked.string('WARNING')\n )\n)\n\n# import of standard configurations\nprocess.load('Configuration.StandardSequences.Services_cff')\nprocess.load('Configuration.EventContent.EventContent_cff')\nprocess.load('Configuration.StandardSequences.GeometryRecoDB_cff')\nprocess.load('Configuration.StandardSequences.MagneticField_AutoFromDBCurrent_cff')\nprocess.load('Configuration.StandardSequences.EDMtoMEAtRunEnd_cff')\nprocess.load('Configuration.StandardSequences.FrontierConditions_GlobalTag_cff')\n\n# global tag\nfrom Configuration.AlCa.GlobalTag import GlobalTag\nprocess.GlobalTag = GlobalTag(process.GlobalTag, 'auto:mc', '') #for MC\n\n# load DQM frame work\nprocess.load(\"DQMServices.Core.DQM_cfg\")\nprocess.load(\"DQMServices.Components.DQMEnvironment_cfi\")\n\n# raw data source\n#process.source = cms.Source(\"PoolSource\",\n# fileNames = cms.untracked.vstring('file:/afs/cern.ch/user/j/jkaspar/public/run268608_ls0001_streamA_StorageManager.root')\n#)\n\nprocess.load('TotemRawData.Readers.TotemStandaloneRawDataSource_cfi')\nprocess.source.verbosity = 10\nprocess.source.printProgressFrequency = 0\nprocess.source.fileNames.append('/afs/cern.ch/user/j/jkaspar/public/run_9987_EVB11_1.003.srs')\n\nprocess.maxEvents = cms.untracked.PSet(\n input = cms.untracked.int32(100)\n)\n\n# raw-to-digi conversion\nprocess.load('CondFormats.TotemReadoutObjects.TotemDAQMappingESSourceXML_cfi')\nprocess.TotemDAQMappingESSourceXML.mappingFileNames.append(\"CondFormats/TotemReadoutObjects/xml/totem_rp_210far_220_mapping.xml\")\n\n# process.load('EventFilter.TotemRawToDigi.TotemRPRawToDigi_cfi')\n# process.totemRPRawToDigi.rawDataTag = cms.InputTag(\"rawDataCollector\")\n# process.totemRPRawToDigi.fedIds = cms.vuint32(577, 578, 579, 580)\n# process.totemRPRawToDigi.RawToDigi.printErrorSummary = 0\n# process.totemRPRawToDigi.RawToDigi.printUnknownFrameSummary = 0\n\nprocess.load(\"EventFilter.TotemRawToDigi.totemTriggerRawToDigi_cfi\")\nprocess.totemTriggerRawToDigi.rawDataTag = cms.InputTag(\"source\")\nprocess.totemTriggerRawToDigi.fedId = 0x29c\n\nprocess.load('EventFilter.TotemRawToDigi.totemRPRawToDigi_cfi')\nprocess.totemRPRawToDigi.rawDataTag = cms.InputTag(\"source\")\nprocess.totemRPRawToDigi.fedIds = cms.vuint32(0x1a1, 0x1a2, 0x1a9, 0x1aa, 0x1b5, 0x1bd)\n\n# RP geometry\nprocess.load(\"Geometry.VeryForwardGeometry.geometryRP_cfi\")\nprocess.XMLIdealGeometryESSource.geomXMLFiles.append(\"Geometry/VeryForwardData/data/RP_Garage/RP_Dist_Beam_Cent.xml\")\n\n# local RP reconstruction chain with standard settings\nprocess.load(\"RecoCTPPS.TotemRPLocal.totemRPLocalReconstruction_cff\")\n\n# TOTEM DQM modules\nprocess.load(\"DQM.Totem.totemDAQTriggerDQMSource_cfi\")\nprocess.load(\"DQM.Totem.totemRPDQMSource_cfi\")\n\n# DQM output\nprocess.DQMOutput = cms.OutputModule(\"DQMRootOutputModule\",\n fileName = cms.untracked.string(\"OUT_step1.root\")\n)\n\n# execution schedule\nprocess.reco_step = cms.Path(\n process.totemTriggerRawToDigi *\n process.totemRPRawToDigi *\n process.totemRPLocalReconstruction\n)\n\nprocess.dqm_produce_step = cms.Path(\n process.totemDAQTriggerDQMSource *\n process.totemRPDQMSource\n)\n\nprocess.dqm_output_step = cms.EndPath(\n process.DQMOutput\n)\n\nprocess.schedule = cms.Schedule(\n process.reco_step,\n process.dqm_produce_step,\n process.dqm_output_step\n)\n","sub_path":"DQM/Totem/test/step1_cfg.py","file_name":"step1_cfg.py","file_ext":"py","file_size_in_byte":3521,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"621117638","text":"#!/usr/bin/env python3\n\nimport os\nimport glob\n\nimport pandas as pd\n\n# Returns a tuple.\n# Example: '/Users/john/GitProjects/Python-Baseball/stats', 'data.py'\npath, file_name = os.path.split(os.path.abspath(__file__))\ngame_files = glob.glob(os.path.join(os.path.dirname(__file__), '..', 'games', '*.EVE'))\ngame_files.sort()\n\ngame_frames = []\n# Append game frames\nfor game_file in game_files:\n game_frame = pd.read_csv(game_file, names=['type', 'multi2', 'multi3', 'multi4', 'multi5', 'multi6', 'event'])\n game_frames.append(game_frame)\n\n# Concatenate DataFrames\ngames = pd.concat(game_frames)\n\n# Clean values\ngames.loc[games['multi5'] == '??', ['multi5']] = ''\n\n# Extract identifiers\nidentifiers = games['multi2'].str.extract(r'(.LS(\\d{4})\\d{5})')\n# Forward fill identifiers\nidentifiers = identifiers.fillna(method='ffill')\nidentifiers.columns = ['game_id', 'year']\n\n# Concatenate identifier columns\ngames = pd.concat([games, identifiers], sort=False, axis=1)\n# Fill NaN (Not a Number) values\ngames = games.fillna(' ')\ngames.loc[:, 'type'] = pd.Categorical(games.loc[:, 'type'])\n\n# Print DataFrame\nprint(games.head())\n","sub_path":"stats/data.py","file_name":"data.py","file_ext":"py","file_size_in_byte":1122,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"570648533","text":"# coding: utf-8\n\nimport os, sys, glob\nfrom glob import glob\n\nmdlist = glob('*.md')\ntype(mdlist)\n\n\nfor file in mdlist:\n\n fp = open(file, 'r', encoding='utf-8')\n lines = []\n for line in fp: # 内置的迭代器, 效率很高\n lines.append(line.rstrip())\n fp.close()\n\n lines.pop(3)\n lines.insert(1, '')\n\n s = '\\n'.join(lines)\n\n fp = open(file, 'w', encoding='utf-8')\n fp.write(s)\n fp.close()\n","sub_path":"TOOL_mdAddToc/mdAddToc3.py","file_name":"mdAddToc3.py","file_ext":"py","file_size_in_byte":505,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"283650627","text":"# Copyright 2018 ICON Foundation\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"helper class for TxItem\"\"\"\n\nimport json\nimport sys\n\nfrom loopchain.blockchain.transactions import Transaction, TransactionVersioner, TransactionSerializer\nfrom loopchain.protos import loopchain_pb2\n\n\nclass TxItem:\n tx_serializers = {}\n\n def __init__(self, tx_json: str, channel: str):\n self.channel = channel\n self.__tx_json = tx_json\n self.__len = sys.getsizeof(tx_json) + sys.getsizeof(channel)\n\n def __len__(self):\n return self.__len\n\n def get_tx_message(self):\n message = loopchain_pb2.TxSend(\n tx_json=self.__tx_json,\n channel=self.channel)\n return message\n\n @classmethod\n def create_tx_item(cls, tx_param: tuple, channel: str):\n tx, tx_versioner = tx_param\n tx_serializer = cls.get_serializer(tx, tx_versioner)\n tx_item = TxItem(\n json.dumps(tx_serializer.to_raw_data(tx)),\n channel\n )\n return tx_item\n\n @classmethod\n def get_serializer(cls, tx: Transaction, tx_versioner: TransactionVersioner):\n if tx.version not in cls.tx_serializers:\n cls.tx_serializers[tx.version] = TransactionSerializer.new(tx.version, tx.type(), tx_versioner)\n return cls.tx_serializers[tx.version]\n","sub_path":"loopchain/baseservice/tx_item_helper.py","file_name":"tx_item_helper.py","file_ext":"py","file_size_in_byte":1832,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"596498614","text":"import numpy as np\nimport pandas as pd\n# import model\nimport model\nimport itertools\nimport multiprocessing as mp\nimport pickle\n\n# Prepare network and data\nprep_data = model.data_and_network_prep()\n\n# Create data for the fit\ndata_for_fit_i = model.create_data_for_fit_influenza()\n\ndata_for_fit_v = model.create_data_for_fit(prep_data)\n\n\n# Get parameters - vaccination model\nwith open('../../Data/vaccination_model/grid_search_5_res.pickle', 'rb') as pickle_in:\n grid_search_res_1 = pickle.load(pickle_in)\n\nwith open('../../Data/vaccination_model/grid_search_6_res.pickle', 'rb') as pickle_in:\n grid_search_res_2 = pickle.load(pickle_in)\n\nwith open('../../Data/vaccination_model/grid_search_7_res.pickle', 'rb') as pickle_in:\n grid_search_res_3 = pickle.load(pickle_in)\n\ngrid_search_res_v = grid_search_res_1 + grid_search_res_2 + grid_search_res_3\n\n# Max likelihood subdist\nliklihood_subdist_v = max(grid_search_res_v, key=lambda x: x['log_likelihood_subdist'])\nliklihood_subdist_v['parameters']['beta'] = liklihood_subdist_v['parameters']['beta_2']\n\n# Get parameters - influenza model\n# with open('L:/Dor/data/coupled_model/grid_search_2016_1_res.pickle', 'rb') as pickle_in:\n# grid_search_res_i = pickle.load(pickle_in)\n\n# # Max likelihood subdist\n# liklihood_subdist_i = max(grid_search_res_i, key=lambda x: x['log_likelihood_subdist'])\n\n# Load parameters_i all seasons\nwith open('../../data/coupled_model/parameters_i_all_seasons.pickle', 'rb') as pickle_in:\n parameters_i = pickle.load(pickle_in)\n\n# Set parameters\nparameters_v = liklihood_subdist_v['parameters']\n# parameters_i = liklihood_subdist_i['parameters']\n\n\n# intervention type\n# inter_type = 'random'\n# inter_type = 'by_area'\n# inter_type = 'by_subdist'\n# inter_type = 'by_yeshuv'\n\n\n# intervention parameters\nlength = 5\n\n# Number of simulations\nm = 40\n\n# Set intervention parameters\n# Intervention percents\ninter_percents = [0.005, 0.01, 0.02, 0.03, 0.05]\n# inter_percents = [0.01, 0.025, 0.05]\n# inter_percents = [0.04]\n\n# Vaccination season start time: 1.8, 1.9, 1.10, 1.11\nstart_times = [61, 92, 122, 153]\n\n# Intervention times (by start time)\nall_inter_times = [61, 92, 122, 153, 183, 214] # 1.8, 1.9, 1.10, 1.11, 1.12, 1.1\nintervention_times = {61: all_inter_times, 92: all_inter_times[1:], 122: all_inter_times[2:], 153: all_inter_times[3:]}\n\n# Random intervention dict\ninterventions_dict_random = {inter_percent:\n {start_time: [{'time': time, 'percent': inter_percent, 'len': length,\n 'vacc_start': start_time, 'type': 'random'}\n for time in intervention_times[start_time]]\n for start_time in start_times}\n for inter_percent in inter_percents}\n\n# Intervention by area\n# Load nodes by area and age\nwith open(model.nodes_by_area_age_dict_path, 'rb') as pickle_in:\n nodes_by_area_age = pickle.load(pickle_in)\n\n# Load nodes by area and age\nwith open(model.nodes_by_area_age_dict_path, 'rb') as pickle_in:\n nodes_by_area_age = pickle.load(pickle_in)\n\n# Load page ranks\nwith open(model.pagerank_by_area_age_path, 'rb') as pickle_in:\n pageranks = pickle.load(pickle_in)\n\n# Sort areas and age groups by page rank (descending)\nareas_age_by_rank_with_rank = sorted(list(pageranks.items()), key=lambda x: x[1], reverse=True)\nareas_age_by_rank = list(map(lambda x: x[0], areas_age_by_rank_with_rank))\n\n# Filter irrelevant areas\nareas_age_by_rank = list(filter(lambda x: x in nodes_by_area_age, areas_age_by_rank))\nareas_age_by_rank_with_rank = list(filter(lambda x: x[0] in nodes_by_area_age, areas_age_by_rank_with_rank))\n\n# Create a list of nodes by PageRank\nnodes_by_rank = []\nfor (area, age) in areas_age_by_rank:\n nodes_by_rank += list(nodes_by_area_age[(area, age)])\n\n# Create intervention dict\ninterventions_dict_area = {inter_percent:\n {start_time: [{'time': time, 'percent': inter_percent, 'len': length, 'vacc_start': start_time, 'type': 'by_area',\n 'nodes_by_rank': nodes_by_rank} for time in intervention_times[start_time]]\n for start_time in start_times}\n for inter_percent in inter_percents}\n\n# Load page ranks by subdist\nwith open('../../Data/vaccination_data/pagerank_by_subdist_age.pickle', 'rb') as pickle_in:\n pageranks_subdist = pickle.load(pickle_in)\n\n# Sort areas and age groups by page rank (descending)\nsubdists_age_by_rank_with_rank = sorted(list(pageranks_subdist.items()), key=lambda x: x[1], reverse=True)\nsubdists_age = list(map(lambda x: x[0], subdists_age_by_rank_with_rank))\nsubdist_ranks = list(map(lambda x: x[1], subdists_age_by_rank_with_rank))\n\n# Create intervention dict\ninterventions_dict_subdist = {inter_percent:\n {start_time: [{'time': time, 'percent': inter_percent, 'len': length, 'vacc_start': start_time,\n 'type': 'by_subdist', 'subdists_age': subdists_age, 'subdist_ranks': subdist_ranks}\n for time in intervention_times[start_time]]\n for start_time in start_times}\n for inter_percent in inter_percents}\n\n# Load page ranks by yeshuv\nwith open('../../Data/vaccination_data/pagerank_by_yeshuv_age.pickle', 'rb') as pickle_in:\n pageranks_yeshuv = pickle.load(pickle_in)\n\n# Sort areas and age groups by page rank (descending)\nyeshuv_age_by_rank_with_rank = sorted(list(pageranks_yeshuv.items()), key=lambda x: x[1], reverse=True)\nyeshuv_age = list(map(lambda x: x[0], yeshuv_age_by_rank_with_rank))\nyeshuv_ranks = list(map(lambda x: x[1], yeshuv_age_by_rank_with_rank))\n\n# Create intervention dict\ninterventions_dict_yeshuv = {inter_percent:\n {start_time: [{'time': time, 'percent': inter_percent, 'len': length, 'vacc_start': start_time,\n 'type': 'by_yeshuv', 'yeshuv_age': yeshuv_age, 'yeshuv_ranks': yeshuv_ranks}\n for time in intervention_times[start_time]]\n for start_time in start_times}\n for inter_percent in inter_percents}\n\n\n# Define a function for multiprocessing\ndef intervention_mp(intervention):\n print(mp.current_process())\n\n # Run the model with current intervention\n inter_res = model.intervention_coupled_model(parameters_i, parameters_v, prep_data, data_for_fit_i, data_for_fit_v, intervention,\n intervention['vacc_start'], num_of_simulations=m)\n\n return {(intervention['percent'], intervention['vacc_start'], intervention['time']): inter_res}\n\n\ndef mp_handler(interventions_list):\n # Create a pool of processes\n pool = mp.Pool(24)\n # Process in parallel\n results = pool.map(intervention_mp, interventions_list)\n return results\n\n\nif __name__ == '__main__':\n # Intervention dicts to go over\n # inter_dicts = [interventions_dict_random, interventions_dict_subdist, interventions_dict_yeshuv,\n # interventions_dict_area]\n inter_dicts = [interventions_dict_yeshuv]\n\n # File names for save\n # file_names = [f'random_intervention_res_all_seasons_m{m}', f'subdist_intervention_res_all_seasons_m{m}',\n # f'yeshuv_intervention_res_all_seasons_m{m}', f'area_intervention_res_all_seasons_m{m}']\n file_names = [f'yeshuv_intervention_res_all_seasons_m{m}_all_per']\n\n # Go over the dicts and run all interventions\n for i, interventions_dict in enumerate(inter_dicts):\n # Unpack dictionary\n interventions_list = []\n for inter_percent in interventions_dict:\n for start_time, inter in interventions_dict[inter_percent].items():\n interventions_list.extend(inter)\n\n # Get results using mp\n res = mp_handler(interventions_list)\n\n # Saving the results\n with open(f'../../data/coupled_model/{file_names[i]}.pickle', 'wb') as pickle_out:\n pickle.dump(res, pickle_out)\n","sub_path":"influenza_modeling/I_Coupled model intervention - multiprocessing.py","file_name":"I_Coupled model intervention - multiprocessing.py","file_ext":"py","file_size_in_byte":8166,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"}